How to disable Copy, Paste and Text Selection using CSS

Aired Consulting 4/04/2014 No Comment

Disable Copy/Paste and Text Selection using CSS and Javascript.

CSS Code to disable text selection and highlighting.

This CSS code will disable text highlighting and selecting for your website/blog post. This CSS code works perfectly in Internet Explorer, Mozilla Firefox and Google Chrome. In blogger you may put the code inside the .post-body section.
-webkit-user-select: none;
-khtml-user-drag: none;
-khtml-user-select: none;
-moz-user-select: none;
-moz-user-select: -moz-none;
-ms-user-select: none;
-o-user-select: none;
user-select: none;
One thing to note about the above CSS code is that in Firefox, it will also disable Copy/Paste of the content of your post/website. However, the copy will still be possible in Chrome browser.

CSS Code to enable text selection in Blockquote.

There may be cases where a certain portion of your post may still need the highlighting and selection, let say in the blockquote area where you may write certain sample code. In that case just you need to put
.post blockquote {
-webkit-touch-callout: text;
-khtml-user-select: text;
 -moz-user-select: text;
 -ms-user-select: text;
 user-select: text;
 -o-user-select:text;
 -webkit-user-select: text;
}

Disable Copy/Paste of you content.

To disable copy/paste in the blogger and Wordpress posts copy the below code into you theme template. This will make the task of copying text a little complex for content scrappers.
<script type='text/JavaScript'>
function killCopy(e){
return false
}
function reEnable(){
return true
}
document.onselectstart=new Function (&quot;return false&quot;)
if (window.sidebar){
document.onmousedown=killCopy
document.onclick=reEnable
}
</script>

Online HTML to XML Parser

Aired Consulting 3/04/2014 No Comment

HTML to XML Converter/Parser Tool. Convert your Blogger code from HTML to XML.

Convert your Adsense, Chitika or any other HTML code to XML and make it compatible with blogger templates.

Since Blogger doesn't allow HTML code, this tool will convert all the Javascript and HTML code to XML safely for use in Blogger.
Enter raw code here


Get escaped code here

HTML Table Generator

Aired Consulting 2/19/2014 1 Comment

Create a CSS styled HTML Table of your choice in real quick time.

This online tool will let you build and design your HTML table from a variety of styles. You just need to copy the HTML code generated into your website.

Please use the options provided below to create your custom HTML table.

Table Style: Highlight:

Table Structure: Columns Rows Cell Info:

Copy the HTML code generated below into your website

How HTML Table will look in your website or blog.

Header 1Header 2Header 3Header 4Header 5
Row:1 Cell:1Row:1 Cell:2Row:1 Cell:3Row:1 Cell:4Row:1 Cell:5
Row:2 Cell:1Row:2 Cell:2Row:2 Cell:3Row:2 Cell:4Row:2 Cell:5
Row:3 Cell:1Row:3 Cell:2Row:3 Cell:3Row:3 Cell:4Row:3 Cell:5
Row:4 Cell:1Row:4 Cell:2Row:4 Cell:3Row:4 Cell:4Row:4 Cell:5
Row:5 Cell:1Row:5 Cell:2Row:5 Cell:3Row:5 Cell:4Row:5 Cell:5
Row:6 Cell:1Row:6 Cell:2Row:6 Cell:3Row:6 Cell:4Row:6 Cell:5

CSS Code to make the footer stay at the bottom of a page

Aired Consulting 2/14/2014 No Comment

HTML page footer stay at bottom of the page.

CSS code for Footer at the bottom of a page.

Main body page is stretched to a 100% of the page. The footer is then given the below CSS rules
#footer {
clear: both;
position: relative;
z-index: 10;
text-align: center;
height:30px;
}

CSS code for Fixed Footer at the bottom.

There is another approach which will work on all browsers, though it will appear at the bottom, the footer will be placed over the content. As such, the footer will be at the bottom of the window even when you scroll.
#footer {
position:fixed;
bottom:0;
left:0;
right:0;
width:100%;
z-index: -999;
overflow: hidden;
}

Hadoop Cluster Interview Questions Answers

Aired Consulting 9/25/2013 No Comment

Apache Hadoop Cluster Interview Questions.

Explain about the Hadoop-core configuration files?
Hadoop core is specified by two resources. Its is configured by two well written xml files which are loaded from the classpath:
1. hadoop-default.xml  -  Read-only defaults for Hadoop, suitable for a single machine instance
2. hadoop-site.xml. - It specifies the site configuration for Hadoop distribution. The cluster specific information is also provided by the Hadoop administrator.

Explain in brief the three modes in which Hadoop can be run?
The three modes in which Hadoop can be run are:
1. Standalone (local) mode - No Hadoop daemons running, everything runs on a single Java Virtual machine.
2. Pseudo-distributed mode - Daemons run on the local machine, thereby simulating a cluster on a smaller sclae.
3. Fully distributed mode - Runs on a cluster of machines.

Explain what are the features of Stand alone (local) mode?

In stand-alone or local mode there are no Hadoop daemons running,  and everything runs on a single Java process. Hence we don't get the benefit of distributing the code across a cluster of machines. Since, it has no DFS, it utilizes the local file system. This mode is suitable only for running MapReduce programs by developers during various stages of development. Its the best environment for learning and good for debugging purposes.

What are the main features of Pseudo mode?

In Pseudo-distributed mode, each Hadoop daemon runs in a separate Java process, as such it simulates a cluster though on a small scale. This mode is used both for development and QA environments. Here , we need to do the configuration changes.

What are the features of Fully Distributed mode?

In Fully Distributed mode, the clusters range from a few nodes to 'n' number of nodes. It is used in the production environment, where we have thousands of  machines in the Hadoop cluster. The daemons of Hadoop run on these clusters.We have to configure separate masters and separate slaves in this distribution, the implementation of which is quite complex. In this configuration, Namenode and Datanode runs on different hosts and there are nodes on which task tracker runs. The root of the distribution is referred as HADOOP_HOME.

What are the Hadoop configuration files at present?
There are 3 configuration files in Hadoop:
1. conf/core-site.xml:
<configuration>     <property>         <name>fs.default.name</name>         <value>hdfs://localhost:9000</value>     </property></configuration>
2. conf/hdfs-site.xml:
<configuration>     <property>         <name>dfs.replication</name>         <value>1</value>     </property></configuration>

3. conf/mapred-site.xml:
<configuration>     <property>         <name>mapred.job.tracker</name>         <value>localhost:9001</value>     </property></configuration>
Does Hadoop follows the UNIX pattern?
Yes, Hadoop closely follows the UNIX pattern. Hadoop also has the ‘conf‘ directory as in the
case of UNIX.

In which directory Hadoop is installed?

Cloudera and Apache has the same directory structure. Hadoop is installed in cd
/usr/lib/hadoop-0.20/.

What are the port numbers of Namenode, job tracker and task tracker?

The port number for Namenode is ’70′, for job tracker is ’30′ and for task tracker is ’60′.

What is a spill factor with respect to the RAM?
Spill factor is the size after which your files move to the temp file. Hadoop-temp directory is
used for this.

Is fs.mapr.working.dir a single directory?

Yes, fs.mapr.working.dir it is just one directory.

Which are the three main hdfs-site.xml properties?

The three main hdfs-site.xml properties are:
1. dfs.name.dir which gives you the location on which metadata will be stored and where
DFS is located – on disk or onto the remote.
2. dfs.data.dir which gives you the location where the data is going to be stored.
3. fs.checkpoint.dir which is for secondary Namenode.

How to come out of the insert mode?

To come out of the insert mode, press ESC, type :q (if you have not written anything) OR
type :wq (if you have written anything in the file) and then press ENTER.

What is Cloudera and why it is used?

Cloudera is the distribution of Hadoop. It is a user created on VM by default. Cloudera
belongs to Apache and is used for data processing.

What happens if you get a ‘connection refused java exception’ when you type hadoop

fsck /?
It could mean that the Namenode is not working on your VM.

We are using Ubuntu operating system with Cloudera, but from where we can

download Hadoop or does it come by default with Ubuntu?
This is a default configuration of Hadoop that you have to download from Cloudera or from
Edureka’s dropbox and the run it on your systems. You can also proceed with your own
configuration but you need a Linux box, be it Ubuntu or Red hat. There are installation
steps present at the Cloudera location or in Edureka’s Drop box. You can go either ways.

What does ‘jps’ command do?

This command checks whether your Namenode, datanode, task tracker, job tracker, etc are
working or not.

How can I restart Namenode?

1. Click on stop-all.sh and then click on start-all.sh OR
2. Write sudo hdfs (press enter), su-hdfs (press enter), /etc/init.d/ha (press enter) and
then /etc/init.d/hadoop-0.20-namenode start (press enter).

What is the full form of fsck?

Full form of fsck is File System Check.

How can we check whether Namenode is working or not?

To check whether Namenode is working or not, use the command /etc/init.d/hadoop-
0.20-namenode status or as simple as jps.

What does the command mapred.job.tracker do?

The command mapred.job.tracker lists out which of your nodes is acting as a job tracker.

What does /etc /init.d do?

/etc /init.d specifies where daemons (services) are placed or to see the status of these
daemons. It is very LINUX specific, and nothing to do with Hadoop.

How can we look for the Namenode in the browser?

If you have to look for Namenode in the browser, you don’t have to give localhost:8021, the
port number to look for Namenode in the brower is 50070.

How to change from SU to Cloudera?

To change from SU to Cloudera just type exit.

Which files are used by the startup and shutdown commands?

Slaves and Masters are used by the startup and the shutdown commands.

What do slaves consist of?

Slaves consist of a list of hosts, one per line, that host datanode and task tracker servers.

What do masters consist of?

Masters contain a list of hosts, one per line, that are to host secondary namenode servers.

What does hadoop-env.sh do?

hadoop-env.sh provides the environment for Hadoop to run. JAVA_HOME is set over here.

Can we have multiple entries in the master files?

Yes, we can have multiple entries in the Master files.

Where is hadoop-env.sh file present?

hadoop-env.sh file is present in the conf location.

In Hadoop_PID_DIR, what does PID stands for?

PID stands for ‘Process ID’.

What does /var/hadoop/pids do?

It stores the PID.

What does hadoop-metrics.properties file do?

hadoop-metrics.properties is used for ‘Reporting‘ purposes. It controls the reporting for
Hadoop. The default status is ‘not to report‘.

What are the network requirements for Hadoop?

The Hadoop core uses Shell (SSH) to launch the server processes on the slave nodes. It
requires password-less SSH connection between the master and all the slaves and the
secondary machines.

Why do we need a password-less SSH in Fully Distributed environment?

We need a password-less SSH in a Fully-Distributed environment because when the cluster
is LIVE and running in Fully
Distributed environment, the communication is too frequent. The job tracker should be
able to send a task to task tracker quickly.

Does this lead to security issues?

No, not at all. Hadoop cluster is an isolated cluster. And generally it has nothing to do with
an internet. It has a different kind of a configuration. We needn’t worry about that kind of a
security breach, for instance, someone hacking through the internet, and so on. Hadoop has
a very secured way to connect to other machines to fetch and to process data.

On which port does SSH work?

SSH works on Port No. 22, though it can be configured. 22 is the default Port number.

Can you tell us more about SSH?

SSH is nothing but a secure shell communication, it is a kind of a protocol that works on a
Port No. 22, and when you do an SSH, what you really require is a password.

Why password is needed in SSH localhost?

Password is required in SSH for security and in a situation where passwordless
communication is not set.

Do we need to give a password, even if the key is added in SSH?

Yes, password is still required even if the key is added in SSH.

What if a Namenode has no data?

If a Namenode has no data it is not a Namenode. Practically, Namenode will have some
data.

What happens to job tracker when Namenode is down?

When Namenode is down, your cluster is OFF, this is because Namenode is the single point
of failure in HDFS.

What happens to a Namenode, when job tracker is down?

When a job tracker is down, it will not be functional but Namenode will be present. So,
cluster is accessible if Namenode is working, even if the job tracker is not working.

Can you give us some more details about SSH communication between Masters and

the Slaves?
SSH is a password-less secure communication where data packets are sent across the slave.
It has some format into which data is sent across. SSH is not only between masters and
slaves but also between two hosts.

What is formatting of the DFS?

Just like we do for Windows, DFS is formatted for proper structuring. It is not usually done
as it formats the Namenode too.

Does the HDFS client decide the input split or Namenode?

No, the Client does not decide. It is already specified in one of the configurations through
which input split is already configured.

In Cloudera there is already a cluster, but if I want to form a cluster on Ubuntu can

we do it?
Yes, you can go ahead with this! There are installation steps for creating a new cluster. You
can uninstall your present cluster and install the new cluster.

Can we create a Hadoop cluster from scratch?

Yes we can do that also once we are familiar with the Hadoop environment.

Can we use Windows for Hadoop?

Actually, Red Hat Linux or Ubuntu are the best Operating Systems for Hadoop. Windows is
not used frequently for installing Hadoop as there are many support problems attached
with Windows. Thus, Windows is not a preferred environment for Hadoop.

Big Data Resources - Interview Questions, Downloads, Tutorials

Aired Consulting 9/22/2013 No Comment

Big Data Interview Questions, Downloads, Tutorials.

Apache Hadoop Interview Questions Answers

MapReduce Interview Questions

Apache Pig Interview Questions Answers

Big Data Downloads 

 

Siebel Server Down Troubleshooting

Aired Consulting 9/22/2013 No Comment

Siebel Server is not coming up in Linux/Unix and Windows.

Server Busy Error for Siebel Server. Steps for Troubleshooting.

1. Ensure that you validate that the .srf file is not corrupt by placing the SRF on a dedicated environment. If the dedicated client is facing issues there could be possibly  two reasons as cited below.
  • The SRF file in the server got corrupt.
  • The Oracle Database Server itself is down
2. Check that no Siebel processes for the enterprise are still running.Windows: Check Task Manager for any Siebel process for the enterprise still running.

3. Stop siebel servers after executing ./siebenv.sh and command stop_server all

Solaris: Execute ps -ef | grep [directory path] (eg. ps -ef | grep /app/siebel/siebsrvr).
ps -ef | grep sieb
Ensure that all processes for that enterprise are killed.
use kill -9 pid

4. Delete any file that exists in directory %SIEBEL_ROOT%\sys with name like:
osdf.[SiebelEnterprise].[SiebelServer]
Where
[SiebelEnterprise] = The Siebel Enterprise name
[SiebelServer] = The Siebel Server name.

5. Delete any file that exists in directory %SIEBEL_ROOT%\admin with name like:*.shm
dhm files are shared memory files. This file should be automatically deleted when the Siebel server is
shut down, if it still exists when the Siebel server is down then it has been corrupted and not correctly removed.

6.Delete fdr and core files as these file eat up large amounts of memory.

7.Cleanup unwanted logarchive and log files so that fresh logs can be monitered and space can be freed up.

8.Try to restart server after executing ./siebenv.sh and command start_server all
Note- At this point if the server still does not restart, you need to check the enterprise log for the reason.
The enterprise log is located in:
%SIEBEL_ROOT%\enterprises\[SiebelEnterprise]\[SiebelServer]\log
The enterprise log has name with format:
[SiebelEnterprise].[SiebelServer].log

9. If no enterprise logs are getting creted there are connectivity issues with the database:
i.e change of db password for SADMIN user
run odbcsql from siebsrvr/bin to check connectivity issues
odbcsql /u SADMIN /p SADMIN /s DSN Name

10. Any changes which lead to the corruption of siebns.dat will also result in the servers not coming up.
usually NameSrvr logs tell us connectivity related information and errors like key not found.Try reverting
to an old working siebns.dat file.

11.If the environment is LDAP authenticated any changes in the LDAP trees can also affect the environment.
please verify the same.

12.Check the SCBroker and SRBroker logs;you would get a hint.

13.Use netstat -an|grep 2320 for verifying that the gateway service port is listening.

14.Use netstat -an|grep 2321 for verifing that the SRBroker/SCBroker port is listening

Big Data and Apache Hadoop Free Downloads

Aired Consulting 9/21/2013 No Comment

Big Data and Hadoop related Downloads. 

Download Hadoop for Windows:

Hadoop is a powerful framework for automatic parallelization of computing tasks. Mostly available for Unix/Linux, this installer help develop applications and analyze big data stored in Apache Hadoop running on Microsoft Windows.

Download Hortonworks Hadoop Sandbox:

Learn Hadoop with Hortonworks Sandbox. A free download that comes with many interactive Hadoop tutorials.

Download Cloudera Hadoop :

Revel insights from all your data, store everything forever without data loss or archiving, and make data an integral component of your enterprise.

Download MapR Hadoop :

MapR delivers on the promise of Hadoop with a proven, enterprise-grade Big Data platform that supports a broad set of mission-critical and real-time production.

Download Cloudera Impala :

Cloudera Impala is an open source Massively Parallel Processing (MPP) query engine that runs natively on Apache Hadoop.

Download Cassandra :

Apache Cassandra is an open source distributed database management system designed to handle large amounts of data across many commodity servers.

MongoDB 3rd-Party Admin Tools:

Monitoring is a critical component of all database administration. A number of third party monitoring tools have support for MongoDB. There are some very good tools available in the MongoDB package but the list provided here will help you a lot in administration.

Download Oracle NoSQL Database:

Oracle NoSQL Database is a distributed, highly performance, highly available scalable key-value database. The Oracle NoSQL Database is a noSQL-type distributed key-value database by Oracle

Download Couchbase Server:

Couchbase Server, a distributed, non-relational NoSQL database that can easily accommodate changing data management needs.

Download Neo4J :

Neo4j is an open-source graph database, implemented in Java. Developers describe Neo4j as "embedded, disk-based, fully transnational Java persistence engine that stores data structured in graphs rather than in tables". Neo4j is the most popular graph database.

Continuity AppFabric :

Delivered as a cloud PaaS, the Continuity AppFabric is an application run-time and data platform, which sits on top of open source Hadoop.

Amazon Hadoop/Mapreduce:

Amazon Elastic MapReduce automatically spins up a Hadoop implementation of the MapReduce framework on Amazon EC2 instances. Amazon Elastic MapReduce (Amazon EMR) makes it easy to provision and manageHadoop in the AWS Cloud.

Download Spring For Hadoop:

Spring for Apache Hadoop is a framework for application developers to take advantage of the features of both Hadoop and Spring.

MORTAR Hadoop Platform :

Mortar have a great platform for leveraging Hadoop, Pig and Python. Mortar is the fastest and easiest way to work with Pig.

I will update other Big Data and Apache Hadoop Free Downloads in my next post.
 

Aired | Copyright © 2009-2014| All Rights Reserved 2014 | Privacy Policy and Disclaimer

Contact Us | About Us