cloudera manager installation process
cd /etc/hostname //change hostname
---------------------------------------------------------------------------------------------------------
Hadoop cluster setup
---------------------------------------------------------------------------------------------------------
1. create a new 'cluster' user in ubuntu.
2. create '/home/cluster/work' and '/home/cluster/work/hadoopdata' folders
mkdir /home/cluster/work
mkdir /home/cluster/work/hadoopdata
3. Download 'hadoop-1.0.4.tar.gz' file from hadoop repository, copy 'hadoop-1.0.4.tar.gz' file into this
'/home/cluster/work' directory and extract the tar file in same directory
tar -xvzf hadoop-1.0.4.tar.gz .
4. Open the '~/.bashrc' file on all the machines and add the following lines at the end and save:
command: gedit ~/.bashrc
export JAVA_HOME=/usr/lib/jvm/java-1.6.0-openjdk
export HADOOP_HOME=/home/cluster/work/hadoop-1.0.4
export PATH=$HADOOP_HOME/bin:$JAVA_HOME/bin:$PATH
5. Open the '/etc/hostname' file on all the machines update as per machine details and save
hostname will be namenode in namenode machine
hostname will be jobtracker in jobtracker machine
hostname will be secondarynamenode in secondarynamenode machine
hostname will be slave1 in slave1 machine
hostname will be slave2 in slave2 machine
hostname will be slave3 in slave3 machine
6. Open the '/etc/hosts' file on all the machines and add the following lines as per your network details and save:
192.168.41.1 namenode
192.168.41.2 jobtracker
192.168.41.3 secondarynamenode
192.168.41.4 slave1
192.168.41.5 slave2
192.168.41.6 slave3
7. Enter the below commands on terminal:
1. 'jobtracker', 'secondarynamenode', 'slave1', 'slave2' and 'slave3' machines run the following commands on terminal:
ssh localhost
ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
2. 'namenode' machine run the following commands on terminal:
ssh localhost
ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys
sudo rm -r .ssh
chmod 600 .ssh/authorized_keys
3. copy the authorized keys from 'namenode' machine to all other machines using the below commands:
scp ~/.ssh/id_dsa.pub cluster@jobtracker:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@secondarynamenode:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@slave1:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@slave2:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@slave3:~/.ssh/authorized_keys
8. Update the '$HADOOP_HOME/conf' folder files 'core-site.xml','hdfs-site.xml','mapred-site.xml','masters' and 'slaves' files
as per the above configurations
9. Format the 'namenode' from 'namenode' machine using this command:
hadoop namenode -format
10. Start the hadoop cluster by using this command on 'namenode' machine:
start-all.sh
11. Stop the hadoop cluster by using this command on 'namenode' machine:
stop-all.sh
---------------------------------------
sudo su
<enter ur password>
cp -r /home/<src_username>/Desktop/kalyan_admin_training_class /home/<dest_username>/Desktop/
chown -R <dest_username>:<dest_username> /home/<dest_username>/Desktop/kalyan_admin_training_class
example: copy data from hadoop user to kalyan user
--------------------------------------------------
sudo su
<enter ur password>
cp -r /home/hadoop/Desktop/kalyan_admin_training_class /home/kalyan/Desktop/
chown -R kalyan:kalyan /home/kalyan/Desktop/kalyan_admin_training_class
exit
---------------------------------------
change hostname
2.6 files configuration files;
https://drive.google.com/file/d/0B7ZpDaKA88ZQaDd5MFZLMTh5ZG8/
follow this for more help.
http://dogdogfish.com/2014/04/22/hadoop-from-spare-change/
Download from this link
chmod u+x cloudera-manager-installer.bin
sudo ./cloudera-manager-installer.bin;
More help:
---------------------------------------------------------------------------------------------------------
Hadoop cluster setup
---------------------------------------------------------------------------------------------------------
1. create a new 'cluster' user in ubuntu.
2. create '/home/cluster/work' and '/home/cluster/work/hadoopdata' folders
mkdir /home/cluster/work
mkdir /home/cluster/work/hadoopdata
3. Download 'hadoop-1.0.4.tar.gz' file from hadoop repository, copy 'hadoop-1.0.4.tar.gz' file into this
'/home/cluster/work' directory and extract the tar file in same directory
tar -xvzf hadoop-1.0.4.tar.gz .
4. Open the '~/.bashrc' file on all the machines and add the following lines at the end and save:
command: gedit ~/.bashrc
export JAVA_HOME=/usr/lib/jvm/java-1.6.0-openjdk
export HADOOP_HOME=/home/cluster/work/hadoop-1.0.4
export PATH=$HADOOP_HOME/bin:$JAVA_HOME/bin:$PATH
5. Open the '/etc/hostname' file on all the machines update as per machine details and save
hostname will be namenode in namenode machine
hostname will be jobtracker in jobtracker machine
hostname will be secondarynamenode in secondarynamenode machine
hostname will be slave1 in slave1 machine
hostname will be slave2 in slave2 machine
hostname will be slave3 in slave3 machine
6. Open the '/etc/hosts' file on all the machines and add the following lines as per your network details and save:
192.168.41.1 namenode
192.168.41.2 jobtracker
192.168.41.3 secondarynamenode
192.168.41.4 slave1
192.168.41.5 slave2
192.168.41.6 slave3
7. Enter the below commands on terminal:
1. 'jobtracker', 'secondarynamenode', 'slave1', 'slave2' and 'slave3' machines run the following commands on terminal:
ssh localhost
ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
2. 'namenode' machine run the following commands on terminal:
ssh localhost
ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys
sudo rm -r .ssh
chmod 600 .ssh/authorized_keys
3. copy the authorized keys from 'namenode' machine to all other machines using the below commands:
scp ~/.ssh/id_dsa.pub cluster@jobtracker:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@secondarynamenode:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@slave1:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@slave2:~/.ssh/authorized_keys
scp ~/.ssh/id_dsa.pub cluster@slave3:~/.ssh/authorized_keys
8. Update the '$HADOOP_HOME/conf' folder files 'core-site.xml','hdfs-site.xml','mapred-site.xml','masters' and 'slaves' files
as per the above configurations
9. Format the 'namenode' from 'namenode' machine using this command:
hadoop namenode -format
10. Start the hadoop cluster by using this command on 'namenode' machine:
start-all.sh
11. Stop the hadoop cluster by using this command on 'namenode' machine:
stop-all.sh
---------------------------------------
sudo su
<enter ur password>
cp -r /home/<src_username>/Desktop/kalyan_admin_training_class /home/<dest_username>/Desktop/
chown -R <dest_username>:<dest_username> /home/<dest_username>/Desktop/kalyan_admin_training_class
example: copy data from hadoop user to kalyan user
--------------------------------------------------
sudo su
<enter ur password>
cp -r /home/hadoop/Desktop/kalyan_admin_training_class /home/kalyan/Desktop/
chown -R kalyan:kalyan /home/kalyan/Desktop/kalyan_admin_training_class
exit
---------------------------------------
change hostname
sudo sysctl -p2.6 files configuration files;
https://drive.google.com/file/d/0B7ZpDaKA88ZQaDd5MFZLMTh5ZG8/
follow this for more help.
http://dogdogfish.com/2014/04/22/hadoop-from-spare-change/
Download from this link
chmod u+x cloudera-manager-installer.bin
sudo ./cloudera-manager-installer.bin;
More help:

0 Comments:
Post a Comment
Subscribe to Post Comments [Atom]
<< Home