原文链接 下面自己做了几个小的修正。 (是安装在VirtualBox下的Ubuntu Server里的) Ubuntu is 14.04 主机名(hostname)是 master 用户名是 hadoop 自己也disable了ipv6 Disabling IPv6 Since Hadoop doesn’t work on IPv6, we should disable it. One of another reason is also that it has been developed and tested on IPv4 stacks. Hadoop nodes will be able to communicate if we are having IPv4 cluster. (Once you have disabled IPV6 on your machine, you need to reboot your machine in order to check its effect. In case if you don’t know how to reboot with command use sudo reboot ) For getting your IPv6 disable in your Linux machine, you need to update /etc/sysctl.conf by adding following line of codes at end of the file, # disable ipv6 net.ipv6.conf.all.disable_ipv6 = 1 net.ipv6.conf.default.disable_ipv6 = 1 net.ipv6.conf.lo.disable_ipv6 = 1 ******************************************************************************************************** $ sudo apt-get update $ sudo apt-get install default-jdk $ java -version $ sudo apt-get install ssh $ sudo apt-get install rsync $cd ~/.ssh # 如果没有该目录,先执行一次ssh localhost $ssh-keygen -t rsa # 一直按回车就可以 $cp id_rsa.pub authorized_keys 此时再用 $ wget -c http://mirror./pub/apache/hadoop/common/current/hadoop-2.6.0.tar.gz $ sudo tar -zxvf hadoop-2.6.0.tar.gz $ sudo mv hadoop-2.6.0 /usr/local/hadoop $ update-alternatives --config java $ sudo gedit ~/.bashrc #Hadoop Variables export JAVA_HOME=/usr/lib/jvm/java-7-openjdk-amd64 export HADOOP_HOME=/usr/local/hadoop export PATH=$PATH:$HADOOP_HOME/bin export PATH=$PATH:$HADOOP_HOME/sbin export HADOOP_MAPRED_HOME=$HADOOP_HOME export HADOOP_COMMON_HOME=$HADOOP_HOME export HADOOP_HDFS_HOME=$HADOOP_HOME export YARN_HOME=$HADOOP_HOME export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib" $ source ~/.bashrc $ cd /usr/local/hadoop/etc/hadoop $ sudo vi hadoop-env.sh #The java implementation to use. export JAVA_HOME="/usr/lib/jvm/java-7-openjdk-amd64" $ sudo vi core-site.xml <configuration> <property> <name>fs.defaultFS</name> <value>hdfs://localhost:9000</value> </property> </configuration> $ sudo vi yarn-site.xml <configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value> org.apache.hadoop.mapred.ShuffleHandler</value> </property> </configuration> $ sudo cp mapred-site.xml.template mapred-site.xml $ sudo vi mapred-site.xml <configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> </configuration> $ sudo vi hdfs-site.xml <configuration> <property> <name>dfs.replication</name> <value>1</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:/usr/local/hadoop/hadoop_data/hdfs/namenode</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/usr/local/hadoop/hadoop_store/hdfs/datanode</value> </property> </configuration> $ cd $ mkdir -p /usr/local/hadoop/hadoop_data/hdfs/namenode $ mkdir -p /usr/local/hadoop/hadoop_data/hdfs/datanode $ sudo chown hadoop -R /usr/local/hadoop $ hdfs namenode -format $ start-all.sh (这个已经deprecated了。 要用start-dfs.sh 和 start-yarn.sh) 为了方便,在.bashrc中加入: alias hstart="/usr/local/hadoop/sbin/start-dfs.sh;/usr/local/hadoop/sbin/start-yarn.sh" alias hstop="/usr/local/hadoop/sbin/stop-yarn.sh;/usr/local/hadoop/sbin/stop-dfs.sh" $ jps its result should be : hadoop@master:/usr/local/hadoop$ jps 5253 NodeManager 6084 NameNode 5118 ResourceManager 4791 DataNode 4972 SecondaryNameNode 6713 Jps http://localhost:8088/ http://localhost:50070/ http://localhost:50090/ http://localhost:50075/ |
|