hadoop 完全分布式环境搭建

1、虚拟机角色分配:

192.168.44.184 hadoop02 NameNode/DataNode ResourceManager/NodeManager
192.168.44.185 hadoop03 DataNode NodeManager
192.168.44.186 hadoop04 DataNode NodeManager

每台机器配置好hostname,hosts:/etc/hostname 中配置好hostname

/etc/hosts 中配置好hosts:

192.168.44.184 hadoop02
192.168.44.185 hadoop03
192.168.44.186 hadoop04

2、sskey免登录设置

1)ssh-keygen -t rsa (会在.ssh目录生成公钥和私钥,每台机器都要生成)

2)以hadoop02为主,将公钥copy到各个机器

ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop02
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop03
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop04

3、jdk的安装(在hadoop02上配置,另外两台机器稍后复制过去)

1) 直接下载jdk包放到安装目录即可(jdk-8u131-linux-x64.tar.gz解压到安装目录  tar zxvf ./jdk-8u131-linux-x64.tar.gz -C /home/hadoop/app/)

2) 配置环境变量 直接在 .bash_profile配置

export JAVA_HOME=/home/hadoop/app/jdk1.8.0_131
export PATH=$JAVA_HOME/bin:$PATH

4.下载安装hadoop-2.6.0-cdh5.7.0(在hadoop02上配置,另外两台机器稍后复制过去)

  cdh下载地址:(http://archive.cloudera.com/cdh5/cdh/5/)

1)直接下载hadoop-2.6.0-cdh5.7.0.tar.gz包放到安装目录即可

2)配置环境变量

 export HADOOP_HOME=/home/hadoop/app/hadoop-2.6.0-cdh5.7.0
export PATH=$HADOOP_HOME/bin:$JAVA_HOME/bin:$PATH

5、配置修改hadoop相关配置文件(在hadoop02上配置,另外两台机器稍后复制过去)

1)hadoop_home/etc/hadoop/hadoop-env.sh 配置jdk

export JAVA_HOME=/home/hadoop/app/jdk1.8.0_131

2)etc/hadoop/core-site.xml

<property>
<name>fs.defaultFS</name>
<value>hdfs://hadoop02:8020</value>
</property> <property>
<name>hadoop.tmp.dir</name>
<value>/home/hadoop/app/tmp</value>
</property>

3)etc/hadoop/hdfs-site.xml  设置副本系数和name、data的存放路径

<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/home/hadoop/app/tmp/dfs/name</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/home/hadoop/app/tmp/dfs/data</value>
</property>

4)etc/hadoop/yarn-site.xml

<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>hadoop02</value>
</property>

5)etc/hadoop/mapred-site.xml

<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>

6)/etc/hadoop/slaves(标明从节点)

hadoop02
hadoop03
hadoop04

6、分发安装包到 hadoop03、hadoop03节点上

scp -r ~/app hadoop@hadoop03:~/
scp -r ~/app hadoop@hadoop04:~/
scp ~/.bash_profile hadoop@hadoop03:~/
scp ~/.bash_profile hadoop@hadoop04:~/

在hadoop03,hadoop04上分别source下.bash_profile文件

7、NameNode格式化:只要在hadoop02上格式化

bin/hdfs namenode -format

8、启动hadoop集群。在NameNode启动

sbin/start-all.sh

启动日志:

[hadoop@hadoop02 hadoop-2.6.0-cdh5.7.0]$ ./sbin/start-all.sh
This script is Deprecated. Instead use start-dfs.sh and start-yarn.sh
18/08/19 23:20:23 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Starting namenodes on [hadoop02]
hadoop02: starting namenode, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/hadoop-hadoop-namenode-hadoop02.out
hadoop02: starting datanode, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/hadoop-hadoop-datanode-hadoop02.out
hadoop04: starting datanode, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/hadoop-hadoop-datanode-hadoop04.out
hadoop03: starting datanode, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/hadoop-hadoop-datanode-hadoop03.out
Starting secondary namenodes [0.0.0.0]
The authenticity of host '0.0.0.0 (0.0.0.0)' can't be established.
ECDSA key fingerprint is 78:05:94:d6:0b:e2:78:7d:93:3c:90:d0:ab:7c:ed:e2.
Are you sure you want to continue connecting (yes/no)? yes
0.0.0.0: Warning: Permanently added '0.0.0.0' (ECDSA) to the list of known hosts.
0.0.0.0: starting secondarynamenode, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/hadoop-hadoop-secondarynamenode-hadoop02.out
18/08/19 23:21:13 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
starting yarn daemons
starting resourcemanager, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/yarn-hadoop-resourcemanager-hadoop02.out
hadoop03: starting nodemanager, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/yarn-hadoop-nodemanager-hadoop03.out
hadoop04: starting nodemanager, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/yarn-hadoop-nodemanager-hadoop04.out
hadoop02: starting nodemanager, logging to /home/hadoop/app/hadoop-2.6.0-cdh5.7.0/logs/yarn-hadoop-nodemanager-hadoop02.out

9、验证

1)jps验证:

[hadoop@hadoop02 hadoop-2.6.0-cdh5.7.0]$ jps
3862 SecondaryNameNode
4086 NodeManager
3591 NameNode
3991 ResourceManager
3678 DataNode
4399 Jps
[hadoop@hadoop03 ~]$ jps
3460 NodeManager
3576 Jps
3372 DataNode
[hadoop@hadoop04 ~]$ jps
3284 NodeManager
3400 Jps
3196 DataNode

2)web页面验证:http://hadoop02:50070 (hdfs)

http://hadoop02:8088 (yarn)

10、停止

sbin/stop-all.sh

最新文章

  1. Android课程---添加黑名单的练习(课堂讲解)
  2. CodeSoft随笔 批量连续打印,变量打印,codesoft条码
  3. linux笔记:RPM软件包管理-rpm命令管理
  4. SQL SERVER数据库索引、外键查找
  5. linux 源码编译(转)
  6. [GeekBand] 探讨C++新标准之新语法——C++ 11~14
  7. HDU2686-Matrix &amp; HDU3376-Matrix Again(费用流)
  8. Dublin Core
  9. Web Service单元测试工具实例介绍之SoapUI
  10. ansj人名识别
  11. Vue history模式支持ie9
  12. 【noip 2016】普及组
  13. 事务 c#
  14. ora-12705解决方法
  15. 【转】PowerDesigner删除外键关系,而不删除外键列
  16. 【驱动】网卡驱动&#183;linux内核网络分层结构
  17. win7系统损坏无法进入系统(dsark.sys文件损坏)(未测试过)
  18. loadrunner在win10破解提示:Cannot save the license information because acceses to the registry is denied的解决办法
  19. java面试题8
  20. Ubuntu 16 Java Develop环境快速搭建

热门文章

  1. HttpServer:一款Windows平台下基于IOCP模型的高并发轻量级web服务器
  2. spring——自动装配【非常详细】
  3. C#开发之基于NPOI的操作Excel开发体验
  4. 排查bug:竟然是同事把Redis用成这鬼样子,坑了我
  5. docker0-常用命令-持续更新
  6. oracle中如何处理null
  7. noip2015 总结
  8. Linux命令大全之查看登陆用户信息
  9. Python语言规范之Pylint的使用
  10. (Vue中)cehart在同一个dom上画图图切换时饼图有折线图的坐标系