@zhangyy
2020-07-20T13:50:44.000000Z
字数 6372
阅读 109
大数据运维专栏
- 一:环境准备
- 二:安装Hadoop3.2.1
- 三:运行一个wordcount 测试
- 四:Hadoop 分布式构建与测试
系统:CentOS7.5x64
主机名:
cat /etc/hosts
------
192.168.11.192 fat01.flyfish.com
192.168.11.195 fat02.flyfish.com
192.168.11.197 fat03.flyfish.com
-------
jdk版本: jdk1.8.181
hadoop版本 : hadoop3.2.1
rpm -ivh jdk-8u181-linux-x64.rpm
vim /etc/profile
---
## JAVA_HOME
export JAVA_HOME=/usr/java/jdk1.8.0_181-amd64
export CLASSPATH=$JAVA_HOME/jre/lib:$JAVA_HOME/lib
export PATH=$PATH:$JAVA_HOME/bin
---
source /etc/profile
java -version
下载:
wget https://mirrors.tuna.tsinghua.edu.cn/apache/hadoop/common/hadoop-3.2.1/hadoop-3.2.1.tar.gz
解压:
tar -zxvf hadoop-3.2.1.tar.gz
mkdir -p /opt/bigdata ## 将Hadoop 安装到 /opt/bigdata 目录下面
mv hadoop-3.2.1 /opt/bigdata/hadoop
编译 core-site.xml
cd /opt/bigdata/hadoop/etc/hadoop/
vim core-site.xml
-----
<configuration>
<property>
<name>hadoop.tmp.dir</name>
<value>/opt/bigdata/hadoop/data</value>
<description>hadoop_temp</description>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://fat01.flyfish.com:8020</value>
<description>hdfs_derect</description>
</property>
</configuration>
----
vim hdfs-site.xml
---
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
<description>num</description>
<name>dfs.namenode.http-address</name>
<value>fat01.flyfish.com:50070</value>
</property>
</configuration>
---
vim mapred-site.xml
----
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>fat01.flyfish.com:19888</value>
</property>
<property>
<name>yarn.app.mapreduce.am.env</name>
<value>HADOOP_MAPRED_HOME=/opt/bigdata/hadoop</value>
</property>
<property>
<name>mapreduce.map.env</name>
<value>HADOOP_MAPRED_HOME=/opt/bigdata/hadoop</value>
</property>
<property>
<name>mapreduce.reduce.env</name>
<value>HADOOP_MAPRED_HOME=/opt/bigdata/hadoop</value>
</property>
</configuration>
-----
vim yarn-site.xml
----
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
</configuration>
-----
echo "export JAVA_HOME=/usr/java/jdk1.8.0_181-amd64" >> hadoop-env.sh
echo "export JAVA_HOME=/usr/java/jdk1.8.0_181-amd64" >> mapred-env.sh
echo "export JAVA_HOME=/usr/java/jdk1.8.0_181-amd64" >> yarn-env.sh
cd /opt/bigdata/hadoop
bin/hdfs namenode -format
bin/hdfs --daemon start namenode
bin/hdfs --daemon start datanode
浏览器访问:
http://192.168.11.192:50070
bin/yarn --daemon start resourcemanager
bin/yarn --daemon start nodemanager
bin/mapred --daemon start historyserver
cd /opt/bigdata/hadoop
bin/hdfs dfs -mkdir /input
bin/hdfs dfs -put word.txt /input
bin/yarn jar share/hadoop/mapreduce/hadoop-mapreduce-examples-3.2.1.jar wordcount /input /output1
bin/hdfs dfs -put /output1
承接上文伪分布基础上部署:
角色分配图:
cd /opt/bigdata/hadoop/etc/
vim core-site.xml
----
<configuration>
<property>
<name>hadoop.tmp.dir</name>
<value>/opt/bigdata/hadoop/data</value>
<description>hadoop_temp</description>
</property>
<property>
<name>fs.default.name</name>
<value>hdfs://fat01.flyfish.com:8020</value>
<description>hdfs_derect</description>
</property>
</configuration>
----
编辑hdfs-site.xml 文件:
<configuration>
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<property>
<name>dfs.namenode.http-address</name>
<value>fat01.flyfish.com:50070</value>
</property>
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>fat02.flyfish.com:50090</value>
</property>
</configuration>
编辑mared-site.xml
vim mared-site.xml
-----
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>fat02.flyfish.com:10020</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>fat02.flyfish.com:19888</value>
</property>
<property>
<name>yarn.app.mapreduce.am.env</name>
<value>HADOOP_MAPRED_HOME=/opt/bigdata/hadoop</value>
</property>
<property>
<name>mapreduce.map.env</name>
<value>HADOOP_MAPRED_HOME=/opt/bigdata/hadoop</value>
</property>
<property>
<name>mapreduce.reduce.env</name>
<value>HADOOP_MAPRED_HOME=/opt/bigdata/hadoop</value>
</property>
</configuration>
-----
编辑yarn-site.xml
vim yarn-site.xml
----
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>fat02.flyfish.com</value>
</property>
<property>
<name>yarn.log-aggregation-enable</name>
<value>true</value>
</property>
<property>
<name>yarn.log-aggregation.retain-seconds</name>
<value>604800</value>
</property>
</configuration>
----
编辑hadoop-env.sh 文件:
vim hadoop-env.sh
---
export JAVA_HOME=/usr/java/jdk1.8.0_181-amd64
export HADOOP_PID_DIR=/opt/bigdata/hadoop/data/tmp
export HADOOP_SECURE_DN_PID_DIR=/opt/bigdata/hadoop/data/tmp
----
编辑mapred-env.sh 文件:
vim mapred-env.sh
----
export JAVA_HOME=/usr/java/jdk1.8.0_181-amd64
export HADOOP_MAPRED_PID_DIR=/opt/bigdata/hadoop/data/tmp
----
编辑yarn-env.sh 文件:
vim yarn-env.sh
----
export JAVA_HOME=/usr/java/jdk1.8.0_181-amd64
-----
编辑works 文件
vim works
---
fat01.flyfish.com
fat02.flyfish.com
fat03.flyfish.com
---
同步所有节点
cd /opt/
tar –zcvf bigdata.tar.gz bigdata
scp bigdata.tar.gz root@192.168.11.195:/opt/
scp bigdata.tar.gz root@192.168.11.197:/opt/
然后到每个节点上面去 解压这个 bigdata.tar.gz
删掉原有节点上面格式数据
cd /opt/bigdata/hadoop/data
rm -rf *
fat01.flyfish.com 主机上执行:
cd /opt/bigdata/hadoop/bin
./hdfs namenode –format
启动之前先决定 启动用户是是是什么这边默认是root
cd /opt/bigdata/hadoop/sbin
对于start-dfs.sh和stop-dfs.sh文件,添加下列参数:
----
#!/usr/bin/env bash
HDFS_DATANODE_USER=root
HADOOP_SECURE_DN_USER=hdfs
HDFS_NAMENODE_USER=root
HDFS_SECONDARYNAMENODE_USER=root
----
vim start-yarn.sh
---
YARN_RESOURCEMANAGER_USER=root
HADOOP_SECURE_DN_USER=yarn
YARN_NODEMANAGER_USER=root
----
启动hdfs
fat01.flyfish.com 主机上执行:
cd hadoop/sbin/
./start-dfs.sh
打开 hdfs web
http://192.168.11.192:50070
启动yarn
fat02.flyfish.com
cd hadoop/sbin/
start-yarn.sh
启动historyserver
mapred --daemon start jobhistoryserver
打开yarn的界面:
http://192.168.11.195:8088
打开jobhistoryserver 的页面
http://192.168.11.195:19888
启动角色对比
fat01.flyfish.com
jps
fat02.flyfish.com
jps
fat03.flyfish.com
jps
至此 Hadoop 分布式环境搭建完成
cd /opt/bigdata/hadoop
bin/hdfs dfs -mkdir /input
bin/hdfs dfs -put word.txt /input
bin/yarn jar share/hadoop/mapreduce/hadoop-mapreduce-examples-3.2.1.jar wordcount /input /output1
hdfs dfs -ls /input
hdfs dfs -cat /output1/part-r-00000