环境

192.168.10.150 xiangys0134-haddop01 master

192.168.10.151 xiangys0134-haddop02 slave

192.168.10.152 xiangys0134-haddop03 slave

一、配置hosts

192.168.10.150 xiangys0134-haddop01
192.168.10.151 xiangys0134-haddop02
192.168.10.152 xiangys0134-haddop03

二、安装jdk

[root@xiangys0134-haddop01 tmp]# yum install java-1.8.0-openjdk java-1.8.0-openjdk-devel

三、配置ssh

[opadm@xiangys0134-haddop02 ~]ssh-keygen -t rsa
[opadm@xiangys0134-haddop02 ~] cd ~/
[opadm@xiangys0134-haddop02 ~]tar -czvf /tmp/ssh-key.tar.gz .ssh

#将压缩包拷贝至另外两台机器
[opadm@xiangys0134-haddop02 ~] scp -P 31235 /tmp/ssh-key.tar.gz opadm@192.168.10.150:/tmp
[opadm@xiangys0134-haddop02 ~]scp -P 31235 /tmp/ssh-key.tar.gz opadm@192.168.10.152:/tmp

[opadm@xiangys0134-haddop01 tmp] tar -zxvf ssh-key.tar.gz -C ~/    #分别在其他两台机器解压
[opadm@xiangys0134-haddop03 tmp]tar -zxvf ssh-key.tar.gz -C ~/

[opadm@xiangys0134-haddop01 .ssh] vi authorized_keys
ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDGwt4dcJx7MHi9w6X2na4EcpHOtmXhFGTSQs2rD3tYrUw0cqG/Tz6jbc1lpAGK+/8pAzcQTACpweFnURrma9um/5Xj9LGXoPGzrHPJ+X197rTFrWA8pZJ93HHTAR59cEUfTGrjoOn+c41ccozMH4BiBFZNgC7dIIUTVDWnTsxQzy5R9VgHzaCNvCaf5uTgDTlWiz1IHZ6YPUPwqarZRibBYSDqowyppIzAKo6+6UHU8dRxQ4h5EZ58ueYWwLW8VKtziVtkFIjDWPkPkLbJ1J3evwYhs9QVISs4j2VSa8INrkcXGeunwEhDW9yBiZCbZYLvv0PyGwQRbEzO6DDf27ex opadm@xiangys0134-haddop02

测试:
[opadm@xiangys0134-haddop01 ~]$ ssh -p 31235 opadm@192.168.10.150

三、配置hadoop环境变量

[opadm@xiangys0134-haddop01 tmp]scp -P 31235 hadoop-2.7.7.tar.gz opadm@192.168.10.151:/tmp/
[opadm@xiangys0134-haddop01 tmp] scp -P 31235 hadoop-2.7.7.tar.gz opadm@192.168.10.152:/tmp/

[opadm@xiangys0134-haddop01 tmp]sudo tar -zxvf hadoop-2.7.7.tar.gz -C /usr/local/

添加JAVA_HOME
[opadm@xiangys0134-haddop01 tmp] vi ~/.bashrc
export HADOOP_HOME=/usr/local/hadoop-2.7.7
export JAVA_HOME=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.242.b08-0.el7_7.x86_64
export PATH=JAVA_HOME/bin:HADOOP_HOME/bin:HADOOP_HOME/sbin:PATH

[opadm@xiangys0134-haddop01 tmp]$ source ~/.bashrc

四、创建数据与日志目录

[opadm@xiangys0134-haddop01 tmp]sudo mkdir /data/hadoop/hdfs/{nn,snn,dn} -p
[opadm@xiangys0134-haddop01 tmp] sudo mkdir /var/log/hadoop/yarn -p
[opadm@xiangys0134-haddop01 tmp]sudo chown -R opadm. /data/hadoop/
[opadm@xiangys0134-haddop01 tmp] sudo chown -R opadm. /var/log/hadoop/

五、修改配置文件

应用配置文件:
core-default.xml
hdfs-default.xml
yarn-default.xml
mapred-default.xml

集群必要配置文件:

etc/hadoop/core-site.xml
etc/hadoop/hdfs-site.xml
etc/hadoop/yarn-site.xml
etc/hadoop/mapred-site.xml

[opadm@xiangys0134-haddop01 tmp]cd /usr/local/hadoop-2.7.7/etc/hadoop
[opadm@xiangys0134-haddop01 hadoop] vim core-site.xml
添加如下内容:
<configuration>
    <property>
        <name>io.file.buffer.size</name>
        <value>131072</value>
    </property>

    <property>
        <name>fs.defaultFS</name>
        <value>hdfs://xiangys0134-haddop01:8020</value>
    </property>
</configuration>

备注:
fs.defaultFS    NameNode URI
io.file.buffer.size 131072 

[opadm@xiangys0134-haddop01 hadoop]vim hdfs-site.xml
<configuration>
        <property>
            <name>dfs.namenode.name.dir</name>
            <value>file:///data/hadoop/hdfs/nn</value>
        </property>
        <property>
            <name>dfs.datanode.data.dir</name>
            <value>file:///data/hadoop/hdfs/dn</value>
        </property>
        <property>
                <name>dfs.replication</name>
                <value>2</value>
        </property>
        <property>
                <name>dfs.permissions</name>
                <value>false</value>
        </property>
        <property>
                <name>dfs.namenode.secondary.http-address</name>
                <value>xiangys0134-haddop02:50090</value>
        </property>
</configuration>

备注:在其他slave节点则不需要配置dfs.namenode.name.dir

[root@xiangys0134-haddop01 hadoop]# vim yarn-site.xml
<configuration>
        <property>
                <name>yarn.nodemanager.aux-services</name>
                <value>mapreduce_shuffle</value>
        </property>
        <property>
                <name>yarn.resourcemanager.webapp.address</name>
                <value>xiangys0134-haddop01:8088</value>
        </property>
        <property>
                <name>yarn.resourcemanager.address</name>
                <value>xiangys0134-haddop01:8032</value>
        </property>
        <property>
                <name>yarn.resourcemanager.scheduler.address</name>
                <value>xiangys0134-haddop01:8030</value>
        </property>
        <property>
                <name>yarn.resourcemanager.resource-tracker.address</name>
                <value>xiangys0134-haddop01:8031</value>
        </property>
        <property>
                <name>mapreduce.jobhistory.address</name>
                <value>xiangys0134-haddop01:10020</value>
        </property>
        <property>
                <name>yarn.nodemanager.vmem-check-enabled</name>
                <value>false</value>
        </property>
</configuration>

[opadm@xiangys0134-haddop01 hadoop] cp mapred-site.xml.template mapred-site.xml
[opadm@xiangys0134-haddop01 hadoop]$ vim mapred-site.xml
<configuration>
        <property>
                <name>mapreduce.framework.name</name>
                <value>yarn</value>
        </property>
        <property>
                <name>mapreduce.jobhistory.address</name>
                <value>xiangys0134-haddop03:10020</value>
        </property>
</configuration>

六、宣告slaves

[opadm@xiangys0134-haddop01 hadoop]$ vim slaves
xiangys0134-haddop01
xiangys0134-haddop02
xiangys0134-haddop03

七、配置集群

7.1如果集群是第一次启动,需要格式化namenode
[opadm@xiangys0134-haddop01 hadoop]hadoop namenode -format
...
20/04/18 13:44:42 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
20/04/18 13:44:42 INFO util.ExitUtil: Exiting with status 0
20/04/18 13:44:42 INFO namenode.NameNode: SHUTDOWN_MSG: 
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at xiangys0134-haddop01/192.168.10.150
************************************************************/

验证:
[opadm@xiangys0134-haddop01 hadoop] ls /data/hadoop/hdfs/nn/
current

八、启动服务

8.1配置ssh的远程连接端口
[opadm@xiangys0134-haddop01 hadoop]$ sudo vim /etc/services
ssh 22/tcp    22端口修改为31235
ssh 22/udp   22端口修改为31235
8.2启动hdfs
[opadm@xiangys0134-haddop01 hadoop]start-dfs.sh

验证:
[opadm@xiangys0134-haddop01 hadoop] start-dfs.sh 
Starting namenodes on [xiangys0134-haddop01]
xiangys0134-haddop01: starting namenode, logging to /usr/local/hadoop-2.7.7/logs/hadoop-opadm-namenode-xiangys0134-haddop01.out
xiangys0134-haddop01: starting datanode, logging to /usr/local/hadoop-2.7.7/logs/hadoop-opadm-datanode-xiangys0134-haddop01.out
xiangys0134-haddop02: starting datanode, logging to /usr/local/hadoop-2.7.7/logs/hadoop-opadm-datanode-xiangys0134-haddop02.out
xiangys0134-haddop03: starting datanode, logging to /usr/local/hadoop-2.7.7/logs/hadoop-opadm-datanode-xiangys0134-haddop03.out
Starting secondary namenodes [xiangys0134-haddop02]
xiangys0134-haddop02: starting secondarynamenode, logging to /usr/local/hadoop-2.7.7/logs/hadoop-opadm-secondarynamenode-xiangys0134-haddop02.out

查看NN与DN
[opadm@xiangys0134-haddop01 hadoop]$ jps
22098 DataNode
22402 Jps
21959 NameNode
8.3启动yarn
[opadm@xiangys0134-haddop01 hadoop]$ start-yarn.sh

九、测试上传

[opadm@xiangys0134-haddop02 hadoop]hdfs dfs -mkdir /test
[opadm@xiangys0134-haddop02 hadoop] hdfs dfs -put /etc/fstab /test/fstab

[opadm@xiangys0134-haddop02 hadoop]$ hdfs dfs -ls /test
Found 1 items
-rw-r--r-- 2 opadm supergroup 465 2020-04-18 14:16 /test/fstab

十、WEB UI

10.1hdfs的ui
http://192.168.10.150:50070
10.2yarn的ui
http://192.168.10.150:8088
10.3namesecondary的ui
http://192.168.10.151:50090
最后修改日期: 2020年4月23日

作者

留言

撰写回覆或留言

发布留言必须填写的电子邮件地址不会公开。