Skip to content
字数
481 字
阅读时间
3 分钟

前置步骤

01.实验1:安装CentOS7

安装配置Hadoop

创建资源目录

sudo mkdir /opt/softwares
sudo mkdir /opt/modules
sudo chown -R hadoop:hadoop /opt
sudo chmod -R 755 /opt
cd /opt/softwares

上传资源

sudo rz

![[Pasted image 20251117144355.png]] 解压资源

sudo tar -zxvf /opt/softwares/hadoop-2.7.2.tar.gz -C /opt/modules/
sudo tar -zxvf /opt/softwares/jdk-8u25-linux-x64.tar.gz -C /opt/modules/

添加环境变量

卸载原本的Java环境

查看已安装的Java

rpm -qa | grep -i java

常见的输出可能是

java-1.8.0-openjdk-1.8.0.xxx
java-1.8.0-openjdk-headless-1.8.0.xxx

根据查找到的包名逐个卸载

sudo yum remove -y java-1.8.0-openjdk
sudo yum remove -y java-1.8.0-openjdk-headless

验证是否卸载成功

java -version

如果提示如下便是卸载成功

bash: java: command not found

添加环境变量

创建hadoop-env.sh文件

sudo vi /etc/profile.d/hadoop-env.sh

写入以下内容

export JAVA_HOME=/opt/modules/jdk1.8.0_25
export HADOOP_HOME=/opt/modules/hadoop-2.7.2
export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin

修改环境变量文件

sudo chmod +x /etc/profile.d/hadoop-env.sh

立即生效

source /etc/profile.d/hadoop-env.sh

修改Hadoop配置文件

将Java环境路径注入Hadoop安装脚本

sudo echo "export JAVA_HOME=/opt/modules/jdk1.8.0_25/" >> /opt/modules/hadoop-2.7.2/etc/hadoop/hadoop-env.sh
sudo echo "export JAVA_HOME=/opt/modules/jdk1.8.0_25/" >> /opt/modules/hadoop-2.7.2/etc/hadoop/mapred-env.sh
sudo echo "export JAVA_HOME=/opt/modules/jdk1.8.0_25/" >> /opt/modules/hadoop-2.7.2/etc/hadoop/yarn-env.sh

core-site.xml

vi /opt/modules/hadoop-2.7.2/etc/hadoop/core-site.xml

配置文件如下:

<configuration>
   <property>
        <name>fs.defaultFS</name>
        <value>hdfs://gyy1:9000</value>
    </property>
    <property>
        <name>hadoop.tmp.dir</name>
        <value>file:/opt/modules/hadoop-2.7.2/tmp</value>
    </property>
</configuration>

hdfs-site.xml

vi /opt/modules/hadoop-2.7.2/etc/hadoop/hdfs-site.xml

配置文件如下:

<configuration>
    <property>
        <name>dfs.replication</name>
        <value>3</value>
    </property>
    <property>
        <name>dfs.namenode.name.dir</name>
        <value>file:/opt/modules/hadoop-2.7.2/tmp/dfs/name</value>
    </property>
    <property>
        <name>dfs.datanode.data.dir</name>
        <value>file:/opt/modules/hadoop-2.7.2/tmp/dfs/data</value>
    </property>
</configuration>

mapred-site.xml

mv /opt/modules/hadoop-2.7.2/etc/hadoop/mapred-site.xml.template /opt/modules/hadoop-2.7.2/etc/hadoop/mapred-site.xml
vi /opt/modules/hadoop-2.7.2/etc/hadoop/mapred-site.xml

配置文件如下

<configuration>
    <property>
        <name>mapreduce.framework.name</name>
        <value>yarn</value>
    </property>
</configuration>

yarn-site.xml

vi /opt/modules/hadoop-2.7.2/etc/hadoop/yarn-site.xml

配置文件如下

<configuration>
    <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce_shuffle</value>
    </property>
    <property>
        <name>yarn.resourcemanager.hostname</name>
        <value>gyy1</value>
    </property>
</configuration>

slaves

vi /opt/modules/hadoop-2.7.2/etc/hadoop/slaves

配置文件如下

gyy1
gyy2
gyy3

在配置完成Hadoop之后,克隆两台虚拟机并03.配置SSH互信

格式化并启动集群

在主机1上运行Hadoop格式化集群

hadoop namenode -format

启动Hadoop集群

start-all.sh

警告

在关闭主机前,需要在主机1上运行stop-all.sh关闭所有集群,否则集群下次运行会报错

贡献者

The avatar of contributor named as lishu620 lishu620

文件历史

撰写