下载链接:
http://archive.apache.org/dist/hadoop/common/hadoop-2.7.5/hadoop-2.7.5.tar.gz
上传至/opt/software
解压命令
cd /opt/software tar -zxvf hadoop-2.7.5.tar.gz -C ../module/第二步:修改配置文件 修改core-site.xml(建议使用notepad++进行修改)
末尾添加如下内容:
修改hdfs-site.xmlfs.default.name hdfs://192.168.211.102:8020 hadoop.tmp.dir /opt/module/hadoop-2.7.5/hadoopDatas/tempDatas io.file.buffer.size 4096 fs.trash.interval 10080
修改hadoop-env.shdfs.namenode.secondary.http-address hadoop102:50090 dfs.namenode.http-address hadoop102:50070 dfs.namenode.name.dir file:///opt/module/hadoop-2.7.5/hadoopDatas/namenodeDatas,file:///opt/module/hadoop-2.7.5/hadoopDatas/namenodeDatas2 dfs.datanode.data.dir file:///opt/module/hadoop-2.7.5/hadoopDatas/datanodeDatas,file:///opt/module/hadoop-2.7.5/hadoopDatas/datanodeDatas2 dfs.namenode.edits.dir file:///opt/module/hadoop-2.7.5/hadoopDatas/nn/edits dfs.namenode.checkpoint.dir file:///opt/module/hadoop-2.7.5/hadoopDatas/snn/name dfs.namenode.checkpoint.edits.dir file:///opt/module/hadoop-2.7.5/hadoopDatas/dfs/snn/edits dfs.replication 3 dfs.permissions false dfs.blocksize 134217728
修改JAVA_HOME的路径
export JAVA_HOME=/opt/module/jdk1.8.0_251修改mapred-site.xml(将mapred-site.xml.template改为mapred-site.xml)
修改yarn-site.xmlmapreduce.framework.name yarn mapreduce.job.ubertask.enable true mapreduce.jobhistory.address hadoop102:10020 mapreduce.jobhistory.webapp.address hadoop102:19888
修改slavesyarn.resourcemanager.hostname hadoop102 yarn.nodemanager.aux-services mapreduce_shuffle yarn.log-aggregation-enable true yarn.log-aggregation.retain-seconds 604800
替换内容:
hadoop102第三步:启动集群 首先执行以下命令
cd /opt/module/hadoop-2.7.5 mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/tempDatas mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/namenodeDatas mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/namenodeDatas2 mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/datanodeDatas mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/datanodeDatas2 mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/nn/edits mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/snn/name mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/dfs/snn/edits准备启动
首次启动 HDFS 时,必须对其进行格式化操作。(只在初始化集群的时候执行一次 )
cd /opt/module/hadoop-2.7.5/ bin/hdfs namenode -format进行启动
sbin/start-dfs.sh sbin/start-yarn.sh
其他:
查看日志:cd /opt/module/hadoop-2.7.5/logs启动历史jobhistory
sbin/mr-jobhistory-daemon.sh start historyserver
http://192.168.211.102:50070/
端口50070:默认查看hdfs集群
http://192.168.211.102:8088/
端口8088:默认看yarn集群
http://192.168.211.102:19888/
端口19888默认查看历史
伪分布式集群搭建:第一台机器执行以下命令
cd /opt/module/hadoop-2.7.5 sbin/stop-dfs.sh sbin/stop-yarn.sh sbin/mr-jobhistory-daemon.sh stop historyserver
rm -rf /opt/module/hadoop-2.7.5/hadoopDatas
mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/tempDatas mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/namenodeDatas mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/namenodeDatas2 mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/datanodeDatas mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/datanodeDatas2 mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/nn/edits mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/snn/name mkdir -p /opt/module/hadoop-2.7.5/hadoopDatas/dfs/snn/edits
第一台机器执行以下命令
cd /opt/module/hadoop-2.7.5/etc/hadoop vim slaves node01 node02 node03
第一台机器执行以下命令
cd /opt/module/ scp -r hadoop-2.7.5 node02:$PWD scp -r hadoop-2.7.5 node03:$PWD
启动集群
方式一:第一台机器执行以下命令
cd /opt/module/hadoop-2.7.5 bin/hdfs namenode -format sbin/start-dfs.sh sbin/start-yarn.sh sbin/mr-jobhistory-daemon.sh start historyserver
方法二:hadoop102服务器上面的服务,挨个启动,但是hadoop103和hadoop104只需要启动datanode和nodemanager
namenode:sbin/hadoop-daemon.sh start namenode datanode: sbin/hadoop-daemon.sh start datanode secondarynamenode: sbin/hadoop-daemon.sh start secondarynamenode resourceManager:sbin/yarn-daemon.sh start resourcemanager nodeManager: sbin/yarn-daemon.sh start nodemanager jobhistory:sbin/mr-jobhistory-daemon.sh start historyserver



