Apache Hadoop软件库是一个框架,开发者只需使用简单的编程模型在大量计算机(配置不高)上对大型数据集进行分布式处理。
YingJoy
Hadoop有两种安装模式
本文采用完全分布式进行安装。
master: 123.206.111.142 hadoop-master
slave : 115.159.221.61 hadoop-slave-1
slave : 123.206.217.178 hadoop-slave-2
执行下面代码
cd /opt/
wget --no-cookies --no-check-certificate --header "Cookie: gpw_e24=http%3A%2F%2Fwww.oracle.com%2F; oraclelicense=accept-securebackup-cookie" "http://download.oracle.com/otn-pub/java/jdk/7u79-b15/jdk-7u79-linux-x64.tar.gz"
tar xzf jdk-7u79-linux-x64.tar.gz
用alternatives安装Java
cd /opt/jdk1.7.0_79/
alternatives --install /usr/bin/java java /opt/jdk1.7.0_79/bin/java 2
alternatives --config java
你会看到这个
There are 3 programs which provide 'java'.
Selection Command
-----------------------------------------------
* 1 /opt/jdk1.7.0_60/bin/java
+ 2 /opt/jdk1.7.0_72/bin/java
3 /opt/jdk1.7.0_79/bin/java
Enter to keep the current selection[+], or type selection number: 3 [Press Enter]
这里选择3
现在设置javac和jar命令
alternatives --install /usr/bin/jar jar /opt/jdk1.7.0_79/bin/jar 2
alternatives --install /usr/bin/javac javac /opt/jdk1.7.0_79/bin/javac 2
alternatives --set jar /opt/jdk1.7.0_79/bin/jar
alternatives --set javac /opt/jdk1.7.0_79/bin/javac
查看Java是否安装成功
java -version
java version "1.7.0_79"
Java(TM) SE Runtime Environment (build 1.7.0_79-b15)
Java HotSpot(TM) 64-Bit Server VM (build 24.79-b02, mixed mode)
配置环境变量
vi ~/.bashrc
在文件尾部加上下面的代码
export JAVA_HOME=/opt/jdk1.7.0_79
export JRE_HOME=/opt/jdk1.7.0_79/jre
export PATH=$PATH:/opt/jdk1.7.0_79/bin:/opt/jdk1.7.0_79/jre/bin
执行下面代码
useradd hadoop
passwd hadoop
会看到下面
Changing password for user hadoop.
New password:
Retype new password:
passwd: all authentication tokens updated successfully.
执行下面代码
vi /etc/hosts
在文件下面增加下面三句
master的ip hadoop-master
slave1的ip hadoop-slave-1
slave2的ip hadoop-slave-2
仅在Hadoop用户下免密登陆,我听到很多人说配置后仍需密码,那是因为他们使用了root用户来连接。
su - hadoop
ssh-keygen -t rsa
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop@hadoop-master
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop@hadoop-slave-1
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop@hadoop-slave-2
chmod 0600 ~/.ssh/authorized_keys
exit
这里使用Hadoop的一个稳定点的版本2.7.3
mkdir /opt/hadoop
cd /opt/hadoop/
wget http://apache.mesi.com.ar/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz
tar -xzf hadoop-2.7.3.tar.gz
mv hadoop-2.7.3 hadoop
chown -R hadoop /opt/hadoop
cd /opt/hadoop/hadoop/
1.修改core-site.xml
在<configuration>标签中间添加下面内容
<property>
<name>fs.default.name</name>
<value>hdfs://hadoop-master:9000/</value>
</property>
<property>
<name>dfs.permissions</name>
<value>false</value>
</property>
2.修改hdfs-site.xml
在<configuration>标签中间添加下面内容
<property>
<name>dfs.data.dir</name>
<value>/opt/hadoop/hadoop/dfs/name/data</value>
<final>true</final>
</property>
<property>
<name>dfs.name.dir</name>
<value>/opt/hadoop/hadoop/dfs/name</value>
<final>true</final>
</property>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
3.修改 mapred-site.xml
首先先运行下面命令
cp mapred-site.xml.template mapred-site.xml
然后在<configuration>标签中间添加下面内容
<property>
<name>mapred.job.tracker</name>
<value>hadoop-master:9001</value>
</property>
4.修改hadoop-env.sh
export JAVA_HOME=/opt/jdk1.7.0_79
export HADOOP_OPTS=-Djava.net.preferIPv4Stack=true
export HADOOP_CONF_DIR=/opt/hadoop/hadoop/conf
你的Java路径
利用SSH复制,速度有点慢,喝杯茶等一下
su hadoop
cd /opt/hadoop
scp -r hadoop hadoop-slave-1:/opt/hadoop
scp -r hadoop hadoop-slave-2:/opt/hadoop
先进入Hadoop路径
su hadoop
cd /opt/hadoop/hadoop
1.修改conf/masters
hadoop-master
2.修改conf/slaves
hadoop-slave-1
hadoop-slave-2
su hadoop
cd /opt/hadoop/hadoop
bin/hadoop namenode -format
使用下面命令,直接启动所有服务
bin/start-all.sh
相关推荐
原创声明:本文系作者授权腾讯云开发者社区发表,未经许可,不得转载。
如有侵权,请联系 cloudcommunity@tencent.com 删除。
原创声明:本文系作者授权腾讯云开发者社区发表,未经许可,不得转载。
如有侵权,请联系 cloudcommunity@tencent.com 删除。