|
hadoop的安装
1. 安装redhat 1) 安装2台虚拟机,一个为master,一个slave 2) 修改启动级别为3,默认控制台登陆 设置hosts:修改/etc/hosts 192.168.1.103 master 192.168.1.102 slave 2. master无密码登陆slave 2.1 Master 、slave生成公钥、密钥: 执行: ssh-keygen 回车到底 2.2 将master上公钥复制到slave上: slave上.ssh目录下建立 authorized_keys文件,复制master上id_rsa.pub文件内容到authorized_keys 3. 安装java 3.1. 将java安装包上传到master、slave上 redhat安装好后,默认会安装sshd服务,安装。安装secureFx,建立会话 上传文件到master、slave上,安装jdk 使用secureCRT远程连接到master: 上传jdk到master: rz –y (如果这种方式长传出错,则使用secureFX进行文件上传) 3.2 解压: tar –zxvf jdk-7u25-linux-i586.tar.gz 复制jdk到slave: scp jdk-7u25-linux-i586.tar.gz slave:/root/ 解压之后,执行java --version可能会遇到这个错误: Error: dl failure on line 864
Error: failed /usr/jdk/jdk1.7.0_05/jre/lib/i386/client/libjvm.so, because /usr/jdk/jdk1.7.0_05/jre/lib/i386/client/libjvm.so: cannot restore segment prot after reloc: Permission denied 执行下面语句可解决: setenforce 0 3.3 设置环境变量:JAVA_HOME、PATH vi /etc/profile JAVA_HOME=/usr/local/java PATH=$JAVA_HOME/bin PATH source /etc/profile 3.4 验证java安装正确: 执行:java 显示: 4. 配置hadoop(1.2版本): 4.1.配置hadoop使用的JAVA_HOME,修改conf/hadoop-env.sh JAVA_HOME=$JAVA_HOME export JAVA_HOME 4.2.配置hdfs:修改conf/core-site.xml <configuration> <property> <name>hadoop.tmp.dir</name> <value>/usr/local/hadoop/tmp/hadoop-${user.name}</value> <description>A base for other temporarydirectories.</description> </property> <property> <name>fs.default.name</name> <value>hdfs://master:9000</value> <description>The name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem. </description> </property> </configuration> 4.3.配置hdfs副本数:修改hdfs-site.xml <configuration> <property> <name>dfs.replication</name> <value>2</value> <description>Default block replication. The actual number of replications can be specified when the file iscreated. The default is used if replication is not specified in create time. </description> </property> </configuration> 4.4.配置jobtracker地址:修改 mapred-site.xml <configuration> <property> <name>mapred.job.tracker</name> <value>master:9001</value> <description>The host and port that the MapReduce job trackerruns at. If "local", then jobs are run in-process as a singlemap and reduce task. </description> </property> </configuration> 4.5.配置secondary namenode:修改 masters Master 4.6.配置slave:修改slaves master slave 4.7.配置HADOOP_HOME Vi /ect/profile JAVA_HOME=/usr/local/java PATH=$JAVA_HOME/bin:$PATH HADOOP_HOME=/usr/local/hadoop export JAVA_HOME PATH HADOOP_HOME 4.8格式化文件系统 hadoop namenode –format 4.9.启动集群 Master上执行:jps,看到如下图所示: ,master上运行了NameNode、DataNode、JobTracker、TaskTracker、SecondaryNameNode Slave上执行:jps,看到如下图所示: 4.10.查看hadoop的jobtracker地址: http://192.168.1.103:50030/jobtracker.jsp 如果不能直接打开,可能是防火墙的问题,需要在master上开启50030端口 Vi /etc/sysconfig/iptables 添加: -A RH-Firewall-1-INPUT -m state --state NEW -m tcp -p tcp --dport 50030 -j ACCEPT 是防火墙生效: iptables-restore < /etc/sysconfig/iptables 4.11.查看hdfs的web地址:如果直接查看不了,确定是否是防火墙未开启,才处理方法同上。 http://192.168.1.103:50070/dfshealth.jsp Hadoop安装成功。 注:本次hadoop的安装用例,没有涉及到map、reduce任务槽的设置、hadoop调度器的设置、hdfs容量的配置、hdfs块的配置等,只是作为测试使用,用于生产,还需上面及其他参数进行设计,对hadoop集群进行调优。
|