设为首页 收藏本站
查看: 1167|回复: 0

[经验分享] Hadoop伪分布模式安装

[复制链接]
累计签到:1 天
连续签到:1 天
发表于 2015-7-12 10:10:54 | 显示全部楼层 |阅读模式
  一、本文说明:
  本次测试在一台虚拟机系统上进行伪分布式搭建。Hadoop伪分布式模式是在单机上模拟Hadoop分布式,单机上的分布式并不是真正的伪分布式,而是使用线程模拟分布式。Hadoop本身是无法区分伪分布式和分布式的,两种配置也很相似。唯一不同的地方是伪分布式是在单机器上配置,数据节点和名字节点均是一个机器。
      环境说明:
  操作系统:red hat 5.4 x86
  hadoop版本:hadoop-0.20.2
  JDK版本:jdk1.7
  二、JDK安装及Java环境变量的配置



----首先把压缩包解压出来----
1 [iyunv@localhost ~]# tar -zxvf jdk-7u9-linux-i586.tar.gz
2
----修改目录名----
3 [iyunv@localhost ~]# mv jdk1.7.0_09 /jdk1.7
4
----在/etc/profile文件中添加下面几行----
5 [iyunv@localhost ~]# vi /etc/profile
6
7 export JAVA_HOME=/jdk1.7
8 export CLASSPATH=.:$JAVA_HOME/lib:$JAVA_HOME/jre/lib
9 export PATH=$JAVA_HOME/bin:$PATH
10
----验证是否已经成功安装jdk1.7----
11 [iyunv@localhost ~]# java -version
12 java version "1.7.0_09"
13 Java(TM) SE Runtime Environment (build 1.7.0_09-b05)
14 Java HotSpot(TM) Client VM (build 23.5-b02, mixed mode)
  三、SSH无密码验证设置
  Hadoop需要使用SSH协议,namemode将使用SSH协议启动namenode和datanode进程,伪分布式模式数据节点和名称节点均是本身,必须配置SSH localhost无密码验证。



1 [iyunv@localhost bin]# ssh-keygen -t rsa
2 Generating public/private rsa key pair.
3 Enter file in which to save the key (/root/.ssh/id_rsa):
4 /root/.ssh/id_rsa already exists.
5 Overwrite (y/n)? y
6 Enter passphrase (empty for no passphrase):
7 Enter same passphrase again:
8 Your identification has been saved in /root/.ssh/id_rsa.
9 Your public key has been saved in /root/.ssh/id_rsa.pub.
10 The key fingerprint is:
11 2f:eb:6c:c5:c5:3b:0b:26:a4:7f:0f:7a:d7:3b:5e:e5 root@localhost.localdomain
12 You have mail in /var/spool/mail/root
13 [iyunv@localhost bin]# cd
14 [iyunv@localhost ~]# cd .ssh
15 [iyunv@localhost .ssh]# ls
16 authorized_keys  id_rsa  id_rsa.pub  known_hosts
17 [iyunv@localhost .ssh]# cat id_rsa.pub > authorized_keys
18 [iyunv@localhost .ssh]# ssh 192.168.20.150
19 Last login: Fri Apr 26 11:07:21 2013 from 192.168.20.103
20 [iyunv@localhost ~]# ssh localhost
21 Last login: Fri Apr 26 12:45:43 2013 from master
  四、Hadoop配置
  4.1、下载hadoop-0.20.2.tar.gz,将其解压缩到/123目录下



1 [iyunv@localhost 123]# tar -zxvf hadoop-0.20.2.tar.gz
  4.2、进入/123/hadoop-0.20.2/conf,配置Hadoop配置文件
    4.3、配置hadoop-env.sh文件



1 [iyunv@localhost conf]# pwd
2 /123/hadoop-0.20.2/conf
3 [iyunv@localhost conf]# vi hadoop-env.sh
4
5 # Set Hadoop-specific environment variables here.
6
7 # The only required environment variable is JAVA_HOME.  All others are
8 # optional.  When running a distributed configuration it is best to
9 # set JAVA_HOME in this file, so that it is correctly defined on
10 # remote nodes.
11
12 # The java implementation to use.  Required.
----下面这句是添加进去的----
13  export JAVA_HOME=/jdk1.7
14
15 # Extra Java CLASSPATH elements.  Optional.
16 # export HADOOP_CLASSPATH=
  4.4、配置core-site.xml



1 [iyunv@localhost conf]# cat core-site.xml
2
3
4
5
6
7
8
9 fs.default.name
10 hdfs://192.168.20.150:9000
11
12
13 hadoop.tmp.dir
14 /123/hadooptmp
15
16
  4.6、配置hdfs-site.xml



1 [iyunv@localhost conf]# cat hdfs-site.xml
2
3
4
5
6
7
8
9 dfs.name.dir
10 /123/hdfs/name
11
12
13 dfs.data.dir
14 /123/hdfs/data
15
16
17 dfs.replication
18 1
19
20
  4.7、配置mapred-site.xml



1 [iyunv@localhost conf]# cat mapred-site.xml
2
3
4
5
6
7
8
9 mapred.job.tracker
10 localhost:9001
11
12
  4.8、配置masters文件和slaves文件



1 [iyunv@localhost conf]# cat masters
2 192.168.20.150
3 [iyunv@localhost conf]# cat slaves
4 192.168.20.150
  注:因为在伪分布式模式下,作为master的namenode与作为slave的datanode是同一台服务器,所以配置文件中的ip是一样的
    4.9、编辑主机名



1 [iyunv@localhost conf]# cat /etc/hosts
2 # Do not remove the following line, or various programs
3 # that require network functionality will fail.
4 127.0.0.1        localhost.localdomain localhost
5 ::1        localhost6.localdomain6 localhost6
6 192.168.20.150 master
7 192.168.20.150 slave
  4.10、创建上面被编辑文件中的目录



1 [iyunv@localhost conf]# mkdir -p /123/hadooptmp
2
3 [iyunv@localhost conf]# mkdir -p /123/hdfs/name
4
5 [iyunv@localhost conf]# mkdir -p /123/hdfs/data
  五、启动Hadoop并进行验证
  5.1、对namenode进行格式化



1 [iyunv@localhost bin]# ./hadoop namenode -format
2 13/04/26 11:08:05 INFO namenode.NameNode: STARTUP_MSG:
3 /************************************************************
4 STARTUP_MSG: Starting NameNode
5 STARTUP_MSG:   host = localhost.localdomain/127.0.0.1
6 STARTUP_MSG:   args = [-format]
7 STARTUP_MSG:   version = 0.20.2
8 STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
9 ************************************************************/
10 Re-format filesystem in /123/hdfs/name ? (Y or N) Y
11 13/04/26 11:08:09 INFO namenode.FSNamesystem: fsOwner=root,root,bin,daemon,sys,adm,disk,wheel
12 13/04/26 11:08:09 INFO namenode.FSNamesystem: supergroup=supergroup
13 13/04/26 11:08:09 INFO namenode.FSNamesystem: isPermissionEnabled=true
14 13/04/26 11:08:09 INFO common.Storage: Image file of size 94 saved in 0 seconds.
15 13/04/26 11:08:09 INFO common.Storage: Storage directory /123/hdfs/name has been successfully formatted.
16 13/04/26 11:08:09 INFO namenode.NameNode: SHUTDOWN_MSG:
17 /************************************************************
18 SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
19 ************************************************************/
  5.2、启动hadoop所有进程



1 [iyunv@localhost bin]# ./start-all.sh
2 starting namenode, logging to /123/hadoop-0.20.2/bin/../logs/hadoop-root-namenode-localhost.localdomain.out
3 192.168.20.150: starting datanode, logging to /123/hadoop-0.20.2/bin/../logs/hadoop-root-datanode-localhost.localdomain.out
4 192.168.20.150: starting secondarynamenode, logging to /123/hadoop-0.20.2/bin/../logs/hadoop-root-secondarynamenode-localhost.localdomain.out
5 starting jobtracker, logging to /123/hadoop-0.20.2/bin/../logs/hadoop-root-jobtracker-localhost.localdomain.out
6 192.168.20.150: starting tasktracker, logging to /123/hadoop-0.20.2/bin/../logs/hadoop-root-tasktracker-localhost.localdomain.out
  5.2、使用jps命令查看hadoop进程是否启动完全。



1 [iyunv@localhost bin]# jps
2 15219 JobTracker
3 15156 SecondaryNameNode
4 15495 Jps
5 15326 TaskTracker
6 15044 DataNode
7 14959 NameNode
  5.3、查看集群状态:



1 [iyunv@localhost bin]# ./hadoop dfsamin -report
2 Error: Could not find or load main class dfsamin
3 [iyunv@localhost bin]# ./hadoop dfsadmin -report
4 Configured Capacity: 19751522304 (18.4 GB)
5 Present Capacity: 14953619456 (13.93 GB)
6 DFS Remaining: 14953582592 (13.93 GB)
7 DFS Used: 36864 (36 KB)
8 DFS Used%: 0%
9 Under replicated blocks: 0
10 Blocks with corrupt replicas: 0
11 Missing blocks: 0
12
13 -------------------------------------------------
14 Datanodes available: 1 (1 total, 0 dead)
15
16 Name: 192.168.20.150:50010
17 Decommission Status : Normal
18 Configured Capacity: 19751522304 (18.4 GB)
19 DFS Used: 36864 (36 KB)
20 Non DFS Used: 4797902848 (4.47 GB)
21 DFS Remaining: 14953582592(13.93 GB)
22 DFS Used%: 0%
23 DFS Remaining%: 75.71%
24 Last contact: Fri Apr 26 13:06:15 CST 2013
  

运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-85709-1-1.html 上篇帖子: Hadoop中Writable类之四 下篇帖子: Meet Hadoop
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表