设为首页 收藏本站
查看: 375|回复: 0

[经验分享] 开源云计算技术系列(七)Cloudera (hadoop 0.20)

[复制链接]

尚未签到

发表于 2016-12-10 08:09:18 | 显示全部楼层 |阅读模式
虚拟一套centos 5.3 os.
下载 jdk-6u16-linux-i586-rpm.bin
[iyunv@hadoop ~]# chmod +x jdk-6u16-linux-i586-rpm.bin
[iyunv@hadoop ~]# ./jdk-6u16-linux-i586-rpm.bin
[iyunv@hadoop ~]# java -version
java version “1.6.0″
OpenJDK Runtime Environment (build 1.6.0-b09)
OpenJDK Client VM (build 1.6.0-b09, mixed mode)

[iyunv@hadoop yum.repos.d]# wget http://archive.cloudera.com/redhat/cdh/cloudera-testing.repo
[iyunv@hadoop yum.repos.d]# ls
CentOS-Base.repo CentOS-Base.repo.bak CentOS-Media.repo cloudera-testing.repo

[iyunv@hadoop ~]# yum install hadoop-0.20 -y
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
Setting up Install Process
Parsing package install arguments
Resolving Dependencies
–> Running transaction check
—> Package hadoop-0.20.noarch 0:0.20.0+69-1 set to be updated
–> Finished Dependency Resolution


Dependencies Resolved
===============================================================================
Package Arch Version Repository Size
===============================================
Installing:
hadoop-0.20 noarch 0.20.0+69-1 cloudera-testing 18 M

Transaction Summary
==================================================

Install 1 Package(s)
Update 0 Package(s)
Remove 0 Package(s)

Total download size: 18 M
Downloading Packages:
hadoop-0.20-0.20.0+69-1.noarch.rpm | 18 MB 01:34
Running rpm_check_debug
Running Transaction Test
Finished Transaction Test
Transaction Test Succeeded
Running Transaction
Installing : hadoop-0.20 [1/1]

Installed: hadoop-0.20.noarch 0:0.20.0+69-1
Complete!


root@hadoop conf]# yum install hadoop-0.20-conf-pseudo -y
Loaded plugins: fastestmirror
Loading mirror speeds from cached hostfile
Setting up Install Process
Parsing package install arguments
Resolving Dependencies
–> Running transaction check
—> Package hadoop-0.20-conf-pseudo.noarch 0:0.20.0+69-1 set to be updated
–> Finished Dependency Resolution

Dependencies Resolved
========================================================
Package Arch Version Repository Size
========================================================

Installing:
hadoop-0.20-conf-pseudo noarch 0.20.0+69-1 cloudera-testing 11 k

Transaction Summary
==========================================================

Install 1 Package(s)
Update 0 Package(s)
Remove 0 Package(s)

Total download size: 11 k
Downloading Packages:
hadoop-0.20-conf-pseudo-0.20.0+69-1.noarch.rpm | 11 kB 00:00
Running rpm_check_debug
Running Transaction Test
Finished Transaction Test
Transaction Test Succeeded
Running Transaction
Installing : hadoop-0.20-conf-pseudo [1/1]

Installed: hadoop-0.20-conf-pseudo.noarch 0:0.20.0+69-1
Complete!


安装完后可以在这个目录下看到。
[iyunv@hadoop conf.pseudo]# rpm -ql hadoop-0.20-conf-pseudo
/etc/hadoop-0.20/conf.pseudo
/etc/hadoop-0.20/conf.pseudo/README
/etc/hadoop-0.20/conf.pseudo/capacity-scheduler.xml
/etc/hadoop-0.20/conf.pseudo/configuration.xsl
/etc/hadoop-0.20/conf.pseudo/core-site.xml
/etc/hadoop-0.20/conf.pseudo/fair-scheduler.xml
/etc/hadoop-0.20/conf.pseudo/hadoop-env.sh
/etc/hadoop-0.20/conf.pseudo/hadoop-metrics.properties
/etc/hadoop-0.20/conf.pseudo/hadoop-policy.xml
/etc/hadoop-0.20/conf.pseudo/hdfs-site.xml
/etc/hadoop-0.20/conf.pseudo/log4j.properties
/etc/hadoop-0.20/conf.pseudo/mapred-site.xml
/etc/hadoop-0.20/conf.pseudo/masters
/etc/hadoop-0.20/conf.pseudo/slaves
/etc/hadoop-0.20/conf.pseudo/ssl-client.xml.example
/etc/hadoop-0.20/conf.pseudo/ssl-server.xml.example
/var/lib/hadoop-0.20
/var/lib/hadoop-0.20/cache

[iyunv@hadoop conf.pseudo]# pwd
/etc/hadoop-0.20/conf.pseudo


[iyunv@hadoop conf.pseudo]# more core-site.xml
<?xml version=”1.0″?>
<?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>

<configuration>
<property>
<name>fs.default.name</name>
<value>hdfs://localhost:8020</value>
</property>

<property>
<name>hadoop.tmp.dir</name>
<value>/var/lib/hadoop-0.20/cache/${user.name}</value>
</property>
</configuration>


启动hadoop相关服务:
[iyunv@hadoop conf.pseudo]# for service in /etc/init.d/hadoop-0.20-*
> do
> sudo $service start
> done
Starting Hadoop datanode daemon (hadoop-datanode): starting datanode, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-datanode-hadoop.out
[ OK ]
Starting Hadoop jobtracker daemon (hadoop-jobtracker): starting jobtracker, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-jobtracker-hadoop.out
[ OK ]
Starting Hadoop namenode daemon (hadoop-namenode): starting namenode, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-namenode-hadoop.out
[ OK ]
Starting Hadoop secondarynamenode daemon (hadoop-secondarynamenode): starting secondarynamenode, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-secondarynamenode-hadoop.out
[ OK ]
Starting Hadoop tasktracker daemon (hadoop-tasktracker): starting tasktracker, logging to /usr/lib/hadoop-0.20/bin/../logs/hadoop-hadoop-tasktracker-hadoop.out
[ OK ]


验证一下启动成功:
hadoop 3503 1 8 18:33 ? 00:00:03 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3577 1 10 18:33 ? 00:00:04 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3657 1 15 18:33 ? 00:00:05 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3734 1 11 18:33 ? 00:00:04 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dcom.sun.manage
hadoop 3827 1 7 18:33 ? 00:00:02 /usr/java/jdk1.6.0_16/bin/java -Xmx1000m -Dhadoop.log.di


测试几个例子:
root@hadoop conf.pseudo]# hadoop-0.20 fs -mkdir input
[iyunv@hadoop conf.pseudo]# hadoop-0.20 fs -put /etc/hadoop-0.20/conf/*.xml input
[iyunv@hadoop conf.pseudo]# hadoop-0.20 fs -ls input
Found 6 items
-rw-r–r– 1 root supergroup 6275 2009-08-25 18:34 /user/root/input/capacity-scheduler.xml
-rw-r–r– 1 root supergroup 338 2009-08-25 18:34 /user/root/input/core-site.xml
-rw-r–r– 1 root supergroup 3032 2009-08-25 18:34 /user/root/input/fair-scheduler.xml
-rw-r–r– 1 root supergroup 4190 2009-08-25 18:34 /user/root/input/hadoop-policy.xml
-rw-r–r– 1 root supergroup 496 2009-08-25 18:34 /user/root/input/hdfs-site.xml
-rw-r–r– 1 root supergroup 213 2009-08-25 18:34 /user/root/input/mapred-site.xml


[iyunv@hadoop conf.pseudo]# hadoop-0.20 jar /usr/lib/hadoop-0.20/hadoop-*-examples.jar grep input output ‘dfs[a-z.]+’
09/08/25 18:34:59 INFO mapred.FileInputFormat: Total input paths to process : 6
09/08/25 18:35:00 INFO mapred.JobClient: Running job: job_200908251833_0001
09/08/25 18:35:01 INFO mapred.JobClient: map 0% reduce 0%
09/08/25 18:35:20 INFO mapred.JobClient: map 33% reduce 0%
09/08/25 18:35:33 INFO mapred.JobClient: map 66% reduce 11%
09/08/25 18:35:42 INFO mapred.JobClient: map 66% reduce 22%
09/08/25 18:35:45 INFO mapred.JobClient: map 100% reduce 22%
09/08/25 18:35:57 INFO mapred.JobClient: map 100% reduce 100%
09/08/25 18:35:59 INFO mapred.JobClient: Job complete: job_200908251833_0001
09/08/25 18:35:59 INFO mapred.JobClient: Counters: 18
09/08/25 18:35:59 INFO mapred.JobClient: Job Counters
09/08/25 18:35:59 INFO mapred.JobClient: Launched reduce tasks=1
09/08/25 18:35:59 INFO mapred.JobClient: Launched map tasks=6
09/08/25 18:35:59 INFO mapred.JobClient: Da<wbr>ta-local map tasks=6<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: FileSystemCounters<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: FILE_BYTES_READ=100<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: HDFS_BYTES_READ=14544<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: FILE_BYTES_WRITTEN=422<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=204<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Map-Reduce Framework<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Reduce input groups=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Combine output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Map input records=364<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Reduce shuffle bytes=124<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Reduce output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Spilled Records=8<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Map output bytes=86<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Map input bytes=14544<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Combine input records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Map output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.JobClient: Reduce input records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:35:59 INFO mapred.FileInputFormat: Total input paths to process : 1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:00 INFO mapred.JobClient: Running job: job_200908251833_0002<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:01 INFO mapred.JobClient: map 0% reduce 0%<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:12 INFO mapred.JobClient: map 100% reduce 0%<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:24 INFO mapred.JobClient: map 100% reduce 100%<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Job complete: job_200908251833_0002<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Counters: 18<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Job Counters<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Launched reduce tasks=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Launched map tasks=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Da<wbr>ta-local map tasks=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: FileSystemCounters<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: FILE_BYTES_READ=100<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: HDFS_BYTES_READ=204<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: FILE_BYTES_WRITTEN=232<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=62<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Map-Reduce Framework<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Reduce input groups=1<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Combine output records=0<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Map input records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Reduce shuffle bytes=0<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Reduce output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Spilled Records=8<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Map output bytes=86<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Map input bytes=118<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Combine input records=0<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Map output records=4<br style="padding-bottom:0px; margin:0px; padding-left:0px; padding-right:0px; padding-top:0px">09/08/25 18:36:26 INFO mapred.JobClient: Reduce input records=4</wbr></wbr>


[iyunv@hadoop conf.pseudo]# hadoop-0.20 fs -ls
Found 2 items
drwxr-xr-x – root supergroup 0 2009-08-25 18:34 /user/root/input
drwxr-xr-x – root supergroup 0 2009-08-25 18:36 /user/root/output


[iyunv@hadoop conf.pseudo]# hadoop-0.20 fs -ls output
Found 2 items
drwxr-xr-x – root supergroup 0 2009-08-25 18:36 /user/root/output/_logs
-rw-r–r– 1 root supergroup 62 2009-08-25 18:36 /user/root/output/part-00000


[iyunv@hadoop conf.pseudo]# hadoop-0.20 fs -cat output/part-00000 | head
1 dfs.name.dir
1 dfs.permissions
1 dfs.replication
1 dfsadmin

转自:http://rdc.taobao.com/blog/dw/archives/436

运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-312088-1-1.html 上篇帖子: Hadoop The Definitive Guide 2nd Edition 读书笔记5 下篇帖子: hadoop 2.x-HDFS HA --Part II: installation
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表