设为首页 收藏本站
查看: 632|回复: 0

[经验分享] CentOS 安装 hadoop(伪分布模式)

[复制链接]

尚未签到

发表于 2016-5-10 10:54:01 | 显示全部楼层 |阅读模式
在本机上装的CentOS 5.5 虚拟机,  软件准备:jdk 1.6 U26
  hadoop:hadoop-0.20.203.tar.gz
  

  
  ssh检查配置
  
Linux代码 DSC0000.png


  • [iyunv@localhost~]#ssh-keygen-trsa
  • Generatingpublic/privatersakeypair.
  • Enterfileinwhichtosavethekey(/root/.ssh/id_rsa):
  • Createddirectory'/root/.ssh'.
  • Enterpassphrase(emptyfornopassphrase):
  • Entersamepassphraseagain:
  • Youridentificationhasbeensavedin/root/.ssh/id_rsa.
  • Yourpublickeyhasbeensavedin/root/.ssh/id_rsa.pub.
  • Thekeyfingerprintis:
  • a8:7a:3e:f6:92:85:b8:c7:be:d9:0e:45:9c:d1:36:3broot@localhost.localdomain
  • [iyunv@localhost~]#
  • [iyunv@localhost~]#cd..
  • [iyunv@localhost/]#cdroot
  • [iyunv@localhost~]#ls
  • anaconda-ks.cfgDesktopinstall.loginstall.log.syslog
  • [iyunv@localhost~]#cd.ssh
  • [iyunv@localhost.ssh]#catid_rsa.pub>authorized_keys
  • [iyunv@localhost.ssh]#

  • [iyunv@localhost.ssh]#sshlocalhost
  • Theauthenticityofhost'localhost(127.0.0.1)'can'tbeestablished.
  • RSAkeyfingerprintis41:c8:d4:e4:60:71:6f:6a:33:6a:25:27:62:9b:e3:90.
  • Areyousureyouwanttocontinueconnecting(yes/no)?yes
  • Warning:Permanentlyadded'localhost'(RSA)tothelistofknownhosts.
  • Lastlogin:TueJun2122:40:312011
  • [iyunv@localhost~]#

  
  安装jdk
Linux代码


  • [iyunv@localhostjava]#chmod+xjdk-6u26-linux-i586.bin
  • [iyunv@localhostjava]#./jdk-6u26-linux-i586.bin
  • ......
  • ......
  • ......
  • FormoreinformationonwhatdataRegistrationcollectsand
  • howitismanagedandused,see:
  • http://java.sun.com/javase/registration/JDKRegistrationPrivacy.html

  • PressEntertocontinue.....


  • Done.

  安装完成后生成文件夹:jdk1.6.0_26
  
  配置环境变量
  
Linux代码


  • [iyunv@localhostjava]#vi/etc/profile
  • #添加如下信息
  • #setjavaenvironment
  • exportJAVA_HOME=/usr/java/jdk1.6.0_26
  • exportCLASSPATH=$CLASSPATH:$JAVA_HOME/lib:$JAVA_HOME/jre/lib
  • exportPATH=$JAVA_HOME/lib:$JAVA_HOME/jre/bin:$PATH:$HOME/bin
  • exportHADOOP_HOME=/usr/local/hadoop/hadoop-0.20.203
  • exportPATH=$PATH:$HADOOP_HOME/bin

  • [iyunv@localhostjava]#chmod+x/etc/profile
  • [iyunv@localhostjava]#source/etc/profile
  • [iyunv@localhostjava]#
  • [iyunv@localhostjava]#java-version
  • javaversion"1.6.0_26"
  • Java(TM)SERuntimeEnvironment(build1.6.0_26-b03)
  • JavaHotSpot(TM)ClientVM(build20.1-b02,mixedmode,sharing)
  • [iyunv@localhostjava]#

  修改hosts
Linuxa代码


  • [iyunv@localhostconf]#vi/etc/hosts
  • #Donotremovethefollowingline,orvariousprograms
  • #thatrequirenetworkfunctionalitywillfail.
  • 127.0.0.1localhost.localdomainlocalhost
  • ::1localhost6.localdomain6localhost6
  • 127.0.0.1namenodedatanode01

  
  解压安装hadoop
Linux代码


  • [iyunv@localhosthadoop]#tarzxvfhadoop-0.20.203.tar.gz
  • ......
  • ......
  • ......
  • hadoop-0.20.203.0/src/contrib/ec2/bin/image/create-hadoop-image-remote
  • hadoop-0.20.203.0/src/contrib/ec2/bin/image/ec2-run-user-data
  • hadoop-0.20.203.0/src/contrib/ec2/bin/launch-hadoop-cluster
  • hadoop-0.20.203.0/src/contrib/ec2/bin/launch-hadoop-master
  • hadoop-0.20.203.0/src/contrib/ec2/bin/launch-hadoop-slaves
  • hadoop-0.20.203.0/src/contrib/ec2/bin/list-hadoop-clusters
  • hadoop-0.20.203.0/src/contrib/ec2/bin/terminate-hadoop-cluster
  • [iyunv@localhosthadoop]#

  进入hadoop配置conf
Linux代码


  • ####################################
  • [iyunv@localhostconf]#vihadoop-env.sh
  • #添加代码
  • #setjavaenvironment
  • exportJAVA_HOME=/usr/java/jdk1.6.0_26

  • #####################################
  • [iyunv@localhostconf]#vicore-site.xml
  • <?xmlversion="1.0"?>
  • <?xml-stylesheettype="text/xsl"href="configuration.xsl"?>
  • <!--Putsite-specificpropertyoverridesinthisfile.-->
  • <configuration>
  • <property>
  • <name>fs.default.name</name>
  • <value>hdfs://namenode:9000/</value>
  • </property>
  • <property>
  • <name>hadoop.tmp.dir</name>
  • <value>/usr/local/hadoop/hadooptmp</value>
  • </property>
  • </configuration>

  • #######################################
  • [iyunv@localhostconf]#vihdfs-site.xml
  • <?xmlversion="1.0"?>
  • <?xml-stylesheettype="text/xsl"href="configuration.xsl"?>
  • <!--Putsite-specificpropertyoverridesinthisfile.-->
  • <configuration>
  • <property>
  • <name>dfs.name.dir</name>
  • <value>/usr/local/hadoop/hdfs/name</value>
  • </property>
  • <property>
  • <name>dfs.data.dir</name>
  • <value>/usr/local/hadoop/hdfs/data</value>
  • </property>
  • <property>
  • <name>dfs.replication</name>
  • <value>1</value>
  • </property>
  • </configuration>

  • #########################################
  • [iyunv@localhostconf]#vimapred-site.xml
  • <?xmlversion="1.0"?>
  • <?xml-stylesheettype="text/xsl"href="configuration.xsl"?>
  • <!--Putsite-specificpropertyoverridesinthisfile.-->
  • <configuration>
  • <property>
  • <name>mapred.job.tracker</name>
  • <value>namenode:9001</value>
  • </property>
  • <property>
  • <name>mapred.local.dir</name>
  • <value>/usr/local/hadoop/mapred/local</value>
  • </property>
  • <property>
  • <name>mapred.system.dir</name>
  • <value>/tmp/hadoop/mapred/system</value>
  • </property>
  • </configuration>

  • #########################################
  • [iyunv@localhostconf]#vimasters
  • #localhost
  • namenode

  • #########################################
  • [iyunv@localhostconf]#vislaves
  • #localhost
  • datanode01

  
  启动 hadoop
Linux代码


  • #####################<spanstyle="font-size:small;">格式化namenode##############</span>



  • [iyunv@localhostbin]#hadoopnamenode-format
  • 11/06/2300:43:54INFOnamenode.NameNode:STARTUP_MSG:
  • /************************************************************
  • STARTUP_MSG:StartingNameNode
  • STARTUP_MSG:host=localhost.localdomain/127.0.0.1
  • STARTUP_MSG:args=[-format]
  • STARTUP_MSG:version=0.20.203.0
  • STARTUP_MSG:build=http://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20-security-203-r1099333;compiledby'oom'onWedMay407:57:50PDT2011
  • ************************************************************/
  • 11/06/2300:43:55INFOutil.GSet:VMtype=32-bit
  • 11/06/2300:43:55INFOutil.GSet:2%maxmemory=19.33375MB
  • 11/06/2300:43:55INFOutil.GSet:capacity=2^22=4194304entries
  • 11/06/2300:43:55INFOutil.GSet:recommended=4194304,actual=4194304
  • 11/06/2300:43:56INFOnamenode.FSNamesystem:fsOwner=root
  • 11/06/2300:43:56INFOnamenode.FSNamesystem:supergroup=supergroup
  • 11/06/2300:43:56INFOnamenode.FSNamesystem:isPermissionEnabled=true
  • 11/06/2300:43:56INFOnamenode.FSNamesystem:dfs.block.invalidate.limit=100
  • 11/06/2300:43:56INFOnamenode.FSNamesystem:isAccessTokenEnabled=falseaccessKeyUpdateInterval=0min(s),accessTokenLifetime=0min(s)
  • 11/06/2300:43:56INFOnamenode.NameNode:Cachingfilenamesoccuringmorethan10times
  • 11/06/2300:43:57INFOcommon.Storage:Imagefileofsize110savedin0seconds.
  • 11/06/2300:43:57INFOcommon.Storage:Storagedirectory/usr/local/hadoop/hdfs/namehasbeensuccessfullyformatted.
  • 11/06/2300:43:57INFOnamenode.NameNode:SHUTDOWN_MSG:
  • /************************************************************
  • SHUTDOWN_MSG:ShuttingdownNameNodeatlocalhost.localdomain/127.0.0.1
  • ************************************************************/
  • [iyunv@localhostbin]#

  • ###########################################
  • [iyunv@localhostbin]#./start-all.sh
  • startingnamenode,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-namenode-localhost.localdomain.out
  • datanode01:startingdatanode,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-datanode-localhost.localdomain.out
  • namenode:startingsecondarynamenode,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-secondarynamenode-localhost.localdomain.out
  • startingjobtracker,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-jobtracker-localhost.localdomain.out
  • datanode01:startingtasktracker,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-tasktracker-localhost.localdomain.out
  • [iyunv@localhostbin]#jps
  • 11971TaskTracker
  • 11807SecondaryNameNode
  • 11599NameNode
  • 12022Jps
  • 11710DataNode
  • 11877JobTracker

  
  
  查看集群状态
Linux代码


  • [iyunv@localhostbin]#hadoopdfsadmin-report
  • ConfiguredCapacity:4055396352(3.78GB)
  • PresentCapacity:464142351(442.64MB)
  • DFSRemaining:464089088(442.59MB)
  • DFSUsed:53263(52.01KB)
  • DFSUsed%:0.01%
  • Underreplicatedblocks:0
  • Blockswithcorruptreplicas:0
  • Missingblocks:0

  • -------------------------------------------------
  • Datanodesavailable:1(1total,0dead)

  • Name:127.0.0.1:50010
  • DecommissionStatus:Normal
  • ConfiguredCapacity:4055396352(3.78GB)
  • DFSUsed:53263(52.01KB)
  • NonDFSUsed:3591254001(3.34GB)
  • DFSRemaining:464089088(442.59MB)
  • DFSUsed%:0%
  • DFSRemaining%:11.44%
  • Lastcontact:ThuJun2301:11:15PDT2011


  • [iyunv@localhostbin]#

  
  
  
  
  其他问题:1
Linux代码


  • ####################启动报错##########
  • [iyunv@localhostbin]#./start-all.sh
  • startingnamenode,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-namenode-localhost.localdomain.out
  • Theauthenticityofhost'datanode01(127.0.0.1)'can'tbeestablished.
  • RSAkeyfingerprintis41:c8:d4:e4:60:71:6f:6a:33:6a:25:27:62:9b:e3:90.
  • Areyousureyouwanttocontinueconnecting(yes/no)?y
  • Pleasetype'yes'or'no':yes
  • datanode01:Warning:Permanentlyadded'datanode01'(RSA)tothelistofknownhosts.
  • datanode01:startingdatanode,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-datanode-localhost.localdomain.out
  • <strong><spanstyle="color:#ff0000;">datanode01:Unrecognizedoption:-jvm
  • datanode01:CouldnotcreatetheJavavirtualmachine.</span>


  • </strong>



  • namenode:startingsecondarynamenode,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-secondarynamenode-localhost.localdomain.out
  • startingjobtracker,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-jobtracker-localhost.localdomain.out
  • datanode01:startingtasktracker,loggingto/usr/local/hadoop/hadoop-0.20.203/bin/../logs/hadoop-root-tasktracker-localhost.localdomain.out
  • [iyunv@localhostbin]#jps
  • 10442JobTracker
  • 10533TaskTracker
  • 10386SecondaryNameNode
  • 10201NameNode
  • 10658Jps

  • ################################################
  • [iyunv@localhostbin]#vihadoop
  • elif["$COMMAND"="datanode"];then
  • CLASS='org.apache.hadoop.hdfs.server.datanode.DataNode'
  • if[[$EUID-eq0]];then
  • HADOOP_OPTS="$HADOOP_OPTS-jvmserver$HADOOP_DATANODE_OPTS"
  • else
  • HADOOP_OPTS="$HADOOP_OPTS-server$HADOOP_DATANODE_OPTS"
  • fi

  • #http://javoft.net/2011/06/hadoop-unrecognized-option-jvm-could-not-create-the-java-virtual-machine/
  • #改为
  • elif["$COMMAND"="datanode"];then
  • CLASS='org.apache.hadoop.hdfs.server.datanode.DataNode'
  • #if[[$EUID-eq0]];then
  • #HADOOP_OPTS="$HADOOP_OPTS-jvmserver$HADOOP_DATANODE_OPTS"
  • #else
  • HADOOP_OPTS="$HADOOP_OPTS-server$HADOOP_DATANODE_OPTS"
  • #fi

  • #或者换非root用户启动
  • #启动成功

  2,启动时要关闭防火墙
  
  查看运行情况:
  http://localhost:50070
Firefox显示代码


  • NameNode'localhost.localdomain:9000'
  • Started:ThuJun2301:07:18PDT2011
  • Version:0.20.203.0,r1099333
  • Compiled:WedMay407:57:50PDT2011byoom
  • Upgrades:Therearenoupgradesinprogress.

  • Browsethefilesystem
  • NamenodeLogs
  • ClusterSummary
  • 6filesanddirectories,1blocks=7total.HeapSizeis31.38MB/966.69MB(3%)
  • ConfiguredCapacity:3.78GB
  • DFSUsed:52.01KB
  • NonDFSUsed:3.34GB
  • DFSRemaining:442.38MB
  • DFSUsed%:0%
  • DFSRemaining%:11.44%
  • LiveNodes:1
  • DeadNodes:0
  • DecommissioningNodes:0
  • NumberofUnder-ReplicatedBlocks:0

  • NameNodeStorage:
  • StorageDirectoryTypeState
  • /usr/local/hadoop/hdfs/nameIMAGE_AND_EDITSActive

  http://localhost:50030
Firefox显示代码


  • namenodeHadoopMap/ReduceAdministration
  • QuickLinks

  • *SchedulingInfo
  • *RunningJobs
  • *RetiredJobs
  • *LocalLogs

  • State:RUNNING
  • Started:ThuJun2301:07:30PDT2011
  • Version:0.20.203.0,r1099333
  • Compiled:WedMay407:57:50PDT2011byoom
  • Identifier:201106230107
  • ClusterSummary(HeapSizeis15.31MB/966.69MB)
  • RunningMapTasksRunningReduceTasksTotalSubmissionsNodesOccupiedMapSlotsOccupiedReduceSlotsReservedMapSlotsReservedReduceSlotsMapTaskCapacityReduceTaskCapacityAvg.Tasks/NodeBlacklistedNodesGraylistedNodesExcludedNodes
  • 00010000224.00000

  • SchedulingInformation
  • QueueNameStateSchedulingInformation
  • defaultrunningN/A
  • Filter(Jobid,Priority,User,Name)
  • Example:'user:smith3200'willfilterby'smith'onlyintheuserfieldand'3200'inallfields
  • RunningJobs
  • none
  • RetiredJobs
  • none
  • LocalLogs
  • Logdirectory,JobTrackerHistoryThisisApacheHadooprelease0.20.203.0

  
  测试:
Linux代码


  • ##########建立目录名称##########
  • [iyunv@localhostbin]#hadoopfs-mkdirtestFolder

  • ###############拷贝文件到文件夹中
  • [iyunv@localhostlocal]#ls
  • binetcgameshadoopincludeliblibexecsbinsharesrcSSH_key_file
  • [iyunv@localhostlocal]#hadoopfs-copyFromLocalSSH_key_filetestFolder

  • 进入web页面即可查看

  
  
  参考:http://bxyzzy.blog.iyunv.com/854497/352692
  
  附: 准备FTP :yum install vsftpd (方便文件传输 和hadoop无关)
  关闭防火墙:service iptables start
  启动FTP:service vsftpd start

运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-215101-1-1.html 上篇帖子: [转] Centos下的LNMP架构 下篇帖子: CentOS下NFS服务器配置实例
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表