设为首页 收藏本站
查看: 767|回复: 0

[经验分享] WINDOWS 下 HADOOP 的安装

[复制链接]
YunVN网友  发表于 2018-6-24 06:15:27 |阅读模式
  Windows hadoop installation
  1.Install Cygwin
  2.Install Cygwin components:openssh,openssl,sed,subversion
  3.Add Cygwin/bin and Cygwin/usr/sbin to windows path
  4.Install sshd
  In Cygwin, runssh-host-config
  Should privilege separation used (no)
  Do you want to install sshd as a service (yes)
  Cygwin will also prompt whether you want to create a new windows user to start the service, default user created is “cyg_server”, it is better to use the current domain user
  5.Config ssh login
  In Cygwin, run ssh-keygen
  6.Start sshd service in windowscontrol panel “service”
  Or callnet start sshd, if the service failed to start, check \var\log\ssh.log
  7.Verify ssh login
  In Cygwin, run ssh localhost
  Sometimes the default port 22 is not good for usage
  We can change port by modify file sshd_config:Port xxx, and change command to ssh localhost-p xxx
  For detailed logs using ssh –v localhost
  8.Download and extract hadoop in afile folder
  9.Change JAVA_HOME in conf/hadoop-env.sh
  10.Test setup
  cp conf/*.xml input
  bin/hadoop jar hadoop-examples-*.jar grep input output ‘dfs[a-z.]+’
  Problems encountered during installation
  1.The first time, installsshd service failed
  I need to run
  sc delete sshd to delete the service and run ssh-host-config again
  2.Error:Privilege separation user sshd does not exist
  Manually add the following line
  sshd:x:74:74:Privilege-separated SSH:/var/empty/sshd:/sbin/nologin to file:“etc/passwd
  etc/pwd format:

  username:password:user>  When user logs in, a shell process is started to pass user input to kernel
  3.Error:Connection closed by 1
  If user A need to ssh connect to user B on host B,we need to copy A’s public key to a file called “authorized_keys” under host B’s“home/<user B>” folder
  Create authorized_keys file:vi authorized_keys

  Copy public key to authorized_keys file:cat>  For ssh, access right of .ssh folder and authorized_keys file need to be set correctly
  Chmod 700 /.ssh
  Chmod 600/.ssh/authorized_keys (we cannot grant write access to authorized_keys file)
  4.Error: Starting hadoop: Java.io.IOException:failed to set permissions of path:\tmp\hadoop-jizhan\mapred\staging\jizhan…..\.staging

  This problem occurs because of a compatibility problem in>  We need to manually change the method checkReturnValue,just log warn message instead of throw exception
  Reference
  http://bbym010.iteye.com/blog/1019653
  Running Hadoop
  1.Under stand-alone mode:
  Leave defaultconfiguration
  Put file to process directly under hadoop/input folder(no need for hadoop file system upload). Output file will be written to hadoop/output folder
  2.Under pseudo-distributedmode:
  Core-site.xml
  <configuration>
  <property>
  <name>fs.default.name</name>
  <value>hdfs://localhost:9890</value>
  </property>
  </configuration>
  Mapred-site.xml
  <configuration>
  <property>
  <name>mapred.job.tracker</name>
  <value>hdfs://localhost:9891</value>
  </property>
  </configuration>
  Hdfs-site.xml
  <configuration>
  <property>
  <name>dfs.replication</name>
  <value>1</value>
  </property>
  </configuration>
  Make sure thatlocalhost is in master file
  Make sure thatlocalhost is in slaves file
  Problem encountered running in standalone mode
  1.Reducer does not execute.
  There are a few things to check when encountering this problem
  It is good to explicitly specify mapper and reducer’s output keyclass and value
  actual mapper and reducer’s parameter type must match specification,mapper’s output parameter type must match reducer’s input parameter type
  Raw Context object will not be accepted for map or reduce method,you need to use a strong typed context.
  Mapper<InputKey, InputValue,OutputKey, OutputValue>.Context
  Reducer<InputKey, InputValue,OutputKey, OutputValue>.Context
  2.Line Reader does not readline correctly, a shorter line carries additional characters from previouslonger line
  This is due to a wrong way of using Text, a text has an internal byte array and an end index, so the Text object may contain additional data due to internal buffer expansion after reading a longer line, those chars will not be cleared and only chars before index should be read for a shorter line.
  Do not usenew String(text.getBytes())to convert text to string, usetext.toString()
  Problem encountered running in pseudo-distributed mode
  Error running map-reduce program
  14/01/19 12:21:25 WARN mapred.JobClient: Error reading task output http://L-SHC-0436751.corp.ebay.com:50060/tasklog?plaintext=true&attemptid=attempt_20140119128_0002_m_000001_2&filter=stderr
  Hadoop uses unix file link to redirect output in {HADOOP_DIR}/logs to tmp/hadoop-jizhan/mapred/local(note that hadoop.tmp.dir-> tmp/hadoop-jizhan/)
  This is not recognized as a directory in windows by jdk and exception is thrown
  To avoid redirection, we can set property HADOOP_LOG_DIR directly pointing to /tmp/mapred/local this is the Cygwin /tmp folder, and we need to use unix ln command to map it to local folder c:/tmp/hadoop-jizhan/mapred/local

运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-529774-1-1.html 上篇帖子: How to hack windows 8 using kali linux 下篇帖子: windows 域建用户
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表