设为首页 收藏本站
查看: 1141|回复: 0

[经验分享] hadoop HA 配置文件

[复制链接]

尚未签到

发表于 2018-10-29 10:21:34 | 显示全部楼层 |阅读模式
  fs.defaultFS     hdfs://rongxinhadoop
  这里的 mycluster为HA集群的逻辑名,与hdfs-site.xml中的dfs.nameservices配置一致  hadoop.tmp.dir     /data/hadoop1/HAtmp3
  这里的路径默认是NameNode、DataNode、JournalNode等存放数据的公共目录。用户也可单独指定每类数据的存储目录。这里目录结构需要自己先创建好
  ha.zookeeper.quorum   master:2181,slave1:2181,slave2:2181
  这里是zk集群配置中各节点的地址和端口。
  注意:数量一定是奇数而且和zoo.cfg中配置的一致
  --------------------------------------------------------------------------------------------------
  dfs.replication     2     配置副本数量
  dfs.namenode.name.dir     file:/data/hadoop1/HAname3     namenode元数据存储目录  dfs.datanode.data.dir     file:/data/hadoop1/HAdata3     datanode数据存储目录  dfs.nameservices  rongxinhadoop 指定HA命名服务,可随意起名,   core-site.xml中fs.defaultFS配置需要引用它
  dfs.ha.namenodes.rongxinhadoop nn1,nn2 指定集群下NameNode逻辑名
  dfs.namenode.rpc-address.rongxinhadoop.nn1   master:9000
  dfs.namenode.rpc-address.rongxinhadoop.nn2   slave1:9000
  dfs.namenode.http-address.rongxinhadoop.nn1   master:50070
  dfs.namenode.http-address.rongxinhadoop.nn2   slave1:50070
  dfs.namenode.servicerpc-address.rongxinhadoop.nn1    master:53310   dfs.namenode.servicerpc-address.rongxinhadoop.nn2    slave1:53310
  dfs.ha.automatic-failover.enabled.rongxinhadoop  true 故障失败是否自动切换   dfs.namenode.shared.edits.dir qjournal://master:8485;slave1:8485;slave2:8485/rongxinhadoop 配置JournalNode,包含三部分:
  1.qjournal 前缀表名协议;
  2.然后就是三台部署JournalNode的主机host/ip:端口,三台机器之间用分号分隔;
  3.最后的hadoop-journal是journalnode的命名空间,可以随意取名。    dfs.journalnode.edits.dir /data/hadoop1/HAjournal3/ journalnode的本地数据存放目录 dfs.client.failover.proxy.provider.rongxinhadoop
  org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider   指定mycluster出故障时执行故障切换的类
  dfs.ha.fencing.methods        sshfence   ssh的操作方式执行故障切换
  dfs.ha.fencing.ssh.private-key-files        /home/hadoop1/.ssh/id_rsa   如果使用ssh进行故障切换,使用ssh通信时用的密钥存储的位置
  dfs.ha.fencing.ssh.connect-timeout    1000      dfs.namenode.handler.count    10
  --------------------------------------------------------------------------------------------------
  mapreduce.framework.name yarn   mapreduce.jobhistory.address master:10020  mapreduce.jobhistory.webapp.address master:19888  mapreduce.jobhistory.intermediate-done-dir /data/hadoop1/mr_history/HAtmp3 Directory where history files are written by MapReduce jobs.  mapreduce.jobhistory.done-dir /data/hadoop1/mr_history/HAdone3 Directory where history files are managed by the MR JobHistory Server.
  --------------------------------------------------------------------------------------------------
  
  
  -
  
  -
  yarn.resourcemanager.ha.enabled
  true
  
  -
  yarn.resourcemanager.cluster-id
  clusterrm
  
  -
  yarn.resourcemanager.ha.rm-ids
  rm1,rm2
  
  -
  yarn.resourcemanager.hostname.rm1
  master
  
  -
  yarn.resourcemanager.hostname.rm2
  slave1
  
  -
  yarn.resourcemanager.recovery.enabled
  true
  
  -
  yarn.resourcemanager.store.class
  org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore
  
  -
  yarn.resourcemanager.zk-address
  master:2181,slave1:2181,slave2:2181
  
  -
  yarn.nodemanager.aux-services
  mapreduce_shuffle
  
  -
  yarn.nodemanager.aux-services.mapreduce.shuffle.class
  org.apache.hadoop.mapred.ShuffleHandler
  
  
  
  -
  yarn.log-aggregation-enable
  true
  
  
  -
  The hostname of the Timeline service web application.
  yarn.timeline-service.hostname
  master
  
  -
  Address for the Timeline server to start the RPC server.
  yarn.timeline-service.address
  master:10200
  
  -
  The http address of the Timeline service web application.
  yarn.timeline-service.webapp.address
  master:8188
  
  -
  The https address of the Timeline service web application.
  yarn.timeline-service.webapp.https.address
  master:8190
  
  -
  Handler thread count to serve the client RPC requests.
  yarn.timeline-service.handler-thread-count
  10
  
  -
  Enables cross-origin support (CORS) for web services where cross-origin web response headers are needed. For example, javascript making a web services request to the timeline server.
  yarn.timeline-service.http-cross-origin.enabled
  false
  
  -
  Comma separated list of origins that are allowed for web services needing cross-origin (CORS) support. Wildcards (*) and patterns allowed
  yarn.timeline-service.http-cross-origin.allowed-origins
  *
  
  -
  Comma separated list of methods that are allowed for web services needing cross-origin (CORS) support.
  yarn.timeline-service.http-cross-origin.allowed-methods
  GET,POST,HEAD
  
  -
  Comma separated list of headers that are allowed for web services needing cross-origin (CORS) support.
  yarn.timeline-service.http-cross-origin.allowed-headers
  X-Requested-With,Content-Type,Accept,Origin
  
  -
  The number of seconds a pre-flighted request can be cached for web services needing cross-origin (CORS) support.
  yarn.timeline-service.http-cross-origin.max-age
  1800
  
  -
  Indicate to clients whether Timeline service is enabled or not. If enabled, the TimelineClient library used by end-users will post entities and events to the Timeline server.
  yarn.timeline-service.enabled
  true
  
  -

  Store>  yarn.timeline-service.store-class
  org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore
  
  -
  Enable age off of timeline store data.
  yarn.timeline-service.ttl-enable
  true
  
  -
  Time to live for timeline store data in milliseconds.
  yarn.timeline-service.ttl-ms
  604800000
  
  


运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-627865-1-1.html 上篇帖子: hadoop集群内存设置 下篇帖子: Hadoop-Hello,World!
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表