设为首页 收藏本站
查看: 774|回复: 0

[经验分享] Install Hadoop with HDP 2.2 on CentOS 6.5

[复制链接]

尚未签到

发表于 2018-10-30 13:47:49 | 显示全部楼层 |阅读模式
  Hadoop Hortonworks
DSC0000.jpg

DSC0001.jpg

DSC0002.jpg

  Masters-- HDFS NameNode, YARN ResourceManager (ApplicationsManager + Scheduler), and HBase Master
  Slaves-- HDFS DataNodes, YARN NodeManagers, and HBase RegionServers (Worker Nodes co-deployed)
  ZooKeeper -- to manage HBase cluster
  HA -- NameNode + Secondary NameNode 或者NameNode HA (Active + Standby),推荐后者
  分区考虑,不要使用LVM
  root -- > 100G
  /tmp -- >100G (run job失败的话请查看此目录空间)
  swap -- 2倍系统内存
  Master node:
  RAID 10, dual Ethernet cards, dual power supplies, etc.
  Slave node:
  1. RAID is not necessary
  2. HDFS分区, not using LVM
  /etc/fstab -- ext3/ext4    defaults,noatime
  挂载到/grid/[0-n] (one partition per disk)
DSC0003.jpg

  HDP 2.2 and Ambari 1.7 repo
  http://public-repo-1.hortonworks.com/HDP/centos6/2.x/GA/2.2.0.0/hdp.repo
  http://public-repo-1.hortonworks.com/ambari/centos6/1.x/updates/1.7.0/ambari.repo
  yum -y install yum-utils
  yum repolist
  reposync -r HDP-2.2.0.0
  reposync -r HDP-UTILS-1.1.0.20
  reposync -r Updates-ambari-1.7.0
  on all cluster nodes and ambari server:
  cat  /etc/yum.repos.d/iso.repo
  [iso]
  name=iso
  baseurl=http://mirrors.aliyun.com/centos/6.5/os/x86_64/
  enable=1
  gpgcheck=0
  EOF
  vi /etc/hosts to add all hosts FQDN, like below:
  192.168.1.19    ambari.local
  192.168.1.20    master1.local # HDFS NameNode
  192.168.1.21    master2.local # YARN ResourceManager
  192.168.1.22    slave1.local
  192.168.1.23    slave2.local
  192.168.1.24    slave3.local
  192.168.1.25    client.local
  vi /etc/sysconfig/network to set FQDN
  yum -y install ntp
  service ntpd start; chkconfig ntpd on
  disable firewall and selinux:
  service iptables stop
  chkconfig iptables off; chkconfig ip6tables off
  setenforce 0
  sed -i 's,SELINUX=enforcing,SELINUX=disabled,g' /etc/selinux/config
  yum -y install openssh-clients wget unzip
  rpm -Uvh http://mirrors.aliyun.com/centos/6.5/updates/x86_64/Packages/openssl-1.0.1e-16.el6_5.15.x86_64.rpm
  rpm -ivh jdk-7u67-linux-x64.rpm
  echo 'export JAVA_HOME=/usr/java/default' >> /etc/profile
  echo 'export PATH=$JAVA_HOME/bin:$PATH' >> /etc/profile
  source /etc/profile
  just on all cluster nodes:
  disable ipv6 and kernel parameters tuning
  echo "net.ipv6.conf.all.disable_ipv6 = 1" >> /etc/sysctl.conf
  echo "vm.swappiness = 0" >> /etc/sysctl.conf
  echo 'net.ipv4.tcp_retries2 = 2' >> /etc/sysctl.conf
  echo 'vm.overcommit_memory = 1' >> /etc/sysctl.conf
  echo "fs.file-max = 6815744" >> /etc/sysctl.conf
  echo "fs.aio-max-nr = 1048576" >> /etc/sysctl.conf
  echo "net.core.rmem_default = 262144" >> /etc/sysctl.conf
  echo "net.core.wmem_default = 262144" >> /etc/sysctl.conf
  echo "net.core.rmem_max = 16777216" >> /etc/sysctl.conf
  echo "net.core.wmem_max = 16777216" >> /etc/sysctl.conf
  echo "net.ipv4.tcp_rmem = 4096 262144 16777216" >> /etc/sysctl.conf
  echo "net.ipv4.tcp_wmem = 4096 262144 16777216" >> /etc/sysctl.conf
  only on ResourceManager and JobHistory Server
  echo "net.core.somaxconn = 1000" >> /etc/sysctl.conf
  sysctl -p
  vi /etc/security/limits.conf
  *         soft    core         unlimited
  *         hard    core         unlimited
  *         soft    nofile       65536
  *         hard    nofile       65536
  *         soft    nproc        unlimited
  *         hard    nproc        unlimited
  *         soft    memlock      unlimited
  *         hard    memlock      unlimited
  echo "echo never > /sys/kernel/mm/redhat_transparent_hugepage/enabled" >> /etc/rc.local
  echo "echo never > /sys/kernel/mm/redhat_transparent_hugepage/defrag" >> /etc/rc.local
  echo "echo no > /sys/kernel/mm/redhat_transparent_hugepage/khugepaged/defrag" >> /etc/rc.local
  vi /etc/grub.conf
  add "elevator=deadline"(no quotes) at the end of kernel line
  reboot to take effect
  Just on ambari server:
  ssh-keygen -q -t rsa -f ~/.ssh/id_rsa -C '' -N ''
  vi ~/.ssh/config
  StrictHostKeyChecking no
  ssh-copy-id master1.local
  ssh-copy-id master2.local
  ssh-copy-id slave1.local
  ssh-copy-id slave2.local
  ssh-copy-id slave3.local
  yum -y install mysql-server httpd
  vi /etc/my.cnf
  [mysqld]
  default_storage_engine=InnoDB
  innodb_file_per_table=1
  service mysqld start; chkconfig mysqld on
  mysql_secure_installation
  vi /etc/yum.repos.d/ambari.repo
  [ambari]
  name=ambari
  baseurl=http://192.168.1.19/ambari
  enable=1
  gpgcheck=0
  mkdir /hdp; mount /dev/cdrom /hdp
  ln -s /hdp/ambari/ /var/www/html/ambari
  ln -s /hdp/hdp /var/www/html/hdp
  service httpd start; chkconfig httpd on
  For ambari/hive/oozie database:
  yum -y install ambari-server
  mysql -uroot -p
  create database ambari;

  GRANT ALL PRIVILEGES ON *.* TO 'ambari'@'localhost'>
  GRANT ALL PRIVILEGES ON *.* TO 'ambari'@'%'>  flush privileges;
  use ambari;
  source /var/lib/ambari-server/resources/Ambari-DDL-MySQL-CREATE.sql;
  create database hive;

  GRANT ALL PRIVILEGES ON *.* TO 'hive'@'localhost'>
  GRANT ALL PRIVILEGES ON *.* TO 'hive'@'%'>  flush privileges;
  create database oozie;

  GRANT ALL PRIVILEGES ON *.* TO 'oozie'@'localhost'>
  GRANT ALL PRIVILEGES ON *.* TO 'oozie'@'%'>  flush privileges;
  Install mysql jdbc connector:
  tar zxf mysql-connector-java-5.1.33.tar.gz
  mkdir /usr/share/java
  cp mysql-connector-java-5.1.33/mysql-connector-java-5.1.33-bin.jar /usr/share/java/mysql-connector-java.jar
  ambari-server setup --jdbc-db=mysql --jdbc-driver=/usr/share/java/mysql-connector-java.jar
  ambari-server setup -j /usr/java/default
  Enter advanced database configuration [y/n] (n)? y
  choose to use mysql, enter 3
  ambari-server start
  http://{your.ambari.server}:8080
  default user name/password: admin/admin
  使用Pi Estimator验证测试:
  sudo -u hdfs hadoop jar /usr/hdp/version-number/hadoop-mapreduce/hadoop-mapreduce-examples.jar pi 10 1000
  关机的正确步骤:
  1. stop all service in Ambari Web UI (if there is no response for stopping, service ambari-server restart, then try again)
  2. poweroff hosts
  Enable HA for NameNode and ResourceManager
DSC0004.jpg

DSC0005.jpg

DSC0006.jpg

  HUE installation:
  we will install HUE on client node:
  The Hive client configuration (hive-site.xml file) needs to be populated on the host to run Hue. In the beeswax section of hue.ini configure:
  hive_conf_dir=/etc/hive/conf
  Configure HDP
  1. Modify the core-site.xml
DSC0007.jpg 2. Modify the webhcat-site.xml file

DSC0008.jpg 3. Modify the oozie-site.xml file

DSC0009.jpg

  yum -y install hue
  vi /etc/hue/conf/hue.ini
  [desktop]
  time_zone=Asia/Shanghai
  [hadoop]
  `hdfs_clusters`
  [`default`]
  fs_defaultfs=hdfs://fqdn.namenode.host:8020
  webhdfs_url=http://fqdn.namenode.host:50070/webhdfs/v1/
  `yarn_clusters`
  [`default`]
  resourcemanager_api_url=http://fqdn.resourcemanager.host:8088
  resourcemanager_rpc_url=http://fqdn.resourcemanager.host:8050
  proxy_api_url=http://fqdn.resourcemanager.host:8088
  history_server_api_url=http://fqdn.historyserver.host:19888
  node_manager_api_url=http://fqdn.resourcemanager.host:8042
  [liboozie]
  oozie_url=http://fqdn.oozieserver.host:11000/oozie
  [beeswax]
  hive_server_host=fqdn.hiveserver.host
  [hcatalog]
  templeton_url=http://fqdn.webhcatserver.host:50111/templeton/v1/
  service hue start
  http://hue.server:8000
  Configuring Hue for MySQL Database
  mysql -uroot -p
  create database hue;

  GRANT ALL PRIVILEGES ON *.* TO 'hue'@'localhost'>
  GRANT ALL PRIVILEGES ON *.* TO 'hue'@'%'>  flush privileges;
  service hue stop
  vi /etc/hue/conf/hue.ini
  `database`
  engine=mysql
  host=fqdn.mysqlserver.host
  port=3306
  user=hue
  password=hue
  name=hue
  /usr/lib/hue/build/env/bin/hue syncdb --noinput
  service hue start
  This is just for hadoop tarball:
  disable firewall and selinux
  rpm -ivh jdk-7u67-linux-x64.rpm
  yum -y install wget rsync unzip openssh-clients
  ssh-keygen -q -t rsa -f ~/.ssh/id_rsa -C '' -N ''
  vi ~/.ssh/config
  StrictHostKeyChecking no
  ssh-copy-id localhost
  cd /usr/local/
  tar xvf /root/hadoop-2.5.1.tar.gz
  ln -s hadoop-2.5.1 hadoop
  vi ~/.bashrc
  export JAVA_HOME=/usr/java/default
  export HADOOP_HOME=/usr/local/hadoop
  export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin
  . ~/.bashrc
  vi $HADOOP_HOME/etc/hadoop/hadoop-env.sh
  export JAVA_HOME=/usr/java/default
  export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/usr/local/hadoop"}
  Web Interface:
  hadoop namenode: 50070
  hadoop secondarynode: 50090
  hadoop yarn ResourceManager: 8088
  hadoop yarn NodeManager: 8042
  hadoop MapReduce JobHistory Server: 19888


运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-628579-1-1.html 上篇帖子: 完整分布式Hadoop集群搭建过程 下篇帖子: Hadoop 1.2.1安装
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表