设为首页 收藏本站
查看: 1851|回复: 0

[经验分享] hadoop-2.6.0+zookeeper-3.4.6+hbase-1.0.0+hive-1.1.0完全分布式集群HA部署

[复制链接]
发表于 2018-10-30 09:32:05 | 显示全部楼层 |阅读模式
  我这也是整合了大家的东西整理出来的,如果有不足或者错误的地方,希望大家指正
  因为准备生产环境用的,所以都做了HA
  HDFS的NameNode HA
  YARN的Resource Manager HA
  Hbase的Hmaster HA
  Hive的Hiveserver2 HA
  PS:听说还有人对JobHistory做HA,这个我没做,谁有经验的也跟我讲下
  =================================================================
DSC0000.jpg

  部署hadoop完全分布式集群
  一、准备环境
  移除已有jdk(最小化安装没有自带jdk,可略过此步骤)
  yum remove java-1.7.0-openjdk -y
  关闭防火墙和selinux
  service iptables stop
  chkconfig iptables off
  setenforce 0
  vi /etc/selinux/config
  SELINUX=disabled
  所有机器同步时间
  ntpdate 202.120.2.101
  配置主机名和hosts
  vi /etc/networks
  HOSTNAME=hadoop001~005
  编辑hosts文件
  vi /etc/hosts
  192.168.5.2hadoop001.test.com
  192.168.5.3hadoop002.test.com
  192.168.5.4hadoop003.test.com
  192.168.5.5hadoop004.test.com
  192.168.5.6hadoop005.test.com
  配置yum源
  vi /etc/yum.repos.d/rhel.repo
  创建hadoop用户和组
  groupadd hadoop
  useradd -g hadoop hadoop
  passwd hadoop
  赋权限,以备后续步骤安装软件[安装包都在/usr/local/src]
  chown hadoop.hadoop /usr/local/src -R
  切换hadoop用户
  su - hadoop
  配置密钥验证免密码登录[所有namenode都要做一遍]
  ssh-keygen -t rsa -P ''
  cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
  chmod 700 ~/.ssh/
  chmod 600 ~/.ssh/authorized_keys
  ssh-copy-id -i $HOME/.ssh/id_rsa.pub hadoop@hadoop002.test.com
  ssh-copy-id -i $HOME/.ssh/id_rsa.pub hadoop@hadoop003.test.com
  ssh-copy-id -i $HOME/.ssh/id_rsa.pub hadoop@hadoop004.test.com
  ssh-copy-id -i $HOME/.ssh/id_rsa.pub hadoop@hadoop005.test.com
  验证
  ssh hadoop002.test.com ~ hadoop005.test.com
  创建备用目录
  mkdir -pv /home/hadoop/storage/hadoop/tmp
  mkdir -pv /home/hadoop/storage/hadoop/name
  mkdir -pv /home/hadoop/storage/hadoop/data
  mkdir -pv /home/hadoop/storage/hadoop/journal
  mkdir -pv /home/hadoop/storage/yarn/local
  mkdir -pv /home/hadoop/storage/yarn/logs
  mkdir -pv /home/hadoop/storage/hbase
  mkdir -pv /home/hadoop/storage/zookeeper/data
  mkdir -pv /home/hadoop/storage/zookeeper/logs
  scp -r /home/hadoop/storage hadoop002.test.com:/home/hadoop/
  scp -r /home/hadoop/storage hadoop003.test.com:/home/hadoop/
  scp -r /home/hadoop/storage hadoop004.test.com:/home/hadoop/
  scp -r /home/hadoop/storage hadoop005.test.com:/home/hadoop/
  配置hadoop环境变量
  vi ~/.bashrc
  export JAVA_HOME=/usr/java/jdk1.7.0_67-cloudera
  export JRE_HOME=$JAVA_HOME/jre

  export>  export PATH=$PATH:$JAVA_HOME/bin
  ##java env
  export HADOOP_HOME=/home/hadoop/hadoop
  export HIVE_HOME=/home/hadoop/hive
  export HBASE_HOME=/home/hadoop/hbase
  ##hadoop hbase hive
  export HADOOP_MAPRED_HOME=${HADOOP_HOME}
  export HADOOP_COMMON_HOME=${HADOOP_HOME}
  export HADOOP_HDFS_HOME=${HADOOP_HOME}
  export YARN_HOME=${HADOOP_HOME}
  export HADOOP_YARN_HOME=${HADOOP_HOME}
  export HADOOP_CONF_DIR=${HADOOP_HOME}/etc/hadoop
  export HDFS_CONF_DIR=${HADOOP_HOME}/etc/hadoop
  export YARN_CONF_DIR=${HADOOP_HOME}/etc/hadoop
  export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$HBASE_HOME/bin:$HIVE_HOME/bin
  source ~/.bashrc
  scp ~/.bashrc hadoop002.test.com:~/.bashrc
  scp ~/.bashrc hadoop003.test.com:~/.bashrc
  scp ~/.bashrc hadoop004.test.com:~/.bashrc
  scp ~/.bashrc hadoop005.test.com:~/.bashrc
  安装jdk
  rpm -ivh oracle-j2sdk1.7-1.7.0+update67-1.x86_64.rpm
  配置环境变量[准备环境时已做,此步骤可略]
  验证jdk安装成功
  java -version
  二、部署hadoop-2.6.0的namenoe HA、resource manager HA
  解压、改名
  tar xf /usr/local/src/hadoop-2.6.0.tar.gz -C /home/hadoop
  cd /home/hadoop
  mv hadoop-2.6.0 hadoop
  配置hadoop环境变量[准备环境时已做,略]
  验证hadoop安装成功
  hadoop version
  修改hadoop配置文件
  [1]
  vi /home/hadoop/hadoop/etc/hadoop/core-site.xml
  #################################################################################
  
  
  
  
  
  fs.defaultFS
  hdfs://gagcluster:9000
  
  
  io.file.buffer.size
  131072
  
  
  
  hadoop.tmp.dir
  /home/hadoop/storage/hadoop/tmp
  Abase for other temporary directories.
  
  
  
  hadoop.proxyuser.hduser.hosts
  *
  
  
  
  hadoop.proxyuser.hduser.groups
  *
  
  
  
  ha.zookeeper.quorum
  hadoop001.test.com:2181,hadoop002.test.com:2181,hadoop003.test.com:2181
  
  
  #################################################################################
  [2]
  vi /home/hadoop/hadoop/etc/hadoop/hdfs-site.xml
  #################################################################################
  
  
  
  
  
  dfs.hosts.exclude
  /home/hadoop/hadoop/etc/hadoop/exclude
  
  
  
  dfs.block.size
  67108864
  
  
  
  dfs.nameservices
  gagcluster
  
  
  
  dfs.ha.namenodes.gagcluster
  nn1,nn2
  
  
  
  dfs.namenode.rpc-address.gagcluster.nn1
  hadoop001.test.com:9000
  
  
  
  dfs.namenode.http-address.gagcluster.nn1
  hadoop001.test.com:50070
  
  
  
  dfs.namenode.rpc-address.gagcluster.nn2
  hadoop002.test.com:9000
  
  
  
  dfs.namenode.http-address.gagcluster.nn2
  hadoop002.test.com:50070
  
  
  
  dfs.namenode.shared.edits.dir
  qjournal://hadoop001.test.com:8485;hadoop002.test.com:8485;hadoop003.test.com:8485/gagcluster
  
  
  
  dfs.client.failover.proxy.provider.gagcluster
  org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider
  
  
  
  dfs.ha.fencing.methods
  sshfence
  
  
  
  dfs.ha.fencing.ssh.private-key-files
  /home/hadoop/.ssh/id_rsa
  
  
  
  dfs.journalnode.edits.dir
  /home/hadoop/storage/hadoop/journal
  
  
  
  dfs.ha.automatic-failover.enabled
  true
  
  
  
  dfs.namenode.name.dir
  /home/hadoop/storage/hadoop/name
  
  
  
  dfs.datanode.data.dir
  file:/home/hadoop/storage/hadoop/data
  
  
  
  dfs.replication
  3
  
  
  
  dfs.webhdfs.enabled
  true
  
  
  
  dfs.journalnode.http-address
  0.0.0.0:8480
  
  
  dfs.journalnode.rpc-address
  0.0.0.0:8485
  
  
  ha.zookeeper.quorum
  hadoop001.test.com:2181,hadoop002.test.com:2181,hadoop003.test.com:2181
  
  
  #################################################################################
  [3]
  vi /home/hadoop/hadoop/etc/hadoop/mapred-site.xml
  #################################################################################
  
  
  
  mapreduce.framework.name
  yarn
  
  
  
  mapreduce.jobhistory.address
  0.0.0.0:10020
  
  
  
  mapreduce.jobhistory.webapp.address
  0.0.0.0:19888
  
  
  #################################################################################
  [4]
  yarn-site.xml
  #################################################################################
  
  
  
  
  yarn.log-aggregation-enable
  true
  
  
  
  yarn.log-aggregation.retain-seconds
  259200
  
  
  
  yarn.resourcemanager.connect.retry-interval.ms
  2000
  
  
  
  yarn.resourcemanager.ha.enabled
  true
  
  
  
  yarn.resourcemanager.ha.rm-ids
  rm1,rm2
  
  
  ha.zookeeper.quorum
  hadoop001.test.com:2181,hadoop002.test.com:2181,hadoop003.test.com:2181
  
  
  
  yarn.resourcemanager.ha.automatic-failover.enabled
  true
  
  
  yarn.resourcemanager.hostname.rm1
  hadoop001.test.com
  
  
  yarn.resourcemanager.hostname.rm2
  hadoop002.test.com
  
  
  
  yarn.resourcemanager.ha.id
  rm1
  If we want to launch more than one RM in single node, we need this configuration
  
  
  
  yarn.resourcemanager.recovery.enabled
  true
  
  
  
  yarn.resourcemanager.zk-state-store.address
  hadoop001.test.com:2181,hadoop002.test.com:2181,hadoop003.test.com:2181
  
  
  yarn.resourcemanager.store.class
  org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore
  
  
  yarn.resourcemanager.zk-address
  hadoop001.test.com:2181,hadoop002.test.com:2181,hadoop003.test.com:2181
  
  
  yarn.resourcemanager.cluster-id
  gagcluster-yarn
  
  
  
  yarn.app.mapreduce.am.scheduler.connection.wait.interval-ms
  5000
  
  
  
  yarn.resourcemanager.address.rm1
  hadoop001.test.com:8132
  
  
  yarn.resourcemanager.scheduler.address.rm1
  hadoop001.test.com:8130
  
  
  yarn.resourcemanager.webapp.address.rm1
  hadoop001.test.com:8188
  
  
  yarn.resourcemanager.resource-tracker.address.rm1
  hadoop001.test.com:8131
  
  
  yarn.resourcemanager.admin.address.rm1
  hadoop001.test.com:8033
  
  
  yarn.resourcemanager.ha.admin.address.rm1
  hadoop001.test.com:23142
  
  
  
  yarn.resourcemanager.address.rm2
  hadoop002.test.com:8132
  
  
  yarn.resourcemanager.scheduler.address.rm2
  hadoop002.test.com:8130
  
  
  yarn.resourcemanager.webapp.address.rm2
  hadoop002.test.com:8188
  
  
  yarn.resourcemanager.resource-tracker.address.rm2
  hadoop002.test.com:8131
  
  
  yarn.resourcemanager.admin.address.rm2
  hadoop002.test.com:8033
  
  
  yarn.resourcemanager.ha.admin.address.rm2
  hadoop002.test.com:23142
  
  
  yarn.nodemanager.aux-services
  mapreduce_shuffle
  
  
  yarn.nodemanager.aux-services.mapreduce.shuffle.class
  org.apache.hadoop.mapred.ShuffleHandler
  
  
  yarn.nodemanager.local-dirs
  /home/hadoop/storage/yarn/local
  
  
  yarn.nodemanager.log-dirs
  /home/hadoop/storage/yarn/logs
  
  
  mapreduce.shuffle.port
  23080
  
  
  
  yarn.client.failover-proxy-provider
  org.apache.hadoop.yarn.client.ConfiguredRMFailoverProxyProvider
  
  
  yarn.resourcemanager.ha.automatic-failover.zk-base-path
  /yarn-leader-election
  Optional setting. The default value is /yarn-leader-election
  
  
  ##########################################################################
  配置DataNode节点
  vi /home/hadoop/hadoop/etc/hadoop/slaves
  ##########################################################################
  hadoop001.test.com
  hadoop002.test.com
  hadoop003.test.com
  hadoop004.test.com
  hadoop005.test.com
  创建exclude文件,用于以后下线hadoop节点
  touch /home/hadoop/hadoop/etc/hadoop/exclude
  同步hadoop工程到hadoop002~005机器上面
  scp -r /home/hadoop/hadoop hadoop002.test.com:/home/hadoop/
  scp -r /home/hadoop/hadoop hadoop003.test.com:/home/hadoop/
  scp -r /home/hadoop/hadoop hadoop004.test.com:/home/hadoop/
  scp -r /home/hadoop/hadoop hadoop005.test.com:/home/hadoop/
  修改nn2配置文件yarn-site.xml
  三、部署zookeeper3.4.6三节点完全分布式集群
  使用三台服务器安装zookeeper,安装在hadoop用户上
  hadoop001.test.com 192.168.5.2
  hadoop002.test.com 192.168.5.3
  hadoop003.test.com 192.168.5.4
  解压、改名
  tar xf zookeeper-3.4.6.tar.gz -C /home/hadoop/
  mv /home/hadoop/zookeeper-3.4.6/ /home/hadoop/zookeeper
  cd /home/hadoop/zookeeper
  修改配置文件
  vi /usr/local/zookeeper/conf/zoo.cfg
  tickTime=2000
  initLimit=5
  syncLimit=2
  dataDir=/home/hadoop/storage/zookeeper/data
  dataLogDir=/home/hadoop/storage/zookeeper/logs
  clientPort=2181
  server.1=hadoop001.test.com:2888:3888
  server.2=hadoop002.test.com:2888:3888
  server.3=hadoop003.test.com:2888:3888
  同步到hadoop002、hadoop003节点
  scp -r /home/hadoop/zookeeper hadoop002.test.com:/home/hadoop
  scp -r /home/hadoop/zookeeper hadoop003.test.com:/home/hadoop
  创建zookeeper的数据文件和日志存放目录[准备环境已做,此步骤略]
  hadoop001~003分别编辑myid值
  echo 1 > /home/hadoop/storage/zookeeper/data/myid
  echo 2 > /home/hadoop/storage/zookeeper/data/myid
  echo 3 > /home/hadoop/storage/zookeeper/data/myid
  ====================现在不执行,统一启动=========================
  启动命令
  /home/hadoop/zookeeper/bin/zkServer.sh start
  验证安装成功
  通过jps命令可以查看ZooKeeper服务器进程,名称为QuorumPeerMain
  查看状态
  /home/hadoop/zookeeper/bin/zkServer.sh status
  启动命令行
  /home/hadoop/zookeeper/bin/zkCli.sh
  查看是否有hadoop-ha
  ls /
  查看是否有逻辑namenode组名称
  ls /hadoop-ha
  ================================================================
  四、部署hbase-1.0.0的Hmaster HA
  解压部署
  tar xf hbase-1.0.0-bin.tar.gz -C /home/hadoop
  cd /home/hadoop
  mv hbase-1.0.0 hbase
  修改配置文件
  配置regionserver节点
  vi /home/hadoop/hbase/conf/regionservers
  hadoop001.test.com
  hadoop002.test.com
  hadoop003.test.com
  hadoop004.test.com
  hadoop005.test.com
  vi /home/hadoop/hbase/conf/hbase-site.xml
  #########################################################################
  
  
  
  
  
  hbase.rootdir
  hdfs://gagcluster/hbase
  
  
  hbase.cluster.distributed
  true
  
  
  hbase.tmp.dir
  /home/hadoop/storage/hbase
  
  
  hbase.zookeeper.quorum
  hadoop001.test.com,hadoop002.test.com,hadoop003.test.com
  
  
  
  hbase.master.port
  60000
  
  
  zookeeper.session.timeout
  60000
  
  
  hbase.zookeeper.property.clientPort
  2181
  
  
  
  hbase.zookeeper.property.dataDir
  /home/hadoop/storage/zookeeper/data
  
  
  ###########################################################################
  禁用hbase自带的zookeeper
  vi /home/hadoop/hbase/conf/hbase-env.sh
  export HBASE_MANAGES_ZK=false
  同步hbase工程到hadoop002~005机器上
  scp -r /home/hadoop/hbase hadoop002.test.com:/home/hadoop
  scp -r /home/hadoop/hbase hadoop003.test.com:/home/hadoop
  scp -r /home/hadoop/hbase hadoop004.test.com:/home/hadoop
  scp -r /home/hadoop/hbase hadoop005.test.com:/home/hadoop
  ====================现在不执行,统一启动=========================
  注意!!必须在hadoop集群启动后再启动hbase集群
  方法一、
  主hmaster节点上
  /home/hadoop/hbase/bin/start-hbase.sh
  备hmaster节点上
  /home/hadoop/hbase/bin/hbase-daemon.sh start master
  验证安装成功
  hbase shell
  list
  方法二、
  添加配置文件
  vi /home/hadoop/hbase/conf/backup-masters
  hadoop002.test.com
  主master节点启动
  /home/hadoop/hbase/bin/start-hbase.sh
  ================================================================
  五、部署hive-1.1.0的hiveserver2 HA
  解压部署
  tar xf apache-hive-1.1.0-bin.tar.gz -C /home/hadoop/
  mv apache-hive-1.1.0-bin hive
  修改配置文件
  cd /home/hadoop/hive/conf/
  cp hive-default.xml.template hive-site.xml
  vi hive-site.xml
  ###########################################################################################
  
  
  
  
  
  hive.metastore.warehouse.dir
  /home/hadoop/storage/hive/warehouse
  
  
  
  javax.jdo.option.ConnectionURL
  jdbc:mysql://192.168.5.2:3306/hive?createData baseIfNotExist=true
  JDBC connect string for a JDBC metastore
  
  
  
  javax.jdo.option.ConnectionDriverName
  com.mysql.jdbc.Driver

  Driver>  
  
  
  javax.jdo.option.ConnectionUserName
  hive
  username to use against metastore database
  
  
  
  javax.jdo.option.ConnectionPassword
  hive
  password to use against metastore database
  
  
  
  hive.hwi.war.file
  lib/hive-hwi-1.1.0.war
  
  
  
  hive.metastore.local
  false
  
  
  
  hive.metastore.uris
  thrift://192.168.5.2:9083
  
  
  
  hive.zookeeper.quorum
  hadoop001.test.com,hadoop002.test.com,hadoop003.test.com
  
  
  ###########################################################################################
  添加mysql驱动
  cp mysql-connector-java-5.1.35.jar /home/hadoop/hive/lib/
  添加hive web页面的war包
  下载hive源码包,进入hwi/web
  jar cvf hive-hwi-1.1.0.war ./*
  cp hive-hwi-1.1.0.war  /home/hadoop/hive/lib/
  从hbase/lib下复制必要jar包到hive/lib下
  cp /home/hadoop/hbase/lib/hbase-client-1.0.0.jar /home/hadoop/hbase/lib/hbase-common-1.0.0.jar /home/hadoop/hive/lib
  同步hive和hadoop的jline版本
  查看版本
  find ./ -name "*jline*jar"
  cp /home/hadoop/hive/lib/jline-2.12.jar /home/hadoop/hadoop/share/hadoop/yarn/lib
  删除低版本的jline 0.9
  复制jdk的tools.jar到hive/lib下
  cp $JAVA_HOME/lib/tools.jar /home/hadoop/hive/lib
  在mysql节点
  mysql创建hive用户,密码hive

  CREATE USER hive>  GRANT ALL PRIVILEGES ON *.* TO 'hive'@'%' WITH GRANT OPTION;
  flush privileges;
  用创建的hive用户登录mysql并创建hive库
  mysql -uhive -phive
  create database hive;
  ====================现在不执行,统一启动=========================
  Hive的service
  1、 hive 远程服务 (端口号10000) 启动方式
  /home/hadoop/hive/bin/hive --service metastore &
  /home/hadoop/hive/bin/hive --service hiveserver2 &【或 /home/hadoop/hive/bin/hiveserver2 &】
  客户端调用
  !connect jdbc:hive2://xxxxx:2181,xxxx:2181/;serviceDiscoveryMode=zookeeper user pass
  2、hive 命令行模式
  /home/hadoop/hive/bin/hive
  或者输入
  hive --service cli
  3、 hive web界面的 (端口号9999) 启动方式
  /home/hadoop/hive/bin/hive --service hwi&
  用于通过浏览器来访问hive
  http://hadoop001.test.com:9999/hwi/
  ###########################################################################################
  Hadoop集群首次启动过程
  ###########################################################################################
  1.如果zookeeper集群还没有启动的话, 首先把各个zookeeper起来。
  /home/hadoop/zookeeper/bin/zkServer.sh start    (记住所有的zookeeper机器都要启动)
  /home/hadoop/zookeeper/bin/zkServer.sh status (1个leader,n-1个follower)
  输入jps,会显示启动进程:QuorumPeerMain
  2.、然后在主namenode节点执行如下命令,创建命名空间
  /home/hadoop/hadoop/bin/hdfs zkfc -formatZK   (复制命令,-会变成— 建议手敲)
  3、在每个节点用如下命令启日志程序
  /home/hadoop/hadoop/sbin/hadoop-daemon.sh start journalnode
  (每个journalnode节点都需要启动)
  4、在主namenode节点用./bin/hadoop namenode -format格式化namenode和journalnode目录
  /home/hadoop/hadoop/bin/hadoop namenode -format
  验证成功
  在zookeeper节点执行
  /home/hadoop/zookeeper/bin/zkCli.sh
  ls /
  ls /hadoop-ha
  5、在主namenode节点启动namenode进程
  /home/hadoop/hadoop/sbin/hadoop-daemon.sh start namenode
  6、在备namenode节点执行第一行命令,把备namenode节点的目录格式化并把元数据从主namenode节点copy过来,并且这个命令不会把journalnode目录再格式化了!然后用第二个命令启动备namenode进程!
  /home/hadoop/hadoop/bin/hdfs namenode -bootstrapStandby【或者直接scp -r /home/hadoop/storage/hadoop/name hadoop002.test.com:/home/hadoop/storage/hadoop】
  /home/hadoop/hadoop/sbin/hadoop-daemon.sh start namenode
  7、在两个namenode节点都执行以下命令
  /home/hadoop/hadoop/sbin/hadoop-daemon.sh start zkfc
  8、启动datanode
  方法一、
  在所有datanode节点都执行以下命令启动datanode
  /home/hadoop/hadoop/sbin/hadoop-daemons.sh start datanode
  方法二、
  启动datanode节点多的时候,可以直接在主NameNode(nn1)上执行如下命令一次性启动所有datanode
  /home/hadoop/hadoop/sbin/hadoop-daemons.sh start datanode
  9. 主namenode节点启动Yarn
  /home/hadoop/hadoop/sbin/start-yarn.sh
  10.启动hbase
  方法一、
  主hmaster节点上
  /home/hadoop/hbase/bin/start-hbase.sh
  备hmaster节点上
  /home/hadoop/hbase/bin/hbase-daemon.sh start master
  验证安装成功
  hbase shell
  list
  方法二、
  添加配置文件
  vi /home/hadoop/hbase/conf/backup-masters
  hadoop002.test.com
  主master节点启动
  /home/hadoop/hbase/bin/start-hbase.sh
  11.启动hive
  hive 远程服务 (端口号10000) 启动方式
  /home/hadoop/hive/bin/hive --service metastore &
  /home/hadoop/hive/bin/hive --service hiveserver2 &【或 /home/hadoop/hive/bin/hiveserver2 &】
  客户端调用
  !connect jdbc:hive2://xxxxx:2181,xxxx:2181/;serviceDiscoveryMode=zookeeper user pass
  hive 命令行模式
  /home/hadoop/hive/bin/hive
  或者输入
  hive --service cli
  hive web界面的 (端口号9999) 启动方式
  /home/hadoop/hive/bin/hive --service hwi&
  用于通过浏览器来访问hive
  http://hadoop001.test.com:9999/hwi/
  ##########################################################################################
  参考资料:
  ==========================================================================================
  【Hdfs的NameNode HA、Yarn的Resouce Manager HA】
  HBase+ZooKeeper+Hadoop2.6.0的ResourceManager HA集群高可用配置
  http://www.aboutyun.com/thread-11909-1-1.html
  【Hbase HA】
  http://www.cnblogs.com/junrong624/p/3580477.html
  ***linux/Hbasegaokeyong_backup_master_569269_1377277861.html
  【Hive HA】
  Hive HA使用说明及Hive使用HAProxy配置HA(高可用)
  http://www.aboutyun.com/thread-10938-1-1.html
  Hive安装及使用攻略
  http://blog.fens.me/hadoop-hive-intro/
  Hive metastore三种配置方式
  http://blog.csdn.net/reesun/article/details/8556078
  Hive学习之HiveServer2服务端配置与启动
  http://www.aboutyun.com/thread-12278-1-1.html
  Hive学习之HiveServer2 JDBC客户端
  http://blog.csdn.net/skywalker_only/article/details/38366347
  Hive内置服务介绍
  http://www.aboutyun.com/thread-7438-1-1.html
  使用Hive命令行及内置服务
  http://www.aboutyun.com/thread-12280-1-1.html
  Hive配置文件中配置项的含义详解(收藏版)
  http://www.aboutyun.com/thread-7548-1-1.html
  hbase0.96与hive0.12整合高可靠文档及问题总结
  http://www.aboutyun.com/thread-7881-1-1.html
  【hadoop系列】
  http://www.cnblogs.com/junrong624/category/537234.html
  【其他】
  HBase 默认配置说明
  http://www.aboutyun.com/thread-7914-1-1.html
  HBASE启动脚本/Shell解析
  http://zjushch.iteye.com/blog/1736065
  全面了解hive
  http://www.aboutyun.com/thread-7478-1-1.html
  Hadoop添加删除节点
  http://my.oschina.net/MrMichael/blog/291802
  粉丝日志
  http://blog.fens.me/
  【备注】
  启动hadoop jobhistory历史服务器
  /home/hadoop/hadoop/sbin/mr-jobhistory-daemon.sh start historyserver
  查看namenode状态
  hdfs haadmin -getServiceState


运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-628294-1-1.html 上篇帖子: Python菜鸟的Hadoop实战——Hadoop2.6.0集群搭建 下篇帖子: Python菜鸟的Hadoop实战——Hadoop2.6.0中的Yarn
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表