这部分请参考我的博文:Hadoop数据存储
使用root用户设置,完成需要重新登录。
vim /etc/security/limits.conf
root soft nproc 50000 root hard nproc 50000 root soft nofile 25535 root hard nofile 25535 hadoop soft nproc 50000 hadoop hard nproc 50000 hadoop soft nofile 25535 hadoop hard nofile 25535vim /etc/sysctl.conf
net.ipv4.ip_forward= 0 net.ipv4.conf.default.rp_filter= 1 net.ipv4.conf.default.accept_source_route= 0 kernel.core_users_pid= 1 net.ipv4.tcp_syncookies= 1 net.bridge.bridge-nf-call-ip6tables= 0 net.bridge.bridge-nf-call-iptables= 0 net.bridge.bridge-nf-call-arptables= 0 kernel.mggmnb= 65536 kernel.mggmax= 65536 kernel.shmmax= 68719476736 kernel.shmall= 268435456 net.ipv4.tcp_max_syn_backlog= 65000 net.core.netdev_max_backlog= 32768 net.core.somaxconn= 32768 fs.file-max= 65000 net.core.wmem_default= 8388608 net.core.rmem_default= 8388608 net.core.rmem_max= 16777216 net.core.wmem_max= 16777216 net.ipv4.tcp_timestamps= 1 net.ipv4.tcp_synack_retries= 2 net.ipv4.tcp_syn_retries= 2 net.ipv4.tcp_mem= 94500000 915000000 927000000 net.ipv4.tcp_max_orphans= 3276800 net.ipv4.tcp_tw_reuse= 1 net.ipv4.tcp_tw_recycle= 1 net.ipv4.tcp_keepalive_time= 1200 net.ipv4.tcp_syncookies= 1 net.ipv4.tcp_fin_timeout= 10 net.ipv4.tcp_keepalive_intvl= 15 net.ipv4.tcp_keepalive_probes= 3 net.ipv4.ip_local_port_range= 1024 65535 net.ipv4.conf.eml.send_redirects= 0 net.ipv4.conf.lo.send_redirects= 0 net.ipv4.conf.default.send_redirects= 0 net.ipv4.conf.all.send_redirects= 0 net.ipv4.icmp_echo_ignore_broadcasts= 1 net.ipv4.conf.eml.accept_source_route= 0 net.ipv4.conf.lo.accept_source_route= 0 net.ipv4.conf.default.accept_source_route= 0 net.ipv4.conf.all.accept_source_route= 0 net.ipv4.icmp_ignore_bogus_error_responses= 1 kernel.core_pattern= /tmp/core vm.overcommit_memory= 1sysctl -p生效
配置JDK安装路径
exportJAVA_HOME=/home/hadoop/app/jdk1.7.0_79配置Hadoop安装路径
exportHADOOP_HOME=/home/hadoop/hadoop2.7.2设置HBase的日志目录
exportHBASE_LOG_DIR=${HBASE_HOME}/logs设置HBase的pid目录
exportHBASE_PID_DIR=${HBASE_HOME}/pids使用独立的ZooKeeper集群
exportHBASE_MANAGES_ZK=false优化配置项
设置HBase内存堆的大小
exportHBASE_HEAPSIZE=1024设置HMaster最大可用内存
exportHBASE_MASTER_OPTS="-Xmx512m"设置HRegionServer最大可用内存
exportHBASE_REGIONSERVER_OPTS="-Xmx1024m"vim regionservers
slave1
slave2
slave3
$ vim backup-masters
master2
$ cd /home/hadoop/hbase-0.98.20/
$ mkdir tmp
$ cd /home/hadoop/hbase-0.98.20/
$ mkdir logs
$ cd /home/hadoop/hbase-0.98.20/
$ mkdir pids
$ scp -r /home/hadoop/hbase-0.98.20/ master2:/home/hadoop/
$ scp -r /home/hadoop/hbase-0.98.20/ slave1:/home/hadoop/
$ scp -r /home/hadoop/hbase-0.98.20/ slave2:/home/hadoop/
$ scp -r /home/hadoop/hbase-0.98.20/ slave3:/home/hadoop/
先启动zookeeper,再Hadoop,最后HBase。
(一)设置新节点与集群NameNode的SSH无密码登录
(二)在hosts文件中添加主机名和IP地址,并将该文件同步到集群其它节点
(三)修改hadoop配置文件slaves,加入新节点的主机名
(四)在新节点通过命令hadoop-daemon.shstart datanode和yarn-daemon.sh start nodemanager启动datanode和nodemanager进程
(五)在新节点中通过命令start-balancer.sh均衡当前的HDFS块,在执行前可通过命令Hdfs dfsadmin -setBalancerBandwidth字节数 设置带宽
(一)在hadoop配置文件hdfs-site.xml中配置如下:
<property>
<name>dfs.hosts.exclude</name>
<value>/home/hadoop/hadoop-2.7.2/etc/hadoop/exclude</value>
</property>
(二)在配置文件exclude中添加需要删除的节点主机名
(三)通过命令hdfsafsadmin -refreshNodes执行节点刷新操作
(四)通过命令hdfsafsadmin -report可查看该节点状况
(一)在HBase的配置文件regionserver中添加新节点的主机名
(二)在新节点通过命令hbase-daemon.shstart regionserver启动HRegionServer
(三)进入hbaseshell,通过命令balance_switch true进行region平衡(二)在HBase的配置文件regionserver中移除删除节点的主机名