(1)hadoop2.7.1源码编译 | http://aperise.iteye.com/blog/2246856 |
(2)hadoop2.7.1安装准备 | http://aperise.iteye.com/blog/2253544 |
(3)1.x和2.x都支持的集群安装 | http://aperise.iteye.com/blog/2245547 |
(4)hbase安装准备 | http://aperise.iteye.com/blog/2254451 |
(5)hbase安装 | http://aperise.iteye.com/blog/2254460 |
(6)snappy安装 | http://aperise.iteye.com/blog/2254487 |
(7)hbase性能优化 | http://aperise.iteye.com/blog/2282670 |
(8)雅虎YCSBC测试hbase性能测试 | http://aperise.iteye.com/blog/2248863 |
(9)spring-hadoop实战 | http://aperise.iteye.com/blog/2254491 |
(10)基于ZK的Hadoop HA集群安装 | http://aperise.iteye.com/blog/2305809 |
1.hadoop环境安装
hbase是建立在Hadoop之上的一个NOSQL列存储数据库,所以你得首先安装Hadoop,安装Hadoop参见http://aperise.iteye.com/blog/2245547
2.zookeeper3.4.6集群安装
1,hbase regionserver 向zookeeper注册,提供hbase regionserver状态信息(是否在线)
2,hmaster启动时候会将hbase 系统表-ROOT- 加载到 zookeeper cluster,通过zookeeper cluster可以获取当前系统表.META.的存储所对应的regionserver信息。
3.zookeeper集群节点最少3台,这样才能满足选举规则,少数服从多数
1) 下载解压
下载zookeeper-3.4.6.tar.gz到/opt
解压tar -zxvf zookeeper-3.4.6.tar.gz
2) 配置/etc/hosts
# that require network functionality will fail.
127.0.0.1 localhost.localdomain localhost
::1 localhost6.localdomain6 localhost6
192.168.186.234 inddb1-vip
192.168.186.235 inddb2-vip
192.168.88.22 nmsc2
192.168.88.21 nmsc1
192.168.181.66 nmsc0
3) 创建zookeeper数据文件
#zookeeper每台机器上都要创建
sudo rm -r /home/hadoop/zookeeper
cd /home/hadoop
mkdir zookeeper
4) 配置zoo.cfg
将zookeeper-3.3.4/conf目录下面的 zoo_sample.cfg复制一份为zoo.cfg,配置文件内容如下所示
tickTime=2000
# The number of ticks that the initial
# synchronization phase can take
initLimit=10
# The number of ticks that can pass between
# sending a request and getting an acknowledgement
syncLimit=5
# the directory where the snapshot is stored.
# do not use /tmp for storage, /tmp here is just
# example sakes.
dataDir=/home/hadoop/zookeeper
# the port at which the clients will connect
clientPort=2181
# the maximum number of client connections.
# increase this if you need to handle more clients
#maxClientCnxns=60
#
# Be sure to read the maintenance section of the
# administrator guide before turning on autopurge.
#
# http://zookeeper.apache.org/doc/current/zookeeperAdmin.html#sc_maintenance
#
# The number of snapshots to retain in dataDir
#数据文件保存最近的3个快照,默认是都保存,时间长的话会占用很大磁盘空间
# Purge task interval in hours
# Set to "0" to disable auto purge feature
#单位为小时,每小时清理一次快照数据
server.1=nmsc0:2888:3888
server.2=nmsc1:2888:3888
server.3=nmsc2:2888:3888
5) 远程复制分发安装文件到其它zookeeper集群节点
scp -r /opt/zookeeper-3.3.4 root@nmsc1:/opt/
scp -r /opt/zookeeper-3.3.4 root@nmsc2:/opt/
6) 设置myid必须为整数
#zookeeper集群节点之一
ssh nmsc0
echo "1" > /home/hadoop/zookeeper/myid
##zookeeper集群节点之二
ssh nmsc1
echo "2" > /home/hadoop/zookeeper/myid
##zookeeper集群节点之三
ssh nmsc2
echo "3" > /home/hadoop/zookeeper/myid
7) 启动ZooKeeper集群
cd /opt/zookeeper-3.4.6
bin/zkServer.sh start
8) 查看单机ZooKeeper是leader还是follower
cd /opt/zookeeper-3.4.6
bin/zkServer.sh status
9) 停止ZooKeeper集群
cd /opt/zookeeper-3.4.6
bin/zkServer.sh stop
10) 用客户端查看zookeeper上数据
客户端链接zookeeper服务端
cd /opt/zookeeper-3.4.6/ bin/zkCli.sh -server nmsc0:2181,nmsc2:2181,nmsc1:2181
登录后提示信息如下:
[root@nmsc1 bin]# ./zkCli.sh -server nmsc0:2181,nmsc2:2181,nmsc1:2181 Connecting to nmsc0:2181,nmsc2:2181,nmsc1:2181 2016-04-19 10:36:20,475 [myid:] - INFO [main:Environment@100] - Client environment:zookeeper.version=3.4.6-1569965, built on 02/20/2014 09:09 GMT 2016-04-19 10:36:20,479 [myid:] - INFO [main:Environment@100] - Client environment:host.name=nmsc1 2016-04-19 10:36:20,479 [myid:] - INFO [main:Environment@100] - Client environment:java.version=1.7.0_65 2016-04-19 10:36:20,481 [myid:] - INFO [main:Environment@100] - Client environment:java.vendor=Oracle Corporation 2016-04-19 10:36:20,481 [myid:] - INFO [main:Environment@100] - Client environment:java.home=/opt/java/jdk1.7.0_65/jre 2016-04-19 10:36:20,481 [myid:] - INFO [main:Environment@100] - Client environment:java.class.path=/opt/zookeeper-3.4.6/bin/../build/classes:/opt/zookeeper-3.4.6/bin/../build/lib/*.jar:/opt/zookeeper-3.4.6/bin/../lib/slf4j-log4j12-1.6.1.jar:/opt/zookeeper-3.4.6/bin/../lib/slf4j-api-1.6.1.jar:/opt/zookeeper-3.4.6/bin/../lib/netty-3.7.0.Final.jar:/opt/zookeeper-3.4.6/bin/../lib/log4j-1.2.16.jar:/opt/zookeeper-3.4.6/bin/../lib/jline-0.9.94.jar:/opt/zookeeper-3.4.6/bin/../zookeeper-3.4.6.jar:/opt/zookeeper-3.4.6/bin/../src/java/lib/*.jar:/opt/zookeeper-3.4.6/bin/../conf:.:/opt/java/jdk1.7.0_65/jre/lib/rt.jar:/opt/java/jdk1.7.0_65/lib/dt.jar:/opt/java/jdk1.7.0_65/lib/tools.jar 2016-04-19 10:36:20,481 [myid:] - INFO [main:Environment@100] - Client environment:java.library.path=:/opt/jprofiler7/bin/linux-x64:/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib 2016-04-19 10:36:20,481 [myid:] - INFO [main:Environment@100] - Client environment:java.io.tmpdir=/tmp 2016-04-19 10:36:20,482 [myid:] - INFO [main:Environment@100] - Client environment:java.compiler=<NA> 2016-04-19 10:36:20,482 [myid:] - INFO [main:Environment@100] - Client environment:os.name=Linux 2016-04-19 10:36:20,482 [myid:] - INFO [main:Environment@100] - Client environment:os.arch=amd64 2016-04-19 10:36:20,482 [myid:] - INFO [main:Environment@100] - Client environment:os.version=2.6.18-308.el5 2016-04-19 10:36:20,482 [myid:] - INFO [main:Environment@100] - Client environment:user.name=root 2016-04-19 10:36:20,482 [myid:] - INFO [main:Environment@100] - Client environment:user.home=/root 2016-04-19 10:36:20,482 [myid:] - INFO [main:Environment@100] - Client environment:user.dir=/opt/zookeeper-3.4.6/bin 2016-04-19 10:36:20,484 [myid:] - INFO [main:ZooKeeper@438] - Initiating client connection, connectString=nmsc0:2181,nmsc2:2181,nmsc1:2181 sessionTimeout=30000 watcher=org.apache.zookeeper.ZooKeeperMain$MyWatcher@8a61d64 Welcome to ZooKeeper! 2016-04-19 10:36:20,511 [myid:] - INFO [main-SendThread(nmsc2:2181):ClientCnxn$SendThread@975] - Opening socket connection to server nmsc2/192.168.88.22:2181. Will not attempt to authenticate using SASL (unknown error) 2016-04-19 10:36:20,515 [myid:] - INFO [main-SendThread(nmsc2:2181):ClientCnxn$SendThread@852] - Socket connection established to nmsc2/192.168.88.22:2181, initiating session JLine support is enabled 2016-04-19 10:36:20,528 [myid:] - INFO [main-SendThread(nmsc2:2181):ClientCnxn$SendThread@1235] - Session establishment complete on server nmsc2/192.168.88.22:2181, sessionid = 0x353a8f6e9547a07, negotiated timeout = 30000 WATCHER:: WatchedEvent state:SyncConnected type:None path:null
查看根目录下数据节点
[zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 0] ls / [isr_change_notification, hbase, zookeeper, admin, consumers, config, controller, kafka-manager, brokers, controller_epoch] [zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 1] ls /controller [] [zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 2] ls /hbase [meta-region-server, backup-masters, table, draining, region-in-transition, table-lock, running, master, namespace, hbaseid, online-snapshot, replication, splitWAL, recovering-regions, rs, flush-table-proc] [zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 3] ls /zookeeper [quota] [zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 4] ls /brokers [seqid, topics, ids] [zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 5] ls /brokers/topics [signal2015, __consumer_offsets] [zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 6] ls /brokers/topics/signal2015 [partitions] [zk: nmsc0:2181,nmsc2:2181,nmsc1:2181(CONNECTED) 7]
3.时间同步
注意:集群中机器时间同步很关键,时间最好相差在1秒以下,不然hbase启动会失败
方法一
安装NTP服务,参见http://wenku.baidu.com/link?url=0outiNUpCi-RzCtmWCKgqyVE0S_hyEcc3cjfbes_XHtxDN4wCh3-TXAe8YrtW1iOudzloNbMRMFe6niTS_uODhWEdPjD7MEwimib1a02FVm
方法二
我这是公司内部有时间同步的机器,直接从该机器同步数据,执行命令如下:
#从机器192.168.162.23同步系统时间,可以采用linux定时任务每天凌晨执行一次
ntpdate -u 192.168.162.23
相关推荐
NULL 博文链接:https://aperise.iteye.com/blog/2282670
hadoop2.7.1+hbase2.1.4+zookeeper3.6.2集合
描述了centOS6.5(虚拟机)环境下,hadoop2.7.1+hbase1.3.5版本的安装配置过程,以及个人在安装过程中遇到的问题与总结
因为配置大数据的基础环境特别费事,因此这里搭建好了一份基础环境
主要介绍了详解VMware12使用三台虚拟机Ubuntu16.04系统搭建hadoop-2.7.1+hbase-1.2.4(完全分布式),具有一定的参考价值,有兴趣的可以了解一下。
安装HBase时,要考虑选择正确的Hadoop版本,否则可能出现不兼容的情况。 hadoop-common-2.7.1-bin-master.zip
Hadoop_Hbase集群安装部署手册 本系统操作系统使用的是hadoop2.7.1,操作系统为centos 6.5 64位
eclipse链接hbase所需jar包,hbase版本1.2.6,Hadoop版本2.7.1(转载的另一篇下载的,亲测可用)
eclipse链接hbase所需jar包,hbase版本1.2.6,Hadoop版本2.7.1
A.3实验三:熟悉常用的HBase操作 本实验对应第5章的内容。 A.3.1 实验目的 ...(2)Hadoop 版本:2.7.1或以上版本。(3)HBase版本:1.1.2或以上版本。(4) JDK 版本:1.7或以上版本。(5) Java IDE:Eclipse。
hadoop-mysql-hbase环境部署套装包括: hadoop-2.7.1.tar.gz hbase-1.1.5-bin.tar.gz jdk-8u162-linux-x64.tar.gz mysql-connector-java-5.1.40.tar.gz mysql-server_5.7.21-1ubuntu14.04_amd64.deb-bundle.tar
由于网上下的不支持最新的hadoop,hbase 版本自己稍微修改了下,支持最新版本HBase的图形化管理工具,目前修改改为hadoop-2.7.1版本,hbase-1.1.2版本,依赖可以自己切换,源代码已经包括再里边了,如想修改直接修改...
mysql-8.0.19-1.el7.x86_64.rpm-bundle.tar\apache-hive-2.0.0-bin.tar.gz\spark-2.4.5-bin-hadoop2.7.tgz\hbase-1.2.1-bin.tar.gz\hadoop-2.7.1.tar.gz\jdk-8u162-linux-x64.tar.gz
解压hadoop2.7.1.tar.gz文件,文件位置根据情况而定,本机文件正好放在根目录下,所以没有带路径,请一定根据自己文件所在路径把相应路径加上,进行解压,/usr/local为目标路径,也是根据情况而定,因人或者要求而定
本书是Hadoop权威参考,程序员可从中探索如何分析海量数据集,管理员可以从中了解如何安装与运行Hadoop集群。 目录 第1章 初识Hadoop 数据!数据! 数据存储与分析 与其他系统相比 关系型数据库管理系统...
第4章 Hadoop I/O 数据完整性 HDFS的数据完整性 LocalFileSystem ChecksumFileSystem 压缩 codec 压缩和输入切分 在MapReduce中使用压缩 序列化 Writable接口 Writable类 实现定制的Writable类型 序列化框架 Avro ...
jdk-8u25-linux-x64.tar.gz hadoop-2.7.1.tar.gz zookeeper-3.4.6.tar.gz hbase-1.1.5-bin.tar.gz mysql-connector-java-5.1.27-bin.jar apache-hive-1.2.1-bin.tar.gz
(2)Hadoop版本:2.7.1。 (3)MySQL版本:5.7.15。(4)HBase版本:1.1.2。(5)Redis版本:3.0.6。 332 大数据基础编程、实验和案例教程 (6)MongoDB版本:3.2.6。(7)JDK版本:1.7或以上版本。(8)Java IDE:Eclipse。