入门指南
1. 简介
Quickstart 会让你启动和运行一个单节点单机HBase。
2. 快速启动 – 单点HBase
这部分描述单节点单机HBase的配置。一个单例拥有所有的HBase守护线程—Master,RegionServers和ZooKeeper,运行一个单独JVM持久化到本地文件系统。这是我们最基础的部署文档。我们将会向你展示如何通过hbase shell CLI在HBase中创建一个表格,在表中插入行,执行put和scan操作,让表使能和启动和停止HBase等等操作。
除了下载HBase,这个过程大概需要不到10分钟地时间。
HBase 0.94.x之前的版本希望回送IP地址为127.0.0.1,而UBuntu和其他发行版默认是127.0.1.1,这将会给你造成麻烦。查看 Why does HBase care about /etc/hosts? 获得更多细节 在Ubuntu上运行0.94.x之前版本的HBase,/etc/hosts文档应该以下面所写的模板来保证正常运行 127.0.0.1 localhost 127.0.0.1 ubuntu.ubuntu-domain ubuntu hbase-0.96.0版本之后的已经修复了。 |
2.1. JDK 版本要求
HBase 需要安装JDK。查看 Java 来获得每个HBase版本所支持的JDK版本。
2.2. 开始使用 HBase
过程: 下载, 配置, 和启动单机模式HBase
- 从 Apache Download Mirrors列表中选一个下载节点。点击显示的链接。这将会带你到一个HBase发布版本的镜像。点击名字为stable的文件夹然后下载文件结尾为.tar.gz的二进制文件到你的本地文件系统中。不要下载文件结尾为src.tar.gz的文件。
- 提取下载文件并且将它放到新建的目录。
- $ tar xzvf hbase-2.0.0-SNAPSHOT-bin.tar.gz
- $ cd hbase-2.0.0-SNAPSHOT/
- 你需要在启动HBase之前设置好JAVA_HOME环境变量,你可以通过你的操作系统常用方法来设置这个变量,但是HBase提供了一种中央机制,conf/hbse-env.sh。编辑这个文档,将JAVA-HOME这一行的注释给取消,然后将他的值设为你的操作系统中JAVA的安装路径。JAVA_HOME变量应该设置包含可执行文件bin/java的路径。大多数现代的Linux操作系统提供一种机制,例如在RHEL或者CentOS是/usr/bin/alternatives,为了能够显示地切换Java版本。在这种情况,你可以在设置JAVA_HOME为包含bin/java符号链接的目录,通常是/usr。
JAVA_HOME=/usr
6. 编辑conf/hbase-site.xml,该文档是HBase配置文件。在这个时间点你只需要在本地文件系统中指定HBase和ZooKeeper写数据的目录。默认情况下,会在/tmp目录下创建一个新目录。许多服务器会配置为一旦reboot那么会删除/tmp目录下的内容,所以你应该在别的地方存储数据。接下来的配置将会存储HBase的数据在hbase目录下,放在用户testuser的主目录下。新安装的HBase下<configuration> 标签里面的内容是空,粘贴 <property>标签到<configuration> 下进行配置。
Example 1. Example hbase-site.xml for Standalone HBase
<configuration>
<property>
<name>hbase.rootdir</name>
<value>file:///home/testuser/hbase</value>
</property>
<property>
<name>hbase.zookeeper.property.dataDir</name>
<value>/home/testuser/zookeeper</value>
</property>
</configuration>
你不需要创建HBase数据目录。HBase将会为你创建。如果你自己创建了,HBase将会试图一个你并不想要的迁移。
上面例子中hbase.rootdir 指向本地文件系统的目录。我们用‘file:/’前缀来表示本地文件系统。将HBase的home目录配置在已有的HDFS实例上,设置hbase.rootdir 指向你的HDFS实例,例如 hdfs://namenode.example.org:8020/hbase.关于这个变量的细节,请查看下面在HDFS上部署单机HBase部分。 |
7. bin/start-hbase.sh脚本将提供一个简便的方式来启动HBase。发出这个命令并且运行良好的话,一条标准的成功启动的信息会打印在控制台上。你可以通过jps命令来判断你是否已经运行一个HMaster进程。在单价模式下,HBase会在这个单独的JVM中启动HMater,HRegionServer和ZooKeeper守护进程。在 http://localhost:16010查看HBase WebUI 。
需要安装Java并且使之可用。如果你已经安装了,但是却报错提示你尚未安装,可能安装在一个非标准路径下,编辑conf/hbase-env.sh并且修改JAVA_HOME,将包含bin/java的目录赋给它 |
过程: 首次使用HBase
1)连接HBase
使在你HBase安装目录下的bin/ 下用hbase shell命令行来连接HBase。在这个例子中,会打印一些你在启动的HBase shell用时遗漏的用法和版本信息。HBase Shell用>符号来表示结束。
$ ./bin/hbase shell
hbase(main):001:0>
2)显示HBase帮助文本
输出help按下Enter,显示HBase Shell的基础使用信息,以及一些示例命令。需要注意的是表名,行,列都必须用引用符号。
- 创建表
使用create命令来创建一个新表。你必须指定表名和列族名
hbase(main):001:0> create 'test', 'cf'
0 row(s) in 0.4170 seconds
=> Hbase::Table - test
- 列出表的信息
使用list命令
hbase(main):002:0> list 'test'
TABLE
test
1 row(s) in 0.0180 seconds
=> ["test"]
- 插入数据到表中
使用put命令来插入数据.
hbase(main):003:0> put 'test', 'row1', 'cf:a', 'value1'
0 row(s) in 0.0850 seconds
hbase(main):004:0> put 'test', 'row2', 'cf:b', 'value2'
0 row(s) in 0.0110 seconds
hbase(main):005:0> put 'test', 'row3', 'cf:c', 'value3'
0 row(s) in 0.0100 seconds
在这里,我们每次插入一条数据,总共三条。第一次将value1插入到row1,列cf:a中。HBase中的列以列族名为前缀,例子中是cf,后面是冒号和列的限定符后缀,例子中是a。
- 一次查看所有数据
使用scan命令行来扫描表格的数据。你可以限制你的扫描,但是此时此刻,所有的数据都被获取了。
hbase(main):006:0> scan 'test'
ROW COLUMN+CELL
row1 column=cf:a, timestamp=1421762485768, value=value1
row2 column=cf:b, timestamp=1421762491785, value=value2
row3 column=cf:c, timestamp=1421762496210, value=value3
3 row(s) in 0.0230 seconds
- 获得一行的数据
用get命令一次获取一行数据
hbase(main):007:0> get 'test', 'row1'
COLUMN CELL
cf:a timestamp=1421762485768, value=value1
1 row(s) in 0.0350 seconds
- 禁用表
如果你想要删除一个表或者改变它的配置,以及其他一些情况,你首先需要用disable命令来禁用表。
hbase(main):008:0> disable 'test'
0 row(s) in 1.1820 seconds
hbase(main):009:0> enable 'test'
0 row(s) in 0.1770 seconds
在启用‘test’之后再次禁用‘test’
hbase(main):010:0> disable 'test'
0 row(s) in 1.1820 seconds
- 删除表
用drop命令来删除表
hbase(main):011:0> drop 'test'
0 row(s) in 0.1370 seconds
- 退出HBase Shell
使用exit来与HBase断开连接,但HBase仍然在后台运行
过程: 关闭 HBase
跟bin/start-hbase.sh脚本一样方便地启动HBase,用bin/stop-hbase.sh脚本来停止它。
$ ./bin/stop-hbase.sh
stopping hbase....................
$
在发出这个命令之后,将花费几分钟的时间来关闭。使用jps来确保HMaster和HRegionServer已经关闭。
上面的内容已经向你展示了如何启动和停止一个单机HBase。在下一部分我们将提供其他模式的部署。
2.3. 伪分布式本地安装
在通过 quickstart 启动了单机模式之后,你可以重新配置来运行伪分布式模式。伪分布式模式意味着HBase仍然运行在一个节点上,但是每个HBase的守护进程(HMaster, HRegionServer, and ZooKeeper)运行在单独的进程中:在单机模式中所有的守护进程都运行在一个JVM实例中。默认情况下,除非你配置像 quickstart中所描述的配置 hbase.rootdir属性,你的数据仍然存储在/tmp/中。在这次演示中,我们将数据存储在HDFS中,确保你HDFS是可用的。你可以跳过HDFS配置继续将数据存储在本地文件系统中
Hadoop配置 这个过程假设你已经在本地系统或者远程系统中配置好Hadoop和HDFS,并且能够运行和确保可用。也假定你使用Hadoop2. Setting up a Single Node Cluster 将引导如何搭建单节点Hadoop |
1)如果HBase还在运行请停止它
如果你已经完成 quickstart 中的指导并且HBase仍然在运行,请停止他。这个过程将创建一个新的目录来储存它的数据,所以之前你创建的数据库将会丢失。
2)配置HBase
编辑hbase-site.xml 进行配置. 第一,添加下面 property来 指导 HBase运行分布式模式, 每个守护进程运行在一个JVM上。
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
</property>
接下来, 将hbase.rootdir由本地系统改为HDFS实例的地址, 使用 hdfs://// URI 语法. 在这个例子当中, HDFS 运行在端口 8020上.
<property>
<name>hbase.rootdir</name>
<value>hdfs://localhost:8020/hbase</value>
</property>
你不需要在HDFS上创建一个目录。HBase会自己创建。如果你自己创建了,HBase会试图做一些你并不想要的迁移。
3)启动HBase
使用 bin/start-hbase.sh 命令来启动HBase. 如果你的系统配置是正确的话,使用jps命令将会看到HMaster和HRegionServer已经运行。
4)检查HBase在HDFS中的目录
如果所有都运行正确的话,HBase将会在HDFS中创建它的目录。在上面的配置中,它将存储在HDFS的/hbase中。你可以在Hadoop的bin/下使用hadoop fs命令行来列出这个目录下的所有文件。
$ ./bin/hadoop fs -ls /hbase
Found 7 items
drwxr-xr-x - hbase users 0 2014-06-25 18:58 /hbase/.tmp
drwxr-xr-x - hbase users 0 2014-06-25 21:49 /hbase/WALs
drwxr-xr-x - hbase users 0 2014-06-25 18:48 /hbase/corrupt
drwxr-xr-x - hbase users 0 2014-06-25 18:58 /hbase/data
-rw-r--r-- 3 hbase users 42 2014-06-25 18:41 /hbase/hbase.id
-rw-r--r-- 3 hbase users 7 2014-06-25 18:41 /hbase/hbase.version
drwxr-xr-x - hbase users 0 2014-06-25 21:49 /hbase/oldWALs
5)创建一个表格并插入数据
你可以使用HBase Shell来创建一个表格,插入数据,扫描和获取数据,使用方法和 shell exercises所展示的一样。
6)启动和停止一个HMaster备用服务器
在同一个硬件环境上运行多个HMaster实例的情况不能出现在生产环境,同样伪分布式也是不允许的。这个步骤只适用于测试和学习 |
HMaster服务器控制HBase 集群。你可以启动9个HMaster服务器,那么10个HMaster一起执行计算。使用local-master-backup.sh来启动一个HMaster备用服务器。你想要启动的每个备用服务器都要添加一个代表master的端口参数。每个备用HMaster使用三个端口(默认是16010,16020,16030)端口都是以默认默认端口进行偏移的,偏移量为2的话,备用HMaster的端口会是16012,16022,16032。下面的指令用来启动3个端口分别为16012/16022/16032、 16013/16023/16033和16015/16025/16035的HMaster。
$ ./bin/local-master-backup.sh 2 3 5
想要杀掉一个备用master而不是关掉整个进程,你需要找到他的ID(PID)。PID存储在一个名字为/tmp/hbase-USER-X-master.pid的文件中。该文件里面的内容只有PID。你可以使用kill-9命令来杀掉PID。下面的命令杀掉端口为偏移量1的master,而集群仍然运行:
$ cat /tmp/hbase-testuser-1-master.pid |xargs kill -9
7)启动和停止另外的RegionServers
HRegionServer被HMaster指导管理它StoreFiles里的数据。通常来说,集群中的每个节点都运行一个HReigionServer,运行多个HRegionServer在同一系统当中可以用来测试伪分布式模式。使用local-regionservers.sh命令运行多个RegionServers。跟local-master-backup.sh一样,为每个实例提供端口偏移量。每个RegionServer需要两个端口,默认端口为16020和16030。然而,1.0.0版本的基本端口已经被HMaster所使用,所以RegionServer无法使用默认端口。所有基本端口改为16200和16300。你可以在一个服务中运行99额外RegionServer而不是一个HMaster或者HMaster。下面的命令用来启动端口从16202/16302开始连续的额外的RegionServer。
$ .bin/local-regionservers.sh start 2 3 4 5
使用local-regionservers.sh 命令 和要关闭的server的偏移量参数来手动停止RegionServer。
$ .bin/local-regionservers.sh stop 3
8)停止 HBase
你可以使用 quickstart 中阐述的命令 bin/stop-hbase.sh 来停止HBase。
2.4. 高级 – 全分布式
事实上,你需要一个全分布式的配置来测试完整的HBase并且将它用在真实世界的应用场景中。在一个分布式配置中,集群包括多个节点,每个节点运行一个或者多个HBase守护进程。这些包括主要的和备用Master实例,多个ZooKeeper节点和多个RegionServer节点。
这个高级配置比quickstart中多添加了两个节点,结构如下:
Table 1. Distributed Cluster Demo Architecture |
|||
Node Name |
Master |
ZooKeeper |
RegionServer |
node-a.example.com |
yes |
yes |
no |
node-b.example.com |
backup |
yes |
yes |
node-c.example.com |
no |
yes |
yes |
这个快速启动设定每个节点都是一个虚拟机而且他们在同样的网络上。它搭建在之前的quickstart和Pseudo-Distributed Local Install之上,设定你之前配置系统为node-a。在继续操作之前请停止HBase。
防火墙也应该关闭确保所节点都能够互相通信。如果你看到no route to host的报错,检查你的防火墙。 |
过程: 配置无密钥 SSH 登陆
node-a 需要登录到node-b和node-c来启动守护进程。最简单的实现方法是在所有的主机上使用相同用户名,配置无密钥SSH登陆。
1)在 node-a上生成密钥对
登陆那个要运行HBase的用户,使用下面命令生成一个SSH密钥对:
$ ssh-keygen -t rsa
如果该命令成功执行,那么密钥对的路径就会打印到标准输出。公钥的默认名字为 id_rsa.pub。
2)在其他节点创建用来储存密钥的路径。
在node-b和node-c,登陆HBase用户并且在用户的home目录下创建.ssh/目录,如果该目录不存在的话。如果已经存在,要意识到他可能已经包含其他密钥了。
3)复制密钥到其他节点
使用scp或者其他安全的方式将密钥安全地从node-a复制到其他每个节点上。每个节点上如果不存在 .ssh/authorized_keys 这个文件的话,那么创建一个,然后将id_rsa.pub 文件的内容添加到该文件末端。需要说明的是你需要在node-a做同样的操作。
$ cat id_rsa.pub >> ~/.ssh/authorized_keys
4)测试无密钥登陆.
如果一切运行顺利的话,那么你可以使用SSH用相同的用户名而不需要密钥的情况下登陆其他节点。
5)因为node-b将会运行一个备用Master,重复上述的过程,将能看到的node-a都换成node-b。确保不要覆盖已经存在的 .ssh/authorized_keys 的文档,但可以用>>符号将密钥追加到已存在的文档后面。
过程: 预备 node-a
node-a将会运行主master和ZooKeeper进程,但是没有RegionServers。在node-a将RegionServer停掉。
1)编辑conf/regionservers 和移除包含localhost的那一行。添加node-b和node-c的主机名和IP地址。
尽管你想要在node-a运行一个RegionServer,你应该给他指定一个主机名便于其他服务可以和它通讯。在这个例子当中,主机名为node-a.example.com。这使得你可以分布配置到集群每个节点来避免主机名冲突。保存文档。
2)将node-b配置为一个备用master。
所以在conf/目录下创建一个名为backup-master的新文件,然后添加一行node-b的主机名。在这个示例当中,主机名为node-b.example.com
3)配置ZooKeeper
事实上,你应该认真的配置你的ZooKeeper。你可以在 zookeeper找到更多关于ZooKeeper的细节。这个配置会指导HBase的启动和管理集群的每个节点中的ZooKeeper实例。
On node-a, edit conf/hbase-site.xml and add the following properties.
<property>
<name>hbase.zookeeper.quorum</name>
<value>node-a.example.com,node-b.example.com,node-c.example.com</value>
</property>
<property>
<name>hbase.zookeeper.property.dataDir</name>
<value>/usr/local/zookeeper</value>
</property>
4)在你的配置中把node-a配置为主机的地方改变指向主机名的引用以致其他节点可以使用它来代表node-a。在这个示例当中,主机名是node-a.example.com。
过程: 预备 node-b 和 node-c
node-b 将会运行一个备用master 服务器和一个ZooKeeper 实例.
1)下载和解压HBase.
在node-b下下载和解压HBase,跟你在quickstart和伪分布式中所做的一样。
2)从node-a复制配置信息到node-b和node-c
集群中的每个节点需要相同的配置信息。复制conf/下的内容到node-b和node-c下conf/。
过程: 启动和测试你的集群
1)确保任何节点上没有运行HBase
如果你在之前测试中忘记停止HBase,就会出错。用jps命令行检查HBase是否运行。看看HMaster,HRegionServer和HQuorumPeer是否存在,如果存在,那么杀掉。
2)启动集群
在node-a上,运行start-hbase.sh命令。就会打出类似下面的输出:
$ bin/start-hbase.sh
node-c.example.com: starting zookeeper, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-zookeeper-node-c.example.com.out
node-a.example.com: starting zookeeper, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-zookeeper-node-a.example.com.out
node-b.example.com: starting zookeeper, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-zookeeper-node-b.example.com.out
starting master, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-master-node-a.example.com.out
node-c.example.com: starting regionserver, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-regionserver-node-c.example.com.out
node-b.example.com: starting regionserver, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-regionserver-node-b.example.com.out
node-b.example.com: starting master, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-master-nodeb.example.com.out
先启动ZooKeeper,接着master,然后RegionServers,最后是备用masters。
3)检查进程是否运行
在集群中的每个节点,运行jps命令检查每个服务是否正常运行。你可能会看到其他用于其他目的Java进程也运行着。
Example 2. node-a jps Output
$ jps
20355 Jps
20071 HQuorumPeer
20137 HMaster
Example 3. node-b jps Output
$ jps
15930 HRegionServer
16194 Jps
15838 HQuorumPeer
16010 HMaster
Example 4. node-a jps Output
$ jps
13901 Jps
13639 HQuorumPeer
13737 HRegionServer
ZooKeeper 进程名字 HQuorumPeer 进程就是ZooKeeper实例由HBase启动用来控制HBase的。如果你在这里使用ZooKeeper,那么会限制集群中每个节点有一个实例并且只适用于测试。如果ZooKeeper运行在HBase之外,那么进程名为QuorumPeer。请到 zookeeper查看更多关于ZooKeeper配置包括如果用外部ZooKeeper控制HBase。 |
4)浏览Web
Web访问端口改变 |
如果HBase的版本高于0.98.x,那么登陆master的端口由60010改为16010,登陆RegionServer的端口由60030改为16030。
如果配置都正确的话,你应该能够使用浏览器通过 http://node-a.example.com:16010/ 连接Master,通过 http://node-b.example.com:16010/ 连接备用Master。如果你只能通过本地主机登陆而其他主机不能,检查你的防火墙规则。你可以通过ip:16030来连接RegionServers,也可以在Master的Web界面中点击相关链接来登陆。
5)当节点或者服务消失时测试一下发生了什么
正如你配置的三个节点,事情并不总是如你所想。你可以通过杀死进程观察log来看看当主Master或者RegionServer消失时发生了什么?
下面是原文
Getting Started
1. Introduction
Quickstart will get you up and running on a single-node, standalone instance of HBase.
2. Quick Start - Standalone HBase
This section describes the setup of a single-node standalone HBase. A standalone instance has all HBase daemons — the Master, RegionServers, and ZooKeeper — running in a single JVM persisting to the local filesystem. It is our most basic deploy profile. We will show you how to create a table in HBase using the hbase shell
CLI, insert rows into the table, perform put and scan operations against the table, enable or disable the table, and start and stop HBase.
Apart from downloading HBase, this procedure should take less than 10 minutes.
Prior to HBase 0.94.x, HBase expected the loopback IP address to be 127.0.0.1. Ubuntu and some other distributions default to 127.0.1.1 and this will cause problems for you. See Why does HBase care about /etc/hosts? for detail The following /etc/hosts file works correctly for HBase 0.94.x and earlier, on Ubuntu. Use this as a template if you run into trouble. 127.0.0.1 localhost 127.0.0.1 ubuntu.ubuntu-domain ubuntu This issue has been fixed in hbase-0.96.0 and beyond. |
2.1. JDK Version Requirements
HBase requires that a JDK be installed. See Java for information about supported JDK versions.
2.2. Get Started with HBase
-
Choose a download site from this list of Apache Download Mirrors. Click on the suggested top link. This will take you to a mirror of HBase Releases. Click on the folder named stable and then download the binary file that ends in .tar.gz to your local filesystem. Do not download the file ending in src.tar.gz for now.
-
Extract the downloaded file, and change to the newly-created directory.
$ tar xzvf hbase-2.0.0-SNAPSHOT-bin.tar.gz $ cd hbase-2.0.0-SNAPSHOT/
-
You are required to set the
JAVA_HOME
environment variable before starting HBase. You can set the variable via your operating system’s usual mechanism, but HBase provides a central mechanism, conf/hbase-env.sh. Edit this file, uncomment the line starting withJAVA_HOME
, and set it to the appropriate location for your operating system. TheJAVA_HOME
variable should be set to a directory which contains the executable file bin/java. Most modern Linux operating systems provide a mechanism, such as /usr/bin/alternatives on RHEL or CentOS, for transparently switching between versions of executables such as Java. In this case, you can setJAVA_HOME
to the directory containing the symbolic link to bin/java, which is usually /usr.JAVA_HOME=/usr
-
Edit conf/hbase-site.xml, which is the main HBase configuration file. At this time, you only need to specify the directory on the local filesystem where HBase and ZooKeeper write data. By default, a new directory is created under /tmp. Many servers are configured to delete the contents of /tmp upon reboot, so you should store the data elsewhere. The following configuration will store HBase’s data in the hbase directory, in the home directory of the user called
testuser
. Paste the<property>
tags beneath the<configuration>
tags, which should be empty in a new HBase install.Example 1. Example hbase-site.xml for Standalone HBase<configuration> <property> <name>hbase.rootdir</name> <value>file:///home/testuser/hbase</value> </property> <property> <name>hbase.zookeeper.property.dataDir</name> <value>/home/testuser/zookeeper</value> </property> </configuration>
You do not need to create the HBase data directory. HBase will do this for you. If you create the directory, HBase will attempt to do a migration, which is not what you want.
The hbase.rootdir in the above example points to a directory in the local filesystem. The 'file:/' prefix is how we denote local filesystem. To home HBase on an existing instance of HDFS, set the hbase.rootdir to point at a directory up on your instance: e.g. hdfs://namenode.example.org:8020/hbase. For more on this variant, see the section below on Standalone HBase over HDFS. -
The bin/start-hbase.sh script is provided as a convenient way to start HBase. Issue the command, and if all goes well, a message is logged to standard output showing that HBase started successfully. You can use the
jps
command to verify that you have one running process calledHMaster
. In standalone mode HBase runs all daemons within this single JVM, i.e. the HMaster, a single HRegionServer, and the ZooKeeper daemon. Go to http://localhost:16010 to view the HBase Web UI.Java needs to be installed and available. If you get an error indicating that Java is not installed, but it is on your system, perhaps in a non-standard location, edit the conf/hbase-env.sh file and modify the JAVA_HOME
setting to point to the directory that contains bin/javayour system.
-
Connect to HBase.
Connect to your running instance of HBase using the
hbase shell
command, located in the bin/ directory of your HBase install. In this example, some usage and version information that is printed when you start HBase Shell has been omitted. The HBase Shell prompt ends with a>
character.$ ./bin/hbase shell hbase(main):001:0>
-
Display HBase Shell Help Text.
Type
help
and press Enter, to display some basic usage information for HBase Shell, as well as several example commands. Notice that table names, rows, columns all must be enclosed in quote characters. -
Create a table.
Use the
create
command to create a new table. You must specify the table name and the ColumnFamily name.hbase(main):001:0> create 'test', 'cf' 0 row(s) in 0.4170 seconds => Hbase::Table - test
-
List Information About your Table
Use the
list
command tohbase(main):002:0> list 'test' TABLE test 1 row(s) in 0.0180 seconds => ["test"]
-
Put data into your table.
To put data into your table, use the
put
command.hbase(main):003:0> put 'test', 'row1', 'cf:a', 'value1' 0 row(s) in 0.0850 seconds hbase(main):004:0> put 'test', 'row2', 'cf:b', 'value2' 0 row(s) in 0.0110 seconds hbase(main):005:0> put 'test', 'row3', 'cf:c', 'value3' 0 row(s) in 0.0100 seconds
Here, we insert three values, one at a time. The first insert is at
row1
, columncf:a
, with a value ofvalue1
. Columns in HBase are comprised of a column family prefix,cf
in this example, followed by a colon and then a column qualifier suffix,a
in this case. -
Scan the table for all data at once.
One of the ways to get data from HBase is to scan. Use the
scan
command to scan the table for data. You can limit your scan, but for now, all data is fetched.hbase(main):006:0> scan 'test' ROW COLUMN+CELL row1 column=cf:a, timestamp=1421762485768, value=value1 row2 column=cf:b, timestamp=1421762491785, value=value2 row3 column=cf:c, timestamp=1421762496210, value=value3 3 row(s) in 0.0230 seconds
-
Get a single row of data.
To get a single row of data at a time, use the
get
command.hbase(main):007:0> get 'test', 'row1' COLUMN CELL cf:a timestamp=1421762485768, value=value1 1 row(s) in 0.0350 seconds
-
Disable a table.
If you want to delete a table or change its settings, as well as in some other situations, you need to disable the table first, using the
disable
command. You can re-enable it using theenable
command.hbase(main):008:0> disable 'test' 0 row(s) in 1.1820 seconds hbase(main):009:0> enable 'test' 0 row(s) in 0.1770 seconds
Disable the table again if you tested the
enable
command above:hbase(main):010:0> disable 'test' 0 row(s) in 1.1820 seconds
-
Drop the table.
To drop (delete) a table, use the
drop
command.hbase(main):011:0> drop 'test' 0 row(s) in 0.1370 seconds
-
Exit the HBase Shell.
To exit the HBase Shell and disconnect from your cluster, use the
quit
command. HBase is still running in the background.
-
In the same way that the bin/start-hbase.sh script is provided to conveniently start all HBase daemons, the bin/stop-hbase.sh script stops them.
$ ./bin/stop-hbase.sh stopping hbase.................... $
-
After issuing the command, it can take several minutes for the processes to shut down. Use the
jps
to be sure that the HMaster and HRegionServer processes are shut down.
The above has shown you how to start and stop a standalone instance of HBase. In the next sections we give a quick overview of other modes of hbase deploy.
2.3. Pseudo-Distributed Local Install
After working your way through quickstart standalone mode, you can re-configure HBase to run in pseudo-distributed mode. Pseudo-distributed mode means that HBase still runs completely on a single host, but each HBase daemon (HMaster, HRegionServer, and ZooKeeper) runs as a separate process: in standalone mode all daemons ran in one jvm process/instance. By default, unless you configure the hbase.rootdir
property as described in quickstart, your data is still stored in /tmp/. In this walk-through, we store your data in HDFS instead, assuming you have HDFS available. You can skip the HDFS configuration to continue storing your data in the local filesystem.
Hadoop Configuration
This procedure assumes that you have configured Hadoop and HDFS on your local system and/or a remote system, and that they are running and available. It also assumes you are using Hadoop 2. The guide onSetting up a Single Node Cluster in the Hadoop documentation is a good starting point. |
-
Stop HBase if it is running.
If you have just finished quickstart and HBase is still running, stop it. This procedure will create a totally new directory where HBase will store its data, so any databases you created before will be lost.
-
Configure HBase.
Edit the hbase-site.xml configuration. First, add the following property. which directs HBase to run in distributed mode, with one JVM instance per daemon.
<property> <name>hbase.cluster.distributed</name> <value>true</value> </property>
Next, change the
hbase.rootdir
from the local filesystem to the address of your HDFS instance, using thehdfs:////
URI syntax. In this example, HDFS is running on the localhost at port 8020.<property> <name>hbase.rootdir</name> <value>hdfs://localhost:8020/hbase</value> </property>
You do not need to create the directory in HDFS. HBase will do this for you. If you create the directory, HBase will attempt to do a migration, which is not what you want.
-
Start HBase.
Use the bin/start-hbase.sh command to start HBase. If your system is configured correctly, the
jps
command should show the HMaster and HRegionServer processes running. -
Check the HBase directory in HDFS.
If everything worked correctly, HBase created its directory in HDFS. In the configuration above, it is stored in /hbase/ on HDFS. You can use the
hadoop fs
command in Hadoop’s bin/ directory to list this directory.$ ./bin/hadoop fs -ls /hbase Found 7 items drwxr-xr-x - hbase users 0 2014-06-25 18:58 /hbase/.tmp drwxr-xr-x - hbase users 0 2014-06-25 21:49 /hbase/WALs drwxr-xr-x - hbase users 0 2014-06-25 18:48 /hbase/corrupt drwxr-xr-x - hbase users 0 2014-06-25 18:58 /hbase/data -rw-r--r-- 3 hbase users 42 2014-06-25 18:41 /hbase/hbase.id -rw-r--r-- 3 hbase users 7 2014-06-25 18:41 /hbase/hbase.version drwxr-xr-x - hbase users 0 2014-06-25 21:49 /hbase/oldWALs
-
Create a table and populate it with data.
You can use the HBase Shell to create a table, populate it with data, scan and get values from it, using the same procedure as in shell exercises.
-
Start and stop a backup HBase Master (HMaster) server.
Running multiple HMaster instances on the same hardware does not make sense in a production environment, in the same way that running a pseudo-distributed cluster does not make sense for production. This step is offered for testing and learning purposes only. The HMaster server controls the HBase cluster. You can start up to 9 backup HMaster servers, which makes 10 total HMasters, counting the primary. To start a backup HMaster, use the
local-master-backup.sh
. For each backup master you want to start, add a parameter representing the port offset for that master. Each HMaster uses three ports (16010, 16020, and 16030 by default). The port offset is added to these ports, so using an offset of 2, the backup HMaster would use ports 16012, 16022, and 16032. The following command starts 3 backup servers using ports 16012/16022/16032, 16013/16023/16033, and 16015/16025/16035.$ ./bin/local-master-backup.sh 2 3 5
To kill a backup master without killing the entire cluster, you need to find its process ID (PID). The PID is stored in a file with a name like /tmp/hbase-USER-X-master.pid. The only contents of the file is the PID. You can use the
kill -9
command to kill that PID. The following command will kill the master with port offset 1, but leave the cluster running:$ cat /tmp/hbase-testuser-1-master.pid |xargs kill -9
-
Start and stop additional RegionServers
The HRegionServer manages the data in its StoreFiles as directed by the HMaster. Generally, one HRegionServer runs per node in the cluster. Running multiple HRegionServers on the same system can be useful for testing in pseudo-distributed mode. The
local-regionservers.sh
command allows you to run multiple RegionServers. It works in a similar way to thelocal-master-backup.sh
command, in that each parameter you provide represents the port offset for an instance. Each RegionServer requires two ports, and the default ports are 16020 and 16030. However, the base ports for additional RegionServers are not the default ports since the default ports are used by the HMaster, which is also a RegionServer since HBase version 1.0.0. The base ports are 16200 and 16300 instead. You can run 99 additional RegionServers that are not a HMaster or backup HMaster, on a server. The following command starts four additional RegionServers, running on sequential ports starting at 16202/16302 (base ports 16200/16300 plus 2).$ .bin/local-regionservers.sh start 2 3 4 5
To stop a RegionServer manually, use the
local-regionservers.sh
command with thestop
parameter and the offset of the server to stop.$ .bin/local-regionservers.sh stop 3
-
Stop HBase.
You can stop HBase the same way as in the quickstart procedure, using the bin/stop-hbase.sh command.
2.4. Advanced - Fully Distributed
In reality, you need a fully-distributed configuration to fully test HBase and to use it in real-world scenarios. In a distributed configuration, the cluster contains multiple nodes, each of which runs one or more HBase daemon. These include primary and backup Master instances, multiple ZooKeeper nodes, and multiple RegionServer nodes.
This advanced quickstart adds two more nodes to your cluster. The architecture will be as follows:
Node Name | Master | ZooKeeper | RegionServer |
---|---|---|---|
node-a.example.com |
yes |
yes |
no |
node-b.example.com |
backup |
yes |
yes |
node-c.example.com |
no |
yes |
yes |
This quickstart assumes that each node is a virtual machine and that they are all on the same network. It builds upon the previous quickstart, Pseudo-Distributed Local Install, assuming that the system you configured in that procedure is now node-a
. Stop HBase on node-a
before continuing.
Be sure that all the nodes have full access to communicate, and that no firewall rules are in place which could prevent them from talking to each other. If you see any errors like no route to host , check your firewall. |
node-a
needs to be able to log into node-b
and node-c
(and to itself) in order to start the daemons. The easiest way to accomplish this is to use the same username on all hosts, and configure password-less SSH login from node-a
to each of the others.
-
On
node-a
, generate a key pair.While logged in as the user who will run HBase, generate a SSH key pair, using the following command:
$ ssh-keygen -t rsa
If the command succeeds, the location of the key pair is printed to standard output. The default name of the public key is id_rsa.pub.
-
Create the directory that will hold the shared keys on the other nodes.
On
node-b
andnode-c
, log in as the HBase user and create a .ssh/ directory in the user’s home directory, if it does not already exist. If it already exists, be aware that it may already contain other keys. -
Copy the public key to the other nodes.
Securely copy the public key from
node-a
to each of the nodes, by using thescp
or some other secure means. On each of the other nodes, create a new file called .ssh/authorized_keys if it does not already exist, and append the contents of the id_rsa.pub file to the end of it. Note that you also need to do this fornode-a
itself.$ cat id_rsa.pub >> ~/.ssh/authorized_keys
-
Test password-less login.
If you performed the procedure correctly, if you SSH from
node-a
to either of the other nodes, using the same username, you should not be prompted for a password. -
Since
node-b
will run a backup Master, repeat the procedure above, substitutingnode-b
everywhere you seenode-a
. Be sure not to overwrite your existing .ssh/authorized_keys files, but concatenate the new key onto the existing file using the>>
operator rather than the>
operator.
node-a
node-a
will run your primary master and ZooKeeper processes, but no RegionServers. . Stop the RegionServer from starting on node-a
.
-
Edit conf/regionservers and remove the line which contains
localhost
. Add lines with the hostnames or IP addresses fornode-b
andnode-c
.Even if you did want to run a RegionServer on
node-a
, you should refer to it by the hostname the other servers would use to communicate with it. In this case, that would benode-a.example.com
. This enables you to distribute the configuration to each node of your cluster any hostname conflicts. Save the file. -
Configure HBase to use
node-b
as a backup master.Create a new file in conf/ called backup-masters, and add a new line to it with the hostname for
node-b
. In this demonstration, the hostname isnode-b.example.com
. -
Configure ZooKeeper
In reality, you should carefully consider your ZooKeeper configuration. You can find out more about configuring ZooKeeper in zookeeper. This configuration will direct HBase to start and manage a ZooKeeper instance on each node of the cluster.
On
node-a
, edit conf/hbase-site.xml and add the following properties.<property> <name>hbase.zookeeper.quorum</name> <value>node-a.example.com,node-b.example.com,node-c.example.com</value> </property> <property> <name>hbase.zookeeper.property.dataDir</name> <value>/usr/local/zookeeper</value> </property>
-
Everywhere in your configuration that you have referred to
node-a
aslocalhost
, change the reference to point to the hostname that the other nodes will use to refer tonode-a
. In these examples, the hostname isnode-a.example.com
.
node-b
and node-c
node-b
will run a backup master server and a ZooKeeper instance.
-
Download and unpack HBase.
Download and unpack HBase to
node-b
, just as you did for the standalone and pseudo-distributed quickstarts. -
Copy the configuration files from
node-a
tonode-b
.andnode-c
.Each node of your cluster needs to have the same configuration information. Copy the contents of the conf/directory to the conf/ directory on
node-b
andnode-c
.
-
Be sure HBase is not running on any node.
If you forgot to stop HBase from previous testing, you will have errors. Check to see whether HBase is running on any of your nodes by using the
jps
command. Look for the processesHMaster
,HRegionServer
, andHQuorumPeer
. If they exist, kill them. -
Start the cluster.
On
node-a
, issue thestart-hbase.sh
command. Your output will be similar to that below.$ bin/start-hbase.sh node-c.example.com: starting zookeeper, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-zookeeper-node-c.example.com.out node-a.example.com: starting zookeeper, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-zookeeper-node-a.example.com.out node-b.example.com: starting zookeeper, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-zookeeper-node-b.example.com.out starting master, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-master-node-a.example.com.out node-c.example.com: starting regionserver, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-regionserver-node-c.example.com.out node-b.example.com: starting regionserver, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-regionserver-node-b.example.com.out node-b.example.com: starting master, logging to /home/hbuser/hbase-0.98.3-hadoop2/bin/../logs/hbase-hbuser-master-nodeb.example.com.out
ZooKeeper starts first, followed by the master, then the RegionServers, and finally the backup masters.
-
Verify that the processes are running.
On each node of the cluster, run the
jps
command and verify that the correct processes are running on each server. You may see additional Java processes running on your servers as well, if they are used for other purposes.Example 2.node-a
jps
Output$ jps 20355 Jps 20071 HQuorumPeer 20137 HMaster
Example 3.node-b
jps
Output$ jps 15930 HRegionServer 16194 Jps 15838 HQuorumPeer 16010 HMaster
Example 4.node-a
jps
Output$ jps 13901 Jps 13639 HQuorumPeer 13737 HRegionServer
ZooKeeper Process NameThe
HQuorumPeer
process is a ZooKeeper instance which is controlled and started by HBase. If you use ZooKeeper this way, it is limited to one instance per cluster node, , and is appropriate for testing only. If ZooKeeper is run outside of HBase, the process is calledQuorumPeer
. For more about ZooKeeper configuration, including using an external ZooKeeper instance with HBase, see zookeeper. -
Browse to the Web UI.
Web UI Port ChangesWeb UI Port ChangesIn HBase newer than 0.98.x, the HTTP ports used by the HBase Web UI changed from 60010 for the Master and 60030 for each RegionServer to 16010 for the Master and 16030 for the RegionServer.
If everything is set up correctly, you should be able to connect to the UI for the Master
http://node-a.example.com:16010/
or the secondary master athttp://node-b.example.com:16010/
for the secondary master, using a web browser. If you can connect vialocalhost
but not from another host, check your firewall rules. You can see the web UI for each of the RegionServers at port 16030 of their IP addresses, or by clicking their links in the web UI for the Master. -
Test what happens when nodes or services disappear.
With a three-node cluster like you have configured, things will not be very resilient. Still, you can test what happens when the primary Master or a RegionServer disappears, by killing the processes and watching the logs.