先安装hadoop1.2.1,见http://blog.csdn.net/jediael_lu/article/details/38926477

1、配置hbase-site.xml

<property>
<name>hbase.zookeeper.quorum</name>
<value>master,slave1,slave2</value>
</property>
<property>
<name>hbase.rootdir</name>
<value>hdfs://master:9000/hbase</value>
</property>
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
</property>

2、配置regionservers

slave1
slave2

3、配置hbase-env.sh

export JAVA_HOME=/usr/java/jdk1.7.0_51

4、用hadoop的最新版本取代hbase的自带版本

$ cd lib/
$ rm hadoop-core-1.0.4.jar
$ cp /opt/jediael/hadoop-1.2.1/hadoop-core-1.2.1.jar .

5、将hbases整个目录复制到slave中。

scp -r hbase-0.94.26/ slave1:/opt/jediael
scp -r hbase-0.94.26/ slave2:/opt/jediael

6、启动hbase

$ bin/start-hbase.sh
slave1: starting zookeeper, logging to /opt/jediael/hbase-0.90.4/bin/../logs/hbase-jediael-zookeeper-slave1.out
slave2: starting zookeeper, logging to /opt/jediael/hbase-0.90.4/bin/../logs/hbase-jediael-zookeeper-slave2.out
starting master, logging to /opt/jediael/hbase-0.90.4/bin/../logs/hbase-jediael-master-master.out
slave1: starting regionserver, logging to /opt/jediael/hbase-0.90.4/bin/../logs/hbase-jediael-regionserver-slave1.out
slave2: starting regionserver, logging to /opt/jediael/hbase-0.90.4/bin/../logs/hbase-jediael-regionserver-slave2.out

7、验证

(1)http://master:60010/master-status



(2)jps命令

$ jps
5620 JobTracker
5377 NameNode
8641 Jps
5535 SecondaryNameNode
5890 Bootstrap
8476 HMaster $ jps
27888 HQuorumPeer
27977 HRegionServer
26628 TaskTracker
28085 Jps
26536 DataNode $ jps
27736 HQuorumPeer
26654 DataNode
26747 TaskTracker
27816 HRegionServer
27922 Jps

(3)文件操作


异常

2015-01-21 22:06:53,258 INFO org.apache.hadoop.hbase.master.ActiveMasterManager: Master=master:60000

2015-01-21 22:06:53,590 FATAL org.apache.hadoop.hbase.master.HMaster: Unhandled exception. Starting shutdown.

java.io.IOException: Call to master/10.171.29.191:9000 failed on local exception: java.io.EOFException

        at org.apache.hadoop.ipc.Client.wrapException(Client.java:775)

        at org.apache.hadoop.ipc.Client.call(Client.java:743)

        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220)

        at com.sun.proxy.$Proxy5.getProtocolVersion(Unknown Source)

        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359)

        at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:113)

        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:215)

        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:177)

        at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82)

        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378)

        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)

        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390)

        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196)

        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175)

        at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:364)

        at org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:81)

        at org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:346)

        at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:282)

Caused by: java.io.EOFException

        at java.io.DataInputStream.readInt(DataInputStream.java:392)

        at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:501)

        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:446)

2015-01-21 22:06:53,599 INFO org.apache.hadoop.hbase.master.HMaster: Aborting

2015-01-21 22:06:53,599 DEBUG org.apache.hadoop.hbase.master.HMaster: Stopping service threads

2015-01-21 22:06:53,599 INFO org.apache.hadoop.ipc.HBaseServer: Stopping server on 60000

2015-01-21 22:06:53,599 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 0 on 60000: exiting

2015-01-21 22:06:53,599 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 1 on 60000: exiting

2015-01-21 22:06:53,599 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 2 on 60000: exiting

2015-01-21 22:06:53,600 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 3 on 60000: exiting

2015-01-21 22:06:53,600 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 4 on 60000: exiting

2015-01-21 22:06:53,600 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 5 on 60000: exiting

2015-01-21 22:06:53,600 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 6 on 60000: exiting

2015-01-21 22:06:53,600 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 7 on 60000: exiting

2015-01-21 22:06:53,601 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 8 on 60000: exiting

2015-01-21 22:06:53,601 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 9 on 60000: exiting

2015-01-21 22:06:53,601 INFO org.apache.hadoop.ipc.HBaseServer: Stopping IPC Server listener on 60000

2015-01-21 22:06:53,603 INFO org.apache.hadoop.ipc.HBaseServer: Stopping IPC Server Responder

2015-01-21 22:06:53,622 INFO org.apache.zookeeper.ZooKeeper: Session: 0x4b0cd139190000 closed

2015-01-21 22:06:53,622 INFO org.apache.hadoop.hbase.master.HMaster: HMaster main thread exiting

2015-01-21 22:06:53,623 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down

搭建hbase-0.94.26集群环境的更多相关文章

  1. 搭建hbase-0.94.26集群环境 分类: B7_HBASE 2015-01-24 17:14 861人阅读 评论(0) 收藏

    先安装hadoop1.2.1,见http://blog.csdn.net/jediael_lu/article/details/38926477 1.配置hbase-site.xml <prop ...

  2. hbase 0.98.1集群安装

    本文将基于hbase 0.98.1解说其在linux集群上的安装方法,并对一些重要的设置项进行解释,本文原文链接:http://blog.csdn.net/bluishglc/article/deta ...

  3. CentOs7.3 搭建 Redis-4.0.1 Cluster 集群服务

    环境 VMware版本号:12.0.0 CentOS版本:CentOS 7.3.1611 三台虚拟机(IP):192.168.252.101,192.168.102..102,192.168.252. ...

  4. Hbase 0.92.1集群数据迁移到新集群

    老集群 hbase(main):001:0> status 4 servers, 0 dead, 0.0000 average load hbase(main):002:0> list T ...

  5. CentOS7.6搭建redis4.0.1 cluster集群

    1. 操作系统信息: $ uname -a Linux iZbp11d57wmumnwuihb2czZ -.el7.x86_64 # SMP Fri Feb :: UTC x86_64 x86_64 ...

  6. 大数据架构:搭建CDH5.5.1分布式集群环境

    yum install -y ntp gcc make lrzsz wget vim sysstat.x86_64 xinetd screen expect rsync bind-utils ioto ...

  7. hadoop集群环境的搭建

    hadoop集群环境的搭建 今天终于把hadoop集群环境给搭建起来了,能够运行单词统计的示例程序了. 集群信息如下: 主机名 Hadoop角色 Hadoop jps命令结果 Hadoop用户 Had ...

  8. Redis集群环境搭建实践

    0 Redis集群简介 Redis集群(Redis Cluster)是Redis提供的分布式数据库方案,通过分片(sharding)来进行数据共享,并提供复制和故障转移功能.相比于主从复制.哨兵模式, ...

  9. ElasticSearch-5.3.1集群环境搭建,安装ElasticSearch-head插件,安装错误解决

    说起来甚是惭愧,博主在写这篇文章的时候,还没有系统性的学习一下ES,只知道可以拿来做全文检索,功能很牛逼,但是接到了任务不想做也不行, leader让我搭建一下分布式的ES集群环境,用来支持企业信用数 ...

随机推荐

  1. 14--物理引擎Box2D

    物理模拟引擎专注于模拟现实世界中物体以及物体之间的基本运动规律.在游戏中引入物理引擎能提高游戏的真实性和可玩性,如<愤怒的小鸟>中小鸟弹出的抛物线运动.<割绳子>中割后的运动等 ...

  2. NET Core依赖注入解读&使用Autofac替代实现

    NET Core依赖注入解读&使用Autofac替代实现 标签: 依赖注入 Autofac ASPNETCore ASP.NET Core依赖注入解读&使用Autofac替代实现 1. ...

  3. 通过内省机制设置JavaBean

    一.步骤: 1)使用PropertyDescriptor类获取属性描述者对象 //pd引用Student的name属性 PropertyDescriptor pd = new PropertyDesc ...

  4. VS2010 Command Prompt Error:Cannot determine the location of the VS Common Tools folder

    就在VS2010 Command Prompt 用vcvarsall.bat x64重新设置环境变量的时候,出现了标题中的错误.原因就在参考链接中 References: http://stackov ...

  5. cocos2dx lua 加密

    cocos2dx-lua项目发布时,为了保护lua源码,需要对lua进行加密.通常分为两种方式:加密文件和编译为字节码. 1.加密文件 前提是你不用luajit,而使用lua.这样这种方法是真正加密, ...

  6. 如何从Linux系统中获取带宽、流量网络数据

    引入 国外的云主机厂商,例如AWS提供的网络数据是以流量为单位的,例如下面的图片: 从上图来看,其取值方式为 每隔5分钟取值1次,(每次)每个点显示为1分钟内的流量字节数(Bytes) 带宽与流量 我 ...

  7. Qt Creator中,include路径包含过程(或如何找到对应的头文件)

    Qt Creator中,include路径包含过程(或如何找到对应的头文件) 利用Qt Creator开发程序时,需要包含利用#include来添加头文件. 大家都知道,#include <&g ...

  8. tool - 支持TestLink 1.93,将excel格式用例转化成可以导入的xml格式

     tool - 支持TestLink 1.93,将excel格式用例转化成可以导入的xml格式  https://github.com/zhangzheyuk/CaseConvert

  9. Redis安全

    安全 执行在可信环境 Redis的安全设计是在"Redis执行在可信环境"这个前提下做出的.在生产环境执行时不能同意外界直接连接到Redisserver上.而应该通过应用程序进行中 ...

  10. [AngularJS] Angular 1.5 multiple transclude

    If you know ui-router, multi-transclude should be easy for you also. In previou Angular version < ...