- 浏览: 283533 次
- 性别:
- 来自: 广州
文章分类
- 全部博客 (247)
- free talking (11)
- java (18)
- search (16)
- hbase (34)
- open-sources (0)
- architect (1)
- zookeeper (16)
- vm (1)
- hadoop (34)
- nutch (33)
- lucene (5)
- ubuntu/shell (8)
- ant (0)
- mapreduce (5)
- hdfs (2)
- hadoop sources reading (13)
- AI (0)
- distributed tech (1)
- others (1)
- maths (6)
- english (1)
- art & entertainment (1)
- nosql (1)
- algorithms (8)
- hadoop-2.5 (16)
- hbase-0.94.2 source (28)
- zookeeper-3.4.3 source reading (1)
- solr (1)
- TODO (3)
- JVM optimization (1)
- architecture (0)
- hbase-guideline (1)
- data mining (3)
- hive (1)
- mahout (0)
- spark (28)
- scala (3)
- python (0)
- machine learning (1)
最新评论
-
jpsb:
...
为什么需要分布式? -
leibnitz:
hi guy, this is used as develo ...
compile hadoop-2.5.x on OS X(macbook) -
string2020:
撸主真土豪,在苹果里面玩大数据.
compile hadoop-2.5.x on OS X(macbook) -
youngliu_liu:
怎样运行这个脚本啊??大牛,我刚进入搜索引擎行业,希望你能不吝 ...
nutch 数据增量更新 -
leibnitz:
also, there is a similar bug ...
2。hbase CRUD--Lease in hbase
refer to 0.20.6 version
==================
1 ----- use hbase's built in-zookeeper instance as coordinator
2----- use zookeeper cluster as coordinator
within this intallition,i met with a puzzle to this,
here are the steps for installation:
1)configure the hbase cluster
a).as we use hdfs to save files generated by HMaster and HRegionServer(s),so we must reference the hdfs-site.xml from hadoop config dir.u can :
copy the hdfs-site.xml to conf dir of hbase, OR
make a link file from it:
cd conf-dir-of-hbase
ln -s hdfs-site.xml-of-haoop
b)hbase-en.sh
*set the java home
*update: HBASE_MANAGES_ZK=false //not to use zookeeper managed by hbase
c)hbase-site.xml
*hbase.rootdir
let it point to the hdfs url,for example:hdfs://url-to-hdfs:<port>/path-to-share-by-regionservses
NOTE
:this url can NOT contains ip ,use domain is a must.
*hbase.cluster.distributed=true
*hbase.zookeeper.quorum=zk-server-list-splited-by comma
*hbase.zookeeper.property.clientPort=port-which-set-in-zoo.cfg(used to connect by client)
d)regionservers
add HRegionServer host list to this by one per line
2)start the various clusters(
note the sequences
)
a)start hadoop cluster:start-all.sh.yes this is the same like pseudo mode,so u can NOT use start-dfs.sh instead of!
b)start the zookeeper cluster
zkServer.sh start
run this command in all nodes in turn.
c)start the hbase cluster.run this command in a node which the HMater will keep in:
start-hbase.sh
then other region servers will be started up respectively.
this means that the sequence of start hbase cluster are:
hadoop -> zk -> hbase
and these are opposite from stopping it:
hbase -> zk -> hadoop
3) processes in nodes:
a) master:
8027 HMaster //hbase's master
2542 SecondaryNameNode
7920 QuorumPeerMain //zk process.
note
this is started outsite of hbase,it is not named with:HQuorumPeer
2611 JobTracker
2377 NameNode
b)region serves
3870 QuorumPeerMain
1156 TaskTracker
4019 HRegionServer //slave of hbase master
1026 DataNode
4) status in zk
a)nodes in zk tree
[safe-mode, root-region-server, rs, master, shutdown]
*safe-mode
[zk: localhost:2181(CONNECTED) 26] get /hbase/safe-mode //
value is empty
that is hbase has leave from safe mode.
cZxid = 115964117007
ctime = Wed May 18 23:03:03 CST 2011
mZxid = 115964117007
mtime = Wed May 18 23:03:03 CST 2011
pZxid = 115964117007
cversion = 0
dataVersion = 0
aclVersion = 0
ephemeralOwner = 0
dataLength = 0
numChildren = 0
*root-region-servers
[zk: localhost:2181(CONNECTED) 17] get /hbase/root-region-server
192.168.0.2:60020
i think this is "root" region considered by hbase as this is one of the RS(region servers)
*rs (another region server)
[zk: localhost:2181(CONNECTED) 25] get /hbase/rs //value is empty
cZxid = 115964117001
ctime = Wed May 18 23:02:56 CST 2011
mZxid = 115964117001
mtime = Wed May 18 23:02:56 CST 2011
pZxid = 115964117005
cversion = 2
dataVersion = 0
aclVersion = 0
ephemeralOwner = 0
dataLength = 0
numChildren = 2
* master
[zk: localhost:2181(CONNECTED) 21] get /hbase/master
192.168.0.1:60000
this is the ip and port belong to the master,it is used to report by rs(region servers)
*shutdown
[zk: localhost:2181(CONNECTED) 20] get /hbase/shutdown
up
NOTE :
1.the hbase use zk cluster to heartbeat(hbase itself have not this mechanism?),so zk must be started before the running of hbase cluster!
2.how to know to store data in hdfs?
as hbase is hdfs-based table,so it must know site-specific config before storing.so it use the hdfs-site.xml (replication from hadoop/conf) also.
actually,we have set the hbase.rootdir to point to hdfs file system,this is very import for a cluster hbase!
3.does hbase use cluster-mapreduce?
there is nothing mapred-related config in the xml,so i think it use local job runner to do the mapred job instea of JobTracker(that is cluster mapred mode)
BUT if you want to create a secondary index,hbase MAYBE start a job to do it,so it is needness to start MR in this case.
4.how to use zookeeper in hbase?
based on the proerties of ZK_MANAGER_ZOOKEEPER,hbase.zookeeper.quorum and hbase.zookeeper.property.clientPort,hbase will learn to how to connect to zookeeper.for example,if the first param is set to true ,and hbase will proced to use quorum and clientPort to poll the followers(or leader) of zookeeper cluster to read or write files.
see also:hbase architecture
发表评论
-
zookeeper-network partition
2016-02-29 17:36 1855env: hbase,94.26 zookeeper, ... -
hbase-export table to json file
2015-12-25 17:21 1669i wanna export a table to j ... -
hbase-logroll optimizations
2015-09-21 12:10 1054as u know,the hbas's dat ... -
hoya--hbase on yarn
2015-04-23 17:00 446Introducing Hoya – HBase on YA ... -
upgrades of hadoop and hbase
2014-10-28 11:39 7421.the match relationships ... -
hbase-how to know which regionservers are stuck when shutdown a cluster
2014-10-22 15:09 577u know,when u shutdowning a ... -
hbase-bloom filter
2014-10-20 16:06 354there is no notable performa ... -
compile hbase
2014-10-15 17:42 900IF u want to compile a ... -
hbase-region balancer
2014-08-28 13:52 683why along with the time go ... -
HBase Versus Bigtable(comparison)
2014-07-01 23:13 562below views are all from hbas ... -
hbase -how many regions are fit for a table when prespiting or keeping running
2014-06-27 11:46 715how many regions are thinked ... -
hbase -tables replication/snapshot/backup within/cross clusters
2014-06-24 18:09 802serial no soluti ... -
some important optimized advices for hbase-0.94.x
2014-06-20 17:56 1233The following gives you ... -
hbase PerformanceEvaluation benchmark - 0.94.2 VS 0.94.16 VS 0.96
2014-03-23 17:56 1094i worked to benchmark hbase p ... -
install snappy compression in hadoop and hbase
2014-03-08 00:36 4561.what is snappy ... -
downgrade hbase from 0.94.16 to 0.94.2
2014-03-07 02:25 1461we have a cluster about 25 no ... -
2。hbase CRUD--Caching VS Batch
2013-08-26 17:28 8161 in case of ONE region,a ta ... -
5。hbase高级部分:compact/split/balance及其它维护原理-delete table
2013-08-18 02:53 1093* table disabled is NOT ... -
2。hbase CRUD--Read(Scan) operations(server side)
2013-08-15 17:39 962just recovered from a di ... -
2。hbase CRUD--Read(query) operations
2013-08-15 16:41 854read Note: -Ge ...
相关推荐
<name>hbase.cluster.distributed <value>false ``` 6. **启动HBase**: 添加HBase的守护进程启动脚本到系统服务,然后启动HBase: ``` sudo cp $HBASE_HOME/contrib/init/hbase.init.sh /etc/init.d/...
$ sudo apt-get install software-properties-common $ sudo apt-add-repository ppa:ansible/ansible $ sudo apt-get update $ sudo apt-get install ansible 在 Mac 上安装 ansible $ brew install ansible ###...
Customize the HDInsight cluster and install additional Hadoop ecosystem projects using Script Actions Administering HDInsight from the Hadoop command prompt or Microsoft PowerShell Using the Microsoft...
- **安装命令**: `bin/plugin install mobz/elasticsearch-head`,此插件主要用于监控集群健康状态。 ##### 6. 启动Elasticsearch集群 - 拷贝并分发Elasticsearch到各个节点,修改每个节点的`network.host`配置为...
- **Enabling JDBC Support on Client Systems**: Install the JDBC driver on client systems. - **Establishing JDBC Connections**: Configure the connection URL and credentials. #### Upgrading Impala -...
4. **安装Arkgate插件**:通过MySQL客户端连接到实例,执行一系列的`install plugin`命令来安装Arkgate及其相关组件。 完成上述步骤后,用户就可以在Arkcontrol界面中创建和管理数据同步任务,调整参数,监控任务...
NoSQL数据库根据存储类型又分为多种,例如键值存储数据库(如Redis)、列存储数据库(如HBase)和文档型数据库(如MongoDB)。 #### 二、单实例MySQL的安装部署 1. **下载MySQL安装包** 访问MySQL官网...
2. **列存储数据库**:这类数据库以列簇形式存储数据,代表产品包括Cassandra、HBase、Riak等。适合应用于分布式的文件系统。优点是查找速度快,易于分布式扩展,但功能相对有限。 3. **文档型数据库**:这类...
3. **编译安装Redis**:解压下载的Redis压缩包后,进入解压后的目录,执行 `make` 命令进行编译,然后使用 `make install` 安装Redis。 #### 四、Redis客户端 Redis提供了多种客户端工具供用户选择,包括: 1. **...
- **列存储数据库**:如Cassandra、HBase等,适用于分布式的文件系统场景,数据模型为列簇式存储,优点在于查找速度快且易于扩展,但功能相对局限。 - **文档型数据库**:如CouchDB、MongoDB等,常用于Web应用中,...