big data

禁止转载 我所遇见的世界会更美好

HBase shell 、写和读HDFS、HBase建表、编程、开启debug模式

使用Hadoop shell命令导入和导出数据到HDFS

1.2 使用Hadoop shell命令导入和导出数据到HDFS(1)

HDFS提供了许多shell命令来实现访问文件系统的功能,这些命令都是构建在HDFS FileSystem API之上的。Hadoop自带的shell脚本是通过命令行来执行所有操作的。这个脚本的名称叫做hadoop,通常安装在$HADOOP_BIN目录下,其中$HADOOP_BIN是Hadoopbin文件完整的安装目录,同时有必要将$HADOOP_BIN配置到$PATH环境变量中,这样所有的命令都可以通过hadoop fs -command这样的形式来执行。

如果需要获取文件系统的所有命令,可以运行hadoop命令传递不带参数的选项fs。
 

  1. hadoop fs 

 

这些按照功能进行命名的命令的名称与Unix shell命令非常相似。使用help选项可以获得某个具体命令的详细说明。
 

  1. hadoop fs –help ls 

这些shell命令和其简要的参数描述可在官方在线文档http://hadoop.apache.org/docs/r1.0.4/file_system_shell.html中进行查阅。

在这一节中,我们将使用Hadoop shell命令将数据导入HDFS中,以及将数据从HDFS中导出。这些命令更多地用于加载数据,下载处理过的数据,管理文件系统,以及预览相关数据。掌握这些命令是高效使用HDFS的前提。

准备工作

你需要在http://www.packtpub.com/support这个网站上下载数据集weblog_ entries.txt。

操作步骤

完成以下步骤,实现将weblog_entries.txt文件从本地文件系统复制到HDFS上的一个指定文件夹下。 在HDFS中创建一个新文件夹,用于保存weblog_entries.txt文件:
 

  1. hadoop fs -mkdir /data/weblogs 

将weblog_entries.txt文件从本地文件系统复制到HDFS刚创建的新文件夹下:
 

  1. hadoop fs -copyFromLocal weblog_entries.txt /data/weblogs 

列出HDFS上weblog_entries.txt文件的信息:
 

  1. hadoop fs –ls /data/weblogs/weblog_entries.txt 

在Hadoop处理的一些结果数据可能会直接被外部系统使用,可能需要其他系统做更进一步的处理,或者MapReduce处理框架根本就不符合该场景,任何类似的情况下都需要从HDFS上导出数据。下载数据最简单的办法就是使用Hadoop shell。
 

//将本地文件上传到hdfs
  String target="hdfs://localhost:9000/user/Administrator/geoway_portal/tes2.dmp";
  FileInputStream fis=new FileInputStream(new File("C:\tes2.dmp"));//读取本地文件
  Configuration config=new Configuration();
  FileSystem fs=FileSystem.get(URI.create(target), config);
  OutputStream os=fs.create(new Path(target));
  //copy
  IOUtils.copyBytes(fis, os, 4096, true);
  System.out.println("拷贝完成...");
 
查看上传的hdfs的文件信息:

将本地文件上传至HDFS

使用Hadoop fs -ls shell命令查询geoway_portal下的文件情况:

将本地文件上传至HDFS

以下代码实现将本地文件拷到HDFS集群中
 
package com.njupt.Hadoop;
 

import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 

public class CopyToHDFS {
 

public static void main(String[] args) throws Exception{
 


Configuration conf = new Configuration();
 FileSystem fs = FileSystem.get(conf);
 Path source = new Path("/home/hadoop/word.txt");
 Path dst = new Path("/user/root/njupt/");
 fs.copyFromLocalFile(source,dst);
 }
 }

使用HDFS java api 下载文件到本地的代码如下:
 
  String file="hdfs://localhost:9000/user/Administrator/fooo/j-spatial.zip";//hdfs文件 地址
  Configuration config=new Configuration();
  FileSystem fs=FileSystem.get(URI.create(file),config);//构建FileSystem
  InputStream is=fs.open(new Path(file));//读取文件
  IOUtils.copyBytes(is, new FileOutputStream(new File("c:\likehua.zip")),2048, true);//保存到本地  最后 关闭输入输出流

./start

关于配置环境变量

./hbase shell

会了和能做项目是两个层次,完全不同的概念的撒。

 

时间戳可以变成时间表那样??

slave2:

slave1:

[root@slave1 ~]# ll
总用量 52
-rw-------. 1 root root 950 10月 14 16:02 anaconda-ks.cfg
-rw-r--r--. 1 root root 12081 5月 22 21:31 post-install
-rw-r--r--. 1 root root 552 5月 22 21:31 post-install.log
drwxr-xr-x. 2 root root 4096 10月 22 17:43 公共的
drwxr-xr-x. 2 root root 4096 10月 22 17:43 模板
drwxr-xr-x. 2 root root 4096 10月 22 17:43 视频
drwxr-xr-x. 2 root root 4096 10月 22 17:43 图片
drwxr-xr-x. 2 root root 4096 10月 22 17:43 文档
drwxr-xr-x. 2 root root 4096 10月 22 17:43 下载
drwxr-xr-x. 2 root root 4096 10月 22 17:43 音乐
drwxr-xr-x. 2 root root 4096 10月 22 17:43 桌面

[root@slave1 ~]# cd /etc/sysconfig/network-scripts/

如何从master面板快速切换到slave2?  shift+tab
[root@slave1 network-scripts]# cat ifcfg-eth1
TYPE=Ethernet
BOOTPROTO=none
DEFROUTE=yes
IPV4_FAILURE_FATAL=yes
IPV6INIT=no
NAME=eth0
UUID=16a631ff-0a3b-4915-9d6d-99743e5b0b56
ONBOOT=yes
HWADDR=00:0C:29:66:10:7b
IPADDR=192.168.150.4
PREFIX=24
GATEWAY=192.168.150.2
DNS1=192.168.150.2
LAST_CONNECT=1477239855
[root@slave1 network-scripts]# su hadoop

[hadoop@slave1 network-scripts]$ cd

[hadoop@slave1 ~]$ ll
总用量 68
drwxrwxr-x 5 hadoop hadoop 4096 11月 8 20:31 soft
-rw-rw-r-- 1 hadoop hadoop 60181 11月 9 11:44 zookeeper.out

[hadoop@slave1 ~]$ zkServer.sh start
JMX enabled by default
Using config: /home/hadoop/soft/zookeeper-3.4.6/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[hadoop@slave1 ~]$ zkServer.sh status
JMX enabled by default
Using config: /home/hadoop/soft/zookeeper-3.4.6/bin/../conf/zoo.cfg
Mode: leader
[hadoop@slave1 ~]$ jps
2971 Jps
2787 QuorumPeerMain
2907 HRegionServer
2588 NodeManager
2517 DataNode

slave2: 

[root@slave2 ~]# ifconfig

[root@slave2 ~]# su hadoop

[hadoop@slave2 ~]$ zkServer.
zkServer.cmd zkServer.sh
[hadoop@slave2 ~]$ zkServer.sh start
JMX enabled by default
Using config: /home/hadoop/soft/zookeeper-3.4.6/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[hadoop@slave2 ~]$ jps
2609 NodeManager
2797 QuorumPeerMain
2534 DataNode
2824 Jps
[hadoop@slave2 ~]$

master:

[root@master ~]# su hadoop
[hadoop@master ~]$ start-all.sh
This script is Deprecated. Instead use start-dfs.sh and start-yarn.sh
16/11/19 10:55:16 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Starting namenodes on [master]
master: starting namenode, logging to /home/hadoop/soft/hadoop-2.6.4/logs/hadoop-hadoop-namenode-master.out
slave2: starting datanode, logging to /home/hadoop/soft/hadoop-2.6.4/logs/hadoop-hadoop-datanode-slave2.out
slave1: starting datanode, logging to /home/hadoop/soft/hadoop-2.6.4/logs/hadoop-hadoop-datanode-slave1.out
Starting secondary namenodes [0.0.0.0]
0.0.0.0: starting secondarynamenode, logging to /home/hadoop/soft/hadoop-2.6.4/logs/hadoop-hadoop-secondarynamenode-master.out
16/11/19 10:55:40 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
starting yarn daemons
starting resourcemanager, logging to /home/hadoop/soft/hadoop-2.6.4/logs/yarn-hadoop-resourcemanager-master.out
slave2: starting nodemanager, logging to /home/hadoop/soft/hadoop-2.6.4/logs/yarn-hadoop-nodemanager-slave2.out
slave1: starting nodemanager, logging to /home/hadoop/soft/hadoop-2.6.4/logs/yarn-hadoop-nodemanager-slave1.out
[hadoop@master ~]$ jps
5623 SecondaryNameNode
5466 NameNode
6050 Jps
5793 ResourceManager

[hadoop@master ~]$ zkServer.sh start
JMX enabled by default
Using config: /home/hadoop/soft/zookeeper-3.4.6/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED
[hadoop@master ~]$ zkServer.sh status
JMX enabled by default
Using config: /home/hadoop/soft/zookeeper-3.4.6/bin/../conf/zoo.cfg
Mode: follower
[hadoop@master ~]$ jps
5623 SecondaryNameNode
5466 NameNode
6117 QuorumPeerMain
5793 ResourceManager
6174 Jps


[hadoop@master ~]$ start-hbase.sh
starting master, logging to /home/hadoop/soft/hbase-1.2.4/logs/hbase-hadoop-master-master.out
192.168.150.3: starting regionserver, logging to /home/hadoop/soft/hbase-1.2.4/logs/hbase-hadoop-regionserver-master.out
192.168.150.5: starting regionserver, logging to /home/hadoop/soft/hbase-1.2.4/logs/hbase-hadoop-regionserver-slave2.out
192.168.150.4: starting regionserver, logging to /home/hadoop/soft/hbase-1.2.4/logs/hbase-hadoop-regionserver-slave1.out
[hadoop@master ~]$ jps
jps^H^H5623 SecondaryNameNode
6437 HRegionServer
5466 NameNode
6117 QuorumPeerMain
5793 ResourceManager
6547 Jps
6304 HMaster
[hadoop@master ~]$ hdfs dfs -ls /
16/11/19 11:01:23 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your pla
Found 3 items
drwxr-xr-x - hadoop supergroup 0 2016-11-19 10:59 /hbase
drwxr-xr-x - hadoop supergroup 0 2016-10-25 20:40 /tmp
drwxr-xr-x - hadoop supergroup 0 2016-10-25 19:43 /user
[hadoop@master ~]$ free
total used free shared buffers cached
Mem: 1012076 899848 112228 8 872 36104
-/+ buffers/cache: 862872 149204
Swap: 2031612 152652 1878960
[hadoop@master ~]$ free -m
total used free shared buffers cached
Mem: 988 880 107 0 0 35
-/+ buffers/cache: 844 144
Swap: 1983 148 1835
[hadoop@master ~]$ hbase shell
2016-11-19 11:02:55,144 WARN [main] util.NativeCodeLoader: Unable to load native-hadoop librar
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/home/hadoop/soft/hbase-1.2.4/lib/slf4j-log4j12-1.7.5.jar!/or
SLF4J: Found binding in [jar:file:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/slf4j-
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
HBase Shell; enter 'help<RETURN>' for list of supported commands.
Type "exit<RETURN>" to leave the HBase Shell
Version 1.2.4, r67592f3d062743907f8c5ae00dbbe1ae4f69e5af, Tue Oct 25 18:10:20 CDT 2016

hbase(main):001:0> helo
NameError: undefined local variable or method `helo' for #<Object:0x45b4d4a>

hbase(main):002:0> list
TABLE
0 row(s) in 2.3030 seconds

=> []
hbase(main):003:0> creard
NameError: undefined local variable or method `creard' for #<Object:0x45b4d4a>

hbase(main):004:0> put "''
hbase(main):005:0" put
hbase(main):006:0" [hadoop@master ~]$
[hadoop@master ~]$
[hadoop@master ~]$ hbase shell
2016-11-19 11:12:23,455 WARN [main] util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/home/hadoop/soft/hbase-1.2.4/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
HBase Shell; enter 'help<RETURN>' for list of supported commands.
Type "exit<RETURN>" to leave the HBase Shell
Version 1.2.4, r67592f3d062743907f8c5ae00dbbe1ae4f69e5af, Tue Oct 25 18:10:20 CDT 2016

hbase(main):001:0> list
TABLE
0 row(s) in 1.0970 seconds

=> []
hbase(main):002:0>

hbase删除输入错误:

执行删除操作时,需要使用组合键:Ctrl+Back Space同时按下

 

HBase灵活的,如果网页更改,适用于结构化、半结构化;只需要更改列和限定符就行。   以列的方法存储 ,

                                         非稀疏,容易形成数据块。

传统的数据库:稀疏的(就是有空格,有空缺的元素留空白)

逻辑结构 --> 物理结构

第一个表格中每一个时间戳的增加,会可能带来三个新元素,然后找到第二个表的相对应行增加相应行的子行

hbasecontents:html = "<html>..."_百度搜索
https://www.baidu.com/s?wd=hbasecontents%3Ahtml%20%3D%20%22%3Chtml%3E...%22&rsv_spt=1&rsv_iqid=0xe281c553000aa0bf&issp=1&f=3&rsv_bp=1&rsv_idx=2&ie=utf-8&rqlang=cn&tn=site888_3_pg&rsv_enter=1&oq=contents%3Ahtml%20%3D%20%26quot%3B%26lt%3Bhtml%26gt%3B...%26quot%3B&inputT=2068&rsv_t=0cfeUlBQhS2%2B9h%2BkqlPLaajeoQzElnvneQ8%2BdFYDgKQeMeNR3FouOAsDE6nbEiGcSW1s&rsv_sug3=4&rsv_pq=f0944adf000e0101&rsv_sug2=0&prefixsug=hbasecontents%3Ahtml%20%3D%20%22%3Chtml%3E...%22&rsp=6&rsv_sug9=es_0_1&rsv_sug4=3285&rsv_sug=9

HBase使用教程 - 小码哥BASE64 - ITeye技术网站
http://coderbase64.iteye.com/blog/2074601

hbase(main):后面的>变成‘’_百度搜索
https://www.baidu.com/s?wd=hbase(main)%3A%E5%90%8E%E9%9D%A2%E7%9A%84%3E%E5%8F%98%E6%88%90%E2%80%98%E2%80%99&rsv_spt=1&rsv_iqid=0xf52382ec000cdcb3&issp=1&f=3&rsv_bp=1&rsv_idx=2&ie=utf-8&rqlang=cn&tn=site888_3_pg&rsv_enter=1&rsv_t=9ecfwiGV7s4Gk1e0lIpu67VyNgGsC%2BiNO1tA%2FuEb4dWNkbY3Y6CqoOSE838ePQLnjfLO&oq=hbase(main)%3A005%3A0%26quot%3Bhbase(main)%3A005%3A0%26quot%3B&rsv_sug3=32&rsv_pq=8550b5d3000a62b9&rsv_sug2=0&prefixsug=hbase(main)%3A%E5%90%8E%E9%9D%A2%E7%9A%84%3E%E5%8F%98%E6%88%90%E2%80%98%E2%80%99&rsp=7&rsv_sug9=es_0_1&inputT=49183&rsv_sug4=49743&rsv_sug=9

HBase是按列存储的,稀疏(列族少,数据吸收)矩阵。 

HBase是列存储,以列族为存储单位,传统的数据库是按行存储。

即使行、列相同,得到的值也是不一样的(有时间戳)。

回收策略:数据的压缩和整理,没用的都会删掉。

 

HBase 优化需要的源代码

一个物理机有且仅有一个HRegionServer 、可能含有一个HBase(集群中只有一个物理机含有)

见图  HRegionServer 、HBase之间的关系相当于虚拟机的master、slave1、slave2....n

  这个架构也类似于hadoop本身的架构,由master和namenode存储元数据,真正的数据存储在datanode上(会有多个datanode)。

-        Hmaster负责元数据存储,维护整个hbase集群的状态。

-        HRegionServer是真正管理数据的增删改查(数据库的功能),用户提交的数据请求都是通过HRegionServer协调处理

²  每一个物理机器节点只能有一个HRegionServer。

²  在HRegionServer下管理HRegion。

            HRegion中存放的真正存储单元

       HRegion就是我们创建表的时候创建的,在默认情况下建立的表都有一个HRegion(最基本的存储单元,个数由列族决定),region根据行键划分,对应部分范围放到一起

       HRegion到底交个哪个HRegionServer管理,是由Hmaster管理,它会考虑负载均衡和资源管理。

       HRegion中有一系列的结构。

 

 HLog (hbase log)日志记录文件。相当于oracle中的重做日志文件;Redo log记录增删改查的每一笔记录

Store单元存放真实的数据放

 数据以列族为单位存放,列族就存放在store中。

  Store又分为两个部分:MemStore内存存储;StoreFile 存储文件,是以HFile存在的,存在于hdfs系统中datanode上。

     在HRegionServer中  HRegion会有多个,随着数据不断增加就会产生新的HRegion,hlog只有一个。

      在hadoop中,存储是均匀的,hbase的hfile就会分布在datanode上。整体上数据进行了两次分布。在hbase和hadoop上进行两次分布。

 

       第一次分布:HBase  --》HRegionServicer

         第二次分布:  storefile --》  HDFS

一次一个块读取效率高,所以会有memstore(内存缓存区);memstore多了在以数据块的方式存在

最终落实在磁盘上时是HDFS

一个数据块64 

memstore -》 storefile以Hfile的形式才能发给  -》HDFS

这里HBase是虚的,不干具体的实事,找到一个空闲即负载均衡的HRegionServicer去做具体的工作,

HRegionServicer相当于数据库服务器 数据管理单位HRegion

Store真正数据存放的地方,以文件多个HFile的方式持久化存储,MemStore用来存放零碎的,存到一定的数量再存放到StoreFile

再经过客户端存、取

HLog用日志来数据持久化

HBase的访问方式  见1.4文档详细内容

thrift  序列化技术 

每一个store 对应一个列族

 

 

 

2个映像是为了备份HDFS中的文件都要备份? 其他的地方不需要备份吗?

阻抗失谐

Redis 教程 | 菜鸟教程
http://www.runoob.com/redis/redis-tutorial.html

 针对快照:

 拷贝给别人会生成一份日志,要删掉;从别人那里拷贝过来要调网络配置

现在安装的    密码是root

老师的是打开的     是fdq , 进去是111111

跟教授稀里糊涂的学的一塌糊涂

Hadoop、hbase、Mapreduce

hadoop的安装 - [ hadoop学习 ]
http://www.kancloud.cn/zizhilong/hadoop/224101

虾皮工作室 - 赠人玫瑰,手留余香。
http://www.xiapistudio.com/

hadoop中典型Writable类详解 - wuyudong - 博客园
http://www.cnblogs.com/wuyudong/p/hadoop-writable.html

Hadoop学习笔记(1):WordCount程序的实现与总结 - JD_Beatles - 博客园
http://www.cnblogs.com/pengyingzhi/p/5361008.html#3401994

Hadoop示例程序WordCount详解及实例 - xwdreamer - 博客园
http://www.cnblogs.com/xwdreamer/archive/2011/01/04/2297049.html

Hadoop WordCount解读 - 犀利的代码总是很精炼 - ITeye技术网站
http://a123159521.iteye.com/blog/1226924

import java.io.IOException;
import java.net.URI;
import java.util.StringTokenizer;


import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
import org.apache.hadoop.util.GenericOptionsParser;

/**
*
*
* 小结:Hadoop程序处理流程
*
* 1. 分析输入文件
* 2. 文件里面的每一行都会执行一次 TokenizerMapper.map() 分割成单词
* 3. 每一个单词都会被送到 KeyCombiner.reduce() 进行处理,统计次数
* 4. 当所有的单词都统计完成以后,会合并到一起然后送到 IntSumReducer.reduce() 进行最后的处理
*
*/

public class WordCount {

private static Configuration conf;
private static String hdfsPath = "hdfs://192.168.150.3:9000"; //hdfs 主机
private static Path inputPath = new Path(hdfsPath + "/tmp/wordcount/in"); //hdfs完整输入路径
private static String outPathStr = "/tmp/wordcount/out"; //hdfs输出路径
private static Path outPath = new Path(hdfsPath + outPathStr); //hdfs完整输出路径


/**
* Mapper 处理
*
* 四个参数的含义,前两个是输入的值 Key/Value的类型,后两个是输出的值Key/Value的类型 , 此处的输出类型需要KeyCombiner的输入参数保持一致
*/
public static class TokenizerMapper extends Mapper<Object,Text,Text,IntWritable>{

/**
* one 词的个数,刚开始都为1,也可以不定义,直接context.write(keytext, 1);
*/
private final static IntWritable one=new IntWritable(1);

/**
* word 用于临时保存分隔 出 的 单词
*/
private Text word =new Text();

/**
* @param key 即行偏移量,作用不大
* @param value 匹配到的当前行的内容
* @param context
*/
public void map(Object key,Text value,Context context) throws IOException,InterruptedException{
System.out.println("mapper --- 行内容:"+value);

StringTokenizer itr=new StringTokenizer(value.toString()); //把内容行按空格进行分隔

while (itr.hasMoreTokens()) { //返回是否还有分隔符。

word.set(itr.nextToken()); //返回从当前位置到下一个分隔符的字符串。

//System.out.println("mapper --- 分割成单词:"+word);

//以key/value的形式发送给reduce
context.write(word, one);
}
}
}

/**
* Reducer 处理
*
* 四个参数的含义,前两个是输入的值 Key/Value的类型,后两个是输出的值Key/Value的类型
*
* 此处的【输入参数类型】需要和 KeyCombiner 的【输出参数类型】保持一致
*
*/
public static class IntSumReducer extends Reducer<Text,IntWritable,Text,IntWritable> {

/**
* result 用于记录单词的出现数量
*/
private IntWritable result = new IntWritable();

/**
* @param key 单词名称
* @param values 单词对应的值的集合
* @param context
*/
public void reduce(Text key, Iterable<IntWritable> values,Context context) throws IOException, InterruptedException {

int sum = 0;
for (IntWritable val : values) { //遍历值集合
int getval = val.get(); //得到值
sum += getval; //累加数量
//System.out.println("reduce --- getval:"+getval);
}
System.out.println("reduce --- key:"+key +",sum:"+sum);
result.set(sum); //保存累加的数值
context.write(key, result); //将累加后的数值
}
}

/**
* Combiner 处理
*
* 四个参数的含义,前两个是输入的值 Key/Value的类型,后两个是输出的值Key/Value的类型
*
* 此处的【输入参数类型】需要和 TokenizerMapper 的【输出参数类型】保持一致
*
* 此处的【输出参数类型】需要和 IntSumReducer 的【输入参数类型】保持一致
*
*/
public static class KeyCombiner extends Reducer<Text,IntWritable,Text,IntWritable> {

/**
* result 用于记录单词的出现数量
*/
private IntWritable result = new IntWritable();

/**
* @param key 单词名称
* @param values 单词对应的值的集合
* @param context
*/
public void reduce(Text key, Iterable<IntWritable> values,Context context) throws IOException, InterruptedException {

int sum = 0;
for (IntWritable val : values) { //遍历值集合
int getval = val.get(); //得到值
sum += getval; //累加数量
//System.out.println("reduce --- getval:"+getval);
}
System.out.println("combiner --- key:"+key+",sum:"+sum);
result.set(sum); //保存累加的数值
context.write(key, result); //将结果传递给 IntSumReducer
}
}


/**
*
* 执行流程
*
* 输入文件 -> TokenizerMapper -> KeyCombiner -> IntSumReducer -> 输出到文件
*
* @param args
* @throws Exception
*/
public static void main(String[] args) throws Exception {
System.setProperty("hadoop.home.dir", "D:\soft\hadoop-2.6.4");

//删除临时输出目录
WordCount.deleteDir();

conf = new Configuration();
Job job = Job.getInstance(); //得到 job的实例
job.setJobName("word count"); //设置任务名称
job.setJarByClass(WordCount.class); //设置类名
job.setMapperClass(TokenizerMapper.class); //设置用于处理maper的类
job.setCombinerClass(KeyCombiner.class); //用于合并处理的类
job.setReducerClass(IntSumReducer.class); //设置用于处理reduce的类

job.setOutputKeyClass(Text.class); //设置输出KEY的类
job.setOutputValueClass(IntWritable.class); //设置用于输出VALUE的类


FileInputFormat.addInputPath(job, inputPath); //设置输入参数路径

FileOutputFormat.setOutputPath(job, outPath); //设置持久化输出位置
System.exit(job.waitForCompletion(true) ? 0 : 1);
}

/**
* 删除临时输出目录
*/
private static void deleteDir(){
Configuration conf = new Configuration();
FileSystem fs;
try {
fs = FileSystem.get(URI.create(hdfsPath),conf);
Path p=new Path(outPathStr);
boolean a=fs.delete(p,true);
System.out.println(a);
} catch (IOException e) {
// TODO Auto-generated catch block
e.printStackTrace();
System.out.println("delete Dir faild!");
}
}
}

import java.io.IOException;
import java.util.StringTokenizer;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.io.IntWritable;
import org.apache.hadoop.io.Text;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.mapreduce.Mapper;
import org.apache.hadoop.mapreduce.Reducer;
import org.apache.hadoop.mapreduce.lib.input.FileInputFormat;
import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;

public class WordCount {

  public static class TokenizerMapper
       extends Mapper<Object, Text, Text, IntWritable>{

    private final static IntWritable one = new IntWritable(1);
    private Text word = new Text();

    public void map(Object key, Text value, Context context
                    ) throws IOException, InterruptedException {
      StringTokenizer itr = new StringTokenizer(value.toString());
      while (itr.hasMoreTokens()) {
        word.set(itr.nextToken());
        context.write(word, one);
      }
    }
  }

  public static class IntSumReducer
       extends Reducer<Text,IntWritable,Text,IntWritable> {
    private IntWritable result = new IntWritable();

    public void reduce(Text key, Iterable<IntWritable> values,
                       Context context
                       ) throws IOException, InterruptedException {
      int sum = 0;
      for (IntWritable val : values) {
        sum += val.get();
      }
      result.set(sum);
      context.write(key, result);
    }
  }

  public static void main(String[] args) throws Exception {
    Configuration conf = new Configuration();
    Job job = Job.getInstance(conf, "word count");
    job.setJarByClass(WordCount.class);
    job.setMapperClass(TokenizerMapper.class);
    job.setCombinerClass(IntSumReducer.class);
    job.setReducerClass(IntSumReducer.class);
    job.setOutputKeyClass(Text.class);
    job.setOutputValueClass(IntWritable.class);
    FileInputFormat.addInputPath(job, new Path(args[0]));
    FileOutputFormat.setOutputPath(job, new Path(args[1]));
    System.exit(job.waitForCompletion(true) ? 0 : 1);
  }
}

tar jxvf 是解压指今
解压以bzip2压缩的文件

-j 解压命令(*.bz)
-x 释放
-v 释放时的信息
-f 指定解压文件

看看下面这些
tar -cvf a.tar a 创建文件a的tar包

tar -tvf a.tar 查看tar包包含的文件

tar -xvf a.tar 释放tar包文件

tar -rvf a.tar b 追加文件b到tar包a.tar

tar -Avf a.tar c.tar 追加c.tar包到a.tar包

tar -zcvf a.tar.gz a 创建文件a的gzip压缩的tar包

tar -ztvf a.tar.gz 查看文件a的tar压缩包内容

tar -zxvf aa.tar.gz 释放aa.tar.gz包的内容

tar -jcvf aa.tar.bz2 aa 创建文件a的bzip压缩的tar包

tar -jtvf aa.tar.bz2 查看文件a的tar压缩包内容

tar -jxvf aa.tar.bz2 释放aa.tar.bz2包的内

StringTokenizer类_Techome_新浪博客
http://blog.sina.com.cn/s/blog_45c6564d0100kmee.html

start-hbase.sh

stop-hbase.sh

status

指定列族删除

11.22.2016

2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:QTINC=/usr/lib64/qt-3.3/include
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:USER=hadoop
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:QTLIB=/usr/lib64/qt-3.3/lib
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:KDE_IS_PRELINKED=1
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:HBASE_SECURITY_LOGGER=INFO,RFAS
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:HOME=/home/hadoop
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:LESSOPEN=||/usr/bin/lesspipe.sh %s
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_START_FILE=/tmp/hbase-hadoop-regionserver.autorestart
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_LOGFILE=hbase-hadoop-regionserver-master.log
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:LANG=zh_CN.UTF-8
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_LOG_PREFIX=hbase-hadoop-regionserver-master
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_IDENT_STRING=hadoop
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: vmName=Java HotSpot(TM) 64-Bit Server VM, vmVendor=Oracle Corporation, vmVersion=24.65-b04
2016-11-22 17:33:08,111 INFO [main] util.ServerCommandLine: vmInputArguments=[-Dproc_regionserver, -XX:OnOutOfMemoryError=kill -9 %p, -Xmx8G, -XX:+UseConcMarkSweepGC, -XX:PermSize=128m, -XX:MaxPermSize=128m, -Dhbase.log.dir=/home/hadoop/soft/hbase-1.2.4/logs, -Dhbase.log.file=hbase-hadoop-regionserver-master.log, -Dhbase.home.dir=/home/hadoop/soft/hbase-1.2.4, -Dhbase.id.str=hadoop, -Dhbase.root.logger=INFO,RFA, -Djava.library.path=/home/hadoop/soft/hadoop-2.6.4/lib/native, -Dhbase.security.logger=INFO,RFAS]
2016-11-22 17:33:08,594 WARN [main] util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2016-11-22 17:33:09,651 INFO [main] regionserver.RSRpcServices: regionserver/master/192.168.150.3:16020 server-side HConnection retries=350
2016-11-22 17:33:11,094 INFO [main] ipc.SimpleRpcScheduler: Using deadline as user call queue, count=3
2016-11-22 17:33:11,261 INFO [main] ipc.RpcServer: regionserver/master/192.168.150.3:16020: started 10 reader(s) listening on port=16020
2016-11-22 17:33:12,096 INFO [main] impl.MetricsConfig: loaded properties from hadoop-metrics2-hbase.properties
2016-11-22 17:33:12,567 INFO [main] impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
2016-11-22 17:33:12,567 INFO [main] impl.MetricsSystemImpl: HBase metrics system started
2016-11-22 17:33:16,916 INFO [main] fs.HFileSystem: Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks
2016-11-22 17:33:17,793 INFO [main] zookeeper.RecoverableZooKeeper: Process identifier=regionserver:16020 connecting to ZooKeeper ensemble=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.6-1569965, built on 02/20/2014 09:09 GMT
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:host.name=master
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.version=1.7.0_67
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.home=/usr/java/jdk1.7.0_67/jre
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.class.path=/home/hadoop/soft/hbase-1.2.4/conf:/usr/java/jdk1.7.0_67/lib/tools.jar:/home/hadoop/soft/hbase-1.2.4:/home/hadoop/soft/hbase-1.2.4/lib/activation-1.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/aopalliance-1.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/apacheds-i18n-2.0.0-M15.jar:/home/hadoop/soft/hbase-1.2.4/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/home/hadoop/soft/hbase-1.2.4/lib/api-asn1-api-1.0.0-M20.jar:/home/hadoop/soft/hbase-1.2.4/lib/api-util-1.0.0-M20.jar:/home/hadoop/soft/hbase-1.2.4/lib/asm-3.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/avro-1.7.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/aws-java-sdk-core-1.10.77.jar:/home/hadoop/soft/hbase-1.2.4/lib/aws-java-sdk-s3-1.11.34.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-beanutils-1.7.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-cli-1.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-codec-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-collections-3.2.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-configuration-1.6.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-daemon-1.0.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-digester-1.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-el-1.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-httpclient-3.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-io-2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-lang-2.6.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-logging-1.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-math-2.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-math3-3.1.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-net-3.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/disruptor-3.3.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/findbugs-annotations-1.3.9-1.jar:/home/hadoop/soft/hbase-1.2.4/lib/guava-12.0.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/guice-3.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/guice-servlet-3.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-annotations-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-ant-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-archives-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-archives-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-archives-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-auth-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-aws-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-datajoin-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-datajoin-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-datajoin-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-distcp-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-distcp-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-distcp-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-extras-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-extras-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-extras-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-gridmix-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-gridmix-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-gridmix-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-nfs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-kms-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-app-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-app-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-app-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-core-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-core-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-core-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-plugins-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-plugins-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-plugins-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-shuffle-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-shuffle-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-shuffle-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-examples-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-examples-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-examples-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-nfs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-openstack-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-rumen-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-rumen-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-rumen-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-sls-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-sls-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-sls-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-streaming-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-streaming-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-streaming-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-api-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-api-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-distributedshell-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-distributedshell-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-distributedshell-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-client-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-client-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-client-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-registry-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-applicationhistoryservice-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-applicationhistoryservice-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-applicationhistoryservice-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-nodemanager-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-nodemanager-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-nodemanager-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-resourcemanager-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-resourcemanager-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-resourcemanager-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-web-proxy-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-web-proxy-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-web-proxy-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-annotations-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-annotations-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-client-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-common-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-common-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-examples-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-external-blockcache-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-hadoop2-compat-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-hadoop-compat-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-it-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-it-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-prefix-tree-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-procedure-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-protocol-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-resource-bundle-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-rest-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-server-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-server-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-shell-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-thrift-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/htrace-core-3.1.0-incubating.jar:/home/hadoop/soft/hbase-1.2.4/lib/httpclient-4.2.5.jar:/home/hadoop/soft/hbase-1.2.4/lib/httpcore-4.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-jaxrs-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-xc-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jamon-runtime-2.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/jasper-compiler-5.5.23.jar:/home/hadoop/soft/hbase-1.2.4/lib/jasper-runtime-5.5.23.jar:/home/hadoop/soft/hbase-1.2.4/lib/javax.inject-1.jar:/home/hadoop/soft/hbase-1.2.4/lib/java-xmlbuilder-0.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/jaxb-api-2.2.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/jaxb-impl-2.2.3-1.jar:/home/hadoop/soft/hbase-1.2.4/lib/jcodings-1.0.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-client-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-core-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-guice-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-json-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-server-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jets3t-0.9.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/jettison-1.3.3.jar:/home/hadoop/soft/hbase-1.2.4/lib/jetty-6.1.26.jar:/home/hadoop/soft/hbase-1.2.4/lib/jetty-sslengine-6.1.26.jar:/home/hadoop/soft/hbase-1.2.4/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hbase-1.2.4/lib/joni-2.1.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/jruby-complete-1.6.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/jsch-0.1.42.jar:/home/hadoop/soft/hbase-1.2.4/lib/jsp-2.1-6.1.14.jar:/home/hadoop/soft/hbase-1.2.4/lib/jsp-api-2.1-6.1.14.jar:/home/hadoop/soft/hbase-1.2.4/lib/junit-4.12.jar:/home/hadoop/soft/hbase-1.2.4/lib/leveldbjni-all-1.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/libthrift-0.9.3.jar:/home/hadoop/soft/hbase-1.2.4/lib/log4j-1.2.17.jar:/home/hadoop/soft/hbase-1.2.4/lib/metrics-core-2.2.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/netty-all-4.0.23.Final.jar:/home/hadoop/soft/hbase-1.2.4/lib/paranamer-2.3.jar:/home/hadoop/soft/hbase-1.2.4/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/servlet-api-2.5-6.1.14.jar:/home/hadoop/soft/hbase-1.2.4/lib/servlet-api-2.5.jar:/home/hadoop/soft/hbase-1.2.4/lib/slf4j-api-1.7.7.jar:/home/hadoop/soft/hbase-1.2.4/lib/slf4j-log4j12-1.7.5.jar:/home/hadoop/soft/hbase-1.2.4/lib/snappy-java-1.0.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/spymemcached-2.11.6.jar:/home/hadoop/soft/hbase-1.2.4/lib/xmlenc-0.52.jar:/home/hadoop/soft/hbase-1.2.4/lib/xz-1.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/zookeeper-3.4.6.jar:/home/hadoop/soft/hadoop-2.6.4/etc/hadoop:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/curator-recipes-2.6.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jettison-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jsr305-1.3.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jsp-api-2.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-lang-2.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-collections-3.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-configuration-1.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/curator-client-2.6.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-jaxrs-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/gson-2.2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/htrace-core-3.0.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/hadoop-annotations-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/httpclient-4.2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/mockito-all-1.8.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/api-asn1-api-1.0.0-M20.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/hadoop-auth-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/apacheds-i18n-2.0.0-M15.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-httpclient-3.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jersey-json-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jsch-0.1.42.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/activation-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-cli-1.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/paranamer-2.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/zookeeper-3.4.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/httpcore-4.2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-el-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/xmlenc-0.52.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/hamcrest-core-1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/guava-11.0.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-digester-1.8.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jets3t-0.9.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/curator-framework-2.6.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/api-util-1.0.0-M20.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-codec-1.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-logging-1.1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/junit-4.11.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/stax-api-1.0-2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/avro-1.7.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/servlet-api-2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jetty-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-math3-3.1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-net-3.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/xz-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/hadoop-nfs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/hadoop-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/hadoop-common-2.6.4-tests.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/htrace-core-3.0.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/xml-apis-1.3.04.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/xercesImpl-2.9.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-el-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/guava-11.0.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/hadoop-hdfs-nfs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/hadoop-hdfs-2.6.4-tests.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/hadoop-hdfs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jettison-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-lang-2.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-collections-3.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-client-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-xc-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-jaxrs-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/javax.inject-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/aopalliance-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-json-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/activation-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-cli-1.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/zookeeper-3.4.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/guice-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/guava-11.0.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-codec-1.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jline-0.9.94.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/servlet-api-2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jetty-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/xz-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-registry-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-client-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-tests-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-api-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/leveldbjni-all-1.8.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/javax.inject-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/hadoop-annotations-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/guice-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/hamcrest-core-1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/junit-4.11.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/xz-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.6.4-tests.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/contrib/capacity-scheduler/*.jar
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:java.library.path=/home/hadoop/soft/hadoop-2.6.4/lib/native
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:java.compiler=<NA>
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:os.name=Linux
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:os.arch=amd64
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:os.version=2.6.32-642.el6.x86_64
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:user.name=hadoop
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:user.home=/home/hadoop
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:user.dir=/home/hadoop/soft/hbase-1.2.4
2016-11-22 17:33:17,844 INFO [main] zookeeper.ZooKeeper: Initiating client connection, connectString=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181 sessionTimeout=90000 watcher=regionserver:160200x0, quorum=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181, baseZNode=/hbase
2016-11-22 17:33:17,953 INFO [main-SendThread(192.168.150.5:2181)] zookeeper.ClientCnxn: Opening socket connection to server 192.168.150.5/192.168.150.5:2181. Will not attempt to authenticate using SASL (unknown error)
2016-11-22 17:33:17,998 INFO [main-SendThread(192.168.150.5:2181)] zookeeper.ClientCnxn: Socket connection established to 192.168.150.5/192.168.150.5:2181, initiating session
2016-11-22 17:33:18,035 INFO [main-SendThread(192.168.150.5:2181)] zookeeper.ClientCnxn: Session establishment complete on server 192.168.150.5/192.168.150.5:2181, sessionid = 0x3588b8bd1a70002, negotiated timeout = 40000
2016-11-22 17:33:18,201 INFO [RpcServer.responder] ipc.RpcServer: RpcServer.responder: starting
2016-11-22 17:33:18,202 INFO [RpcServer.listener,port=16020] ipc.RpcServer: RpcServer.listener,port=16020: starting
2016-11-22 17:33:18,659 INFO [main] mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2016-11-22 17:33:18,670 INFO [main] http.HttpRequestLog: Http request log for http.requests.regionserver is not defined
2016-11-22 17:33:18,717 INFO [main] http.HttpServer: Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter)
2016-11-22 17:33:18,718 INFO [main] http.HttpServer: Added global filter 'clickjackingprevention' (class=org.apache.hadoop.hbase.http.ClickjackingPreventionFilter)
2016-11-22 17:33:18,726 INFO [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context regionserver
2016-11-22 17:33:18,726 INFO [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
2016-11-22 17:33:18,726 INFO [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
2016-11-22 17:33:18,810 INFO [main] http.HttpServer: Jetty bound to port 16030
2016-11-22 17:33:18,810 INFO [main] mortbay.log: jetty-6.1.26
2016-11-22 17:33:21,264 INFO [main] mortbay.log: Started SelectChannelConnector@0.0.0.0:16030
2016-11-22 17:33:21,631 INFO [regionserver/master/192.168.150.3:16020] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x66bc473e connecting to ZooKeeper ensemble=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181
2016-11-22 17:33:21,632 INFO [regionserver/master/192.168.150.3:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181 sessionTimeout=90000 watcher=hconnection-0x66bc473e0x0, quorum=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181, baseZNode=/hbase
2016-11-22 17:33:21,637 INFO [regionserver/master/192.168.150.3:16020-SendThread(192.168.150.4:2181)] zookeeper.ClientCnxn: Opening socket connection to server 192.168.150.4/192.168.150.4:2181. Will not attempt to authenticate using SASL (unknown error)
2016-11-22 17:33:21,641 INFO [regionserver/master/192.168.150.3:16020-SendThread(192.168.150.4:2181)] zookeeper.ClientCnxn: Socket connection established to 192.168.150.4/192.168.150.4:2181, initiating session
2016-11-22 17:33:21,648 INFO [regionserver/master/192.168.150.3:16020-SendThread(192.168.150.4:2181)] zookeeper.ClientCnxn: Session establishment complete on server 192.168.150.4/192.168.150.4:2181, sessionid = 0x2588b290c690004, negotiated timeout = 40000
2016-11-22 17:33:22,008 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: ClusterId : 3c1b7721-de8f-42a6-90a9-8c0a694da768
2016-11-22 17:33:22,216 INFO [regionserver/master/192.168.150.3:16020] regionserver.MemStoreFlusher: globalMemStoreLimit=3.2 G, globalMemStoreLimitLowMark=3.0 G, maxHeap=8.0 G
2016-11-22 17:33:22,241 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: CompactionChecker runs every 10sec
2016-11-22 17:33:22,362 INFO [regionserver/master/192.168.150.3:16020] regionserver.RegionServerCoprocessorHost: System coprocessor loading is enabled
2016-11-22 17:33:22,362 INFO [regionserver/master/192.168.150.3:16020] regionserver.RegionServerCoprocessorHost: Table coprocessor loading is enabled
2016-11-22 17:33:22,376 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: reportForDuty to master=master,16000,1479807189530 with port=16020, startcode=1479807192653
2016-11-22 17:33:22,996 WARN [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: reportForDuty failed; sleeping and then retrying.
2016-11-22 17:33:26,000 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: reportForDuty to master=master,16000,1479807189530 with port=16020, startcode=1479807192653
2016-11-22 17:33:26,084 INFO [regionserver/master/192.168.150.3:16020] hfile.CacheConfig: Allocating LruBlockCache size=3.20 GB, blockSize=64 KB
2016-11-22 17:33:26,103 INFO [regionserver/master/192.168.150.3:16020] hfile.CacheConfig: Created cacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=3520936, freeSize=3428966488, maxSize=3432487424, heapSize=3520936, minSize=3260862976, minFactor=0.95, multiSize=1630431488, multiFactor=0.5, singleSize=815215744, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-11-22 17:33:26,336 INFO [regionserver/master/192.168.150.3:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-11-22 17:33:26,390 INFO [regionserver/master/192.168.150.3:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=master%2C16020%2C1479807192653.default, suffix=, logDir=hdfs://192.168.150.3:9000/hbase/WALs/master,16020,1479807192653, archiveDir=hdfs://192.168.150.3:9000/hbase/oldWALs
2016-11-22 17:33:26,655 INFO [regionserver/master/192.168.150.3:16020] wal.FSHLog: Slow sync cost: 199 ms, current pipeline: []
2016-11-22 17:33:26,656 INFO [regionserver/master/192.168.150.3:16020] wal.FSHLog: New WAL /hbase/WALs/master,16020,1479807192653/master%2C16020%2C1479807192653.default.1479807206391
2016-11-22 17:33:26,742 INFO [regionserver/master/192.168.150.3:16020] regionserver.MetricsRegionServerWrapperImpl: Computing regionserver metrics every 5000 milliseconds
2016-11-22 17:33:26,761 INFO [regionserver/master/192.168.150.3:16020] regionserver.ReplicationSourceManager: Current list of replicators: [master,16020,1479807192653, slave1,16020,1479712512545, master,16020,1479712420627] other RSs: [master,16020,1479807192653, slave2,16020,1479813638869, slave1,16020,1479807188273]
2016-11-22 17:33:27,468 INFO [main-EventThread] replication.ReplicationTrackerZKImpl: /hbase/rs/slave2,16020,1479813638869 znode expired, triggering replicatorRemoved event
2016-11-22 17:33:27,490 INFO [SplitLogWorker-master:16020] regionserver.SplitLogWorker: SplitLogWorker master,16020,1479807192653 starting
2016-11-22 17:33:27,495 INFO [regionserver/master/192.168.150.3:16020] regionserver.HeapMemoryManager: Starting HeapMemoryTuner chore.
2016-11-22 17:33:27,518 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: Serving as master,16020,1479807192653, RpcServer on master/192.168.150.3:16020, sessionid=0x3588b8bd1a70002
2016-11-22 17:33:27,542 INFO [regionserver/master/192.168.150.3:16020] quotas.RegionServerQuotaManager: Quota support disabled
[hadoop@master soft]$ hbase shell
2016-11-22 17:34:14,980 WARN [main] util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/home/hadoop/soft/hbase-1.2.4/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
HBase Shell; enter 'help<RETURN>' for list of supported commands.
Type "exit<RETURN>" to leave the HBase Shell
Version 1.2.4, r67592f3d062743907f8c5ae00dbbe1ae4f69e5af, Tue Oct 25 18:10:20 CDT 2016

hbase(main):001:0> list
TABLE
0 row(s) in 0.3460 seconds

=> []
hbase(main):002:0> create 'scores','grad','course'
0 row(s) in 1.3900 seconds

=> Hbase::Table - scores
hbase(main):003:0> put 'scores','zkd','grad:','5'
0 row(s) in 0.3950 seconds

hbase(main):004:0> list
TABLE
scores
1 row(s) in 0.0220 seconds

=> ["scores"]
hbase(main):005:0> put 'scores','zkd','course:math','97'
0 row(s) in 0.0260 seconds

hbase(main):006:0> get 'scores','zkd'
COLUMN CELL
course:math timestamp=1479807433445, value=97
grad: timestamp=1479807365109, value=5
2 row(s) in 0.0730 seconds

hbase(main):007:0> scan 'scores'
ROW COLUMN+CELL
zkd column=course:math, timestamp=1479807433445, value=97
zkd column=grad:, timestamp=1479807365109, value=5
1 row(s) in 0.0450 seconds

hbase(main):008:0> list
TABLE
scores
1 row(s) in 0.0180 seconds

=> ["scores"]
hbase(main):009:0> disable 'menmber
hbase(main):010:0' disable 'menmber'
hbase(main):011:0' disable 'member'
hbase(main):012:0'
hbase(main):013:0' disable 'scores'
hbase(main):014:0' '
SyntaxError: (hbase):10: syntax error, unexpected tIDENTIFIER

disable 'menmber'
^

hbase(main):015:0> disable 'scores'
0 row(s) in 2.3940 seconds

hbase(main):016:0> drop 'scores'
0 row(s) in 1.3230 seconds

hbase(main):017:0> describe 'member'

ERROR: Unknown table member!

Here is some help for this command:
Describe the named table. For example:
hbase> describe 't1'
hbase> describe 'ns1:t1'

Alternatively, you can use the abbreviated 'desc' for the same thing.
hbase> desc 't1'
hbase> desc 'ns1:t1'


hbase(main):018:0> create 'member','member_id','address','info'
0 row(s) in 1.2870 seconds

=> Hbase::Table - member
hbase(main):019:0> list
TABLE
member
1 row(s) in 0.0120 seconds

=> ["member"]
hbase(main):020:0> describe 'member'
Table member is ENABLED
member
COLUMN FAMILIES DESCRIPTION
{NAME => 'address', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1', COMPRESSION => 'NONE', MIN_VERSIONS =>
'0', TTL => 'FOREVER', KEEP_DELETED_CELLS => 'FALSE', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}
{NAME => 'info', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1', COMPRESSION => 'NONE', MIN_VERSIONS => '0'
, TTL => 'FOREVER', KEEP_DELETED_CELLS => 'FALSE', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}
{NAME => 'member_id', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1', COMPRESSION => 'NONE', MIN_VERSIONS =
> '0', TTL => 'FOREVER', KEEP_DELETED_CELLS => 'FALSE', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}
3 row(s) in 0.0720 seconds

hbase(main):021:0> disable 'member'
0 row(s) in 2.3090 seconds

hbase(main):022:0> list
TABLE
member
1 row(s) in 0.0170 seconds

=> ["member"]
hbase(main):023:0> is_disabled 'member'
true
0 row(s) in 0.0220 seconds

hbase(main):024:0> is_disabled 'scores'

ERROR: Unknown table scores!

Here is some help for this command:
Is named table disabled? For example:
hbase> is_disabled 't1'
hbase> is_disabled 'ns1:t1'


hbase(main):025:0> enable 'member'
0 row(s) in 1.3500 seconds

hbase(main):026:0> create 'tmp_member','member'
0 row(s) in 1.2740 seconds

=> Hbase::Table - tmp_member
hbase(main):027:0> describe 'tmp_member'
Table tmp_member is ENABLED
tmp_member
COLUMN FAMILIES DESCRIPTION
{NAME => 'member', DATA_BLOCK_ENCODING => 'NONE', BLOOMFILTER => 'ROW', REPLICATION_SCOPE => '0', VERSIONS => '1', COMPRESSION => 'NONE', MIN_VERSIONS => '
0', TTL => 'FOREVER', KEEP_DELETED_CELLS => 'FALSE', BLOCKSIZE => '65536', IN_MEMORY => 'false', BLOCKCACHE => 'true'}
1 row(s) in 0.0320 seconds

hbase(main):028:0> list
TABLE
member
tmp_member
2 row(s) in 0.0160 seconds

=> ["member", "tmp_member"]
hbase(main):029:0> diable 'tmp_member'
NoMethodError: undefined method `diable' for #<Object:0x64d4ebfb>

hbase(main):030:0> disable 'tmp_member'
0 row(s) in 2.2810 seconds

hbase(main):031:0> put 'member','xueba','info:age','25'
0 row(s) in 0.0250 seconds

hbase(main):032:0> put 'member','xueba','info:birthday','1989-06-19'
0 row(s) in 0.0070 seconds

hbase(main):033:0> put 'member','xueba','info:company','tecent'
0 row(s) in 0.0070 seconds

hbase(main):034:0> put 'member','xueba','address:contry','china'
0 row(s) in 0.0070 seconds

hbase(main):035:0> put 'member','xueba','address:province','guangdong'
0 row(s) in 0.0080 seconds

hbase(main):036:0> put 'member','xueba','address:city','shenzhen'
0 row(s) in 0.0080 seconds

hbase(main):037:0> put 'member','xiaoming','info:age','24'
0 row(s) in 0.0070 seconds

hbase(main):038:0> put 'member','xiaoming','info:birthday','1990-03-22'
0 row(s) in 0.0060 seconds

hbase(main):039:0> put 'member','xiaoming','info:company','tecent'
0 row(s) in 0.0120 seconds

hbase(main):040:0> put 'member','xiaoming','info:favorite','movie'
0 row(s) in 0.0070 seconds

hbase(main):041:0> put 'member','xiaoming','address:contry','china'
0 row(s) in 0.0060 seconds

hbase(main):042:0> put 'member','xiaoming','address:province','guangdong'
0 row(s) in 0.0070 seconds

hbase(main):043:0> put 'member','xiaoming','address:city','guangzhou'
0 row(s) in 0.0100 seconds

hbase(main):044:0> scan 'member'
ROW COLUMN+CELL
xiaoming column=address:city, timestamp=1479808315381, value=guangzhou
xiaoming column=address:contry, timestamp=1479808315320, value=china
xiaoming column=address:province, timestamp=1479808315354, value=guangdong
xiaoming column=info:age, timestamp=1479808315155, value=24
xiaoming column=info:birthday, timestamp=1479808315187, value=1990-03-22
xiaoming column=info:company, timestamp=1479808315230, value=tecent
xiaoming column=info:favorite, timestamp=1479808315283, value=movie
xueba column=address:city, timestamp=1479808315123, value=shenzhen
xueba column=address:contry, timestamp=1479808315056, value=china
xueba column=address:province, timestamp=1479808315089, value=guangdong
xueba column=info:age, timestamp=1479808314937, value=25
xueba column=info:birthday, timestamp=1479808314975, value=1989-06-19
xueba column=info:company, timestamp=1479808315013, value=tecent
2 row(s) in 0.0960 seconds

hbase(main):045:0> list
TABLE
member
tmp_member
2 row(s) in 0.0220 seconds

=> ["member", "tmp_member"]
hbase(main):046:0> create 'scores','grad','course'
0 row(s) in 2.3230 seconds

=> Hbase::Table - scores
hbase(main):047:0> put 'scores','zkd','grade','5'

ERROR: Unknown column family! Valid column names: course:*, grad:*

Here is some help for this command:
Put a cell 'value' at specified table/row/column and optionally
timestamp coordinates. To put a cell value into table 'ns1:t1' or 't1'
at row 'r1' under column 'c1' marked with the time 'ts1', do:

hbase> put 'ns1:t1', 'r1', 'c1', 'value'
hbase> put 't1', 'r1', 'c1', 'value'
hbase> put 't1', 'r1', 'c1', 'value', ts1
hbase> put 't1', 'r1', 'c1', 'value', {ATTRIBUTES=>{'mykey'=>'myvalue'}}
hbase> put 't1', 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}
hbase> put 't1', 'r1', 'c1', 'value', ts1, {VISIBILITY=>'PRIVATE|SECRET'}

The same commands also can be run on a table reference. Suppose you had a reference
t to table 't1', the corresponding command would be:

hbase> t.put 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}


hbase(main):048:0> put 'scores','zkd','course:math','97'
0 row(s) in 0.0290 seconds

hbase(main):049:0> get 'scores','zkd'
COLUMN CELL
course:math timestamp=1479808523355, value=97
1 row(s) in 0.0310 seconds

hbase(main):050:0> scan

ERROR: wrong number of arguments (0 for 1)

Here is some help for this command:
Scan a table; pass table name and optionally a dictionary of scanner
specifications. Scanner specifications may include one or more of:
TIMERANGE, FILTER, LIMIT, STARTROW, STOPROW, ROWPREFIXFILTER, TIMESTAMP,
MAXLENGTH or COLUMNS, CACHE or RAW, VERSIONS, ALL_METRICS or METRICS

If no columns are specified, all columns will be scanned.
To scan all members of a column family, leave the qualifier empty as in
'col_family'.

The filter can be specified in two ways:
1. Using a filterString - more information on this is available in the
Filter Language document attached to the HBASE-4176 JIRA
2. Using the entire package name of the filter.

If you wish to see metrics regarding the execution of the scan, the
ALL_METRICS boolean should be set to true. Alternatively, if you would
prefer to see only a subset of the metrics, the METRICS array can be
defined to include the names of only the metrics you care about.

Some examples:

hbase> scan 'hbase:meta'
hbase> scan 'hbase:meta', {COLUMNS => 'info:regioninfo'}
hbase> scan 'ns1:t1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
hbase> scan 't1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
hbase> scan 't1', {COLUMNS => 'c1', TIMERANGE => [1303668804, 1303668904]}
hbase> scan 't1', {REVERSED => true}
hbase> scan 't1', {ALL_METRICS => true}
hbase> scan 't1', {METRICS => ['RPC_RETRIES', 'ROWS_FILTERED']}
hbase> scan 't1', {ROWPREFIXFILTER => 'row2', FILTER => "
(QualifierFilter (>=, 'binary:xyz')) AND (TimestampsFilter ( 123, 456))"}
hbase> scan 't1', {FILTER =>
org.apache.hadoop.hbase.filter.ColumnPaginationFilter.new(1, 0)}
hbase> scan 't1', {CONSISTENCY => 'TIMELINE'}
For setting the Operation Attributes
hbase> scan 't1', { COLUMNS => ['c1', 'c2'], ATTRIBUTES => {'mykey' => 'myvalue'}}
hbase> scan 't1', { COLUMNS => ['c1', 'c2'], AUTHORIZATIONS => ['PRIVATE','SECRET']}
For experts, there is an additional option -- CACHE_BLOCKS -- which
switches block caching for the scanner on (true) or off (false). By
default it is enabled. Examples:

hbase> scan 't1', {COLUMNS => ['c1', 'c2'], CACHE_BLOCKS => false}

Also for experts, there is an advanced option -- RAW -- which instructs the
scanner to return all cells (including delete markers and uncollected deleted
cells). This option cannot be combined with requesting specific COLUMNS.
Disabled by default. Example:

hbase> scan 't1', {RAW => true, VERSIONS => 10}

Besides the default 'toStringBinary' format, 'scan' supports custom formatting
by column. A user can define a FORMATTER by adding it to the column name in
the scan specification. The FORMATTER can be stipulated:

1. either as a org.apache.hadoop.hbase.util.Bytes method name (e.g, toInt, toString)
2. or as a custom class followed by method name: e.g. 'c(MyFormatterClass).format'.

Example formatting cf:qualifier1 and cf:qualifier2 both as Integers:
hbase> scan 't1', {COLUMNS => ['cf:qualifier1:toInt',
'cf:qualifier2:c(org.apache.hadoop.hbase.util.Bytes).toInt'] }

Note that you can specify a FORMATTER by column only (cf:qualifier). You cannot
specify a FORMATTER for all columns of a column family.

Scan can also be used directly from a table, by first getting a reference to a
table, like such:

hbase> t = get_table 't'
hbase> t.scan

Note in the above situation, you can still provide all the filtering, columns,
options, etc as described above.

hbase(main):051:0> scan 'member'
ROW COLUMN+CELL
xiaoming column=address:city, timestamp=1479808315381, value=guangzhou
xiaoming column=address:contry, timestamp=1479808315320, value=china
xiaoming column=address:province, timestamp=1479808315354, value=guangdong
xiaoming column=info:age, timestamp=1479808315155, value=24
xiaoming column=info:birthday, timestamp=1479808315187, value=1990-03-22
xiaoming column=info:company, timestamp=1479808315230, value=tecent
xiaoming column=info:favorite, timestamp=1479808315283, value=movie
xueba column=address:city, timestamp=1479808315123, value=shenzhen
xueba column=address:contry, timestamp=1479808315056, value=china
xueba column=address:province, timestamp=1479808315089, value=guangdong
xueba column=info:age, timestamp=1479808314937, value=25
xueba column=info:birthday, timestamp=1479808314975, value=1989-06-19
xueba column=info:company, timestamp=1479808315013, value=tecent
2 row(s) in 0.1080 seconds

hbase(main):052:0> list
TABLE
member
scores
tmp_member
3 row(s) in 0.0150 seconds

=> ["member", "scores", "tmp_member"]
hbase(main):053:0> get 'member','xueba'
COLUMN CELL
address:city timestamp=1479808315123, value=shenzhen
address:contry timestamp=1479808315056, value=china
address:province timestamp=1479808315089, value=guangdong
info:age timestamp=1479808314937, value=25
info:birthday timestamp=1479808314975, value=1989-06-19
info:company timestamp=1479808315013, value=tecent
6 row(s) in 0.0310 seconds

hbase(main):054:0> list
TABLE
member
scores
tmp_member
3 row(s) in 0.0200 seconds

=> ["member", "scores", "tmp_member"]
hbase(main):055:0> get 'member','xueba'
COLUMN CELL
address:city timestamp=1479808315123, value=shenzhen
address:contry timestamp=1479808315056, value=china
address:province timestamp=1479808315089, value=guangdong
info:age timestamp=1479808314937, value=25
info:birthday timestamp=1479808314975, value=1989-06-19
info:company timestamp=1479808315013, value=tecent
6 row(s) in 0.0280 seconds

hbase(main):056:0> get 'member','xueba'
COLUMN CELL
address:city timestamp=1479808315123, value=shenzhen
address:contry timestamp=1479808315056, value=china
address:province timestamp=1479808315089, value=guangdong
info:age timestamp=1479808314937, value=25
info:birthday timestamp=1479808314975, value=1989-06-19
info:company timestamp=1479808315013, value=tecent
6 row(s) in 0.0190 seconds

hbase(main):057:0> list
TABLE
member
scores
tmp_member
3 row(s) in 0.0230 seconds

=> ["member", "scores", "tmp_member"]
hbase(main):058:0> scan 'member'
ROW COLUMN+CELL
xiaoming column=address:city, timestamp=1479808315381, value=guangzhou
xiaoming column=address:contry, timestamp=1479808315320, value=china
xiaoming column=address:province, timestamp=1479808315354, value=guangdong
xiaoming column=info:age, timestamp=1479808315155, value=24
xiaoming column=info:birthday, timestamp=1479808315187, value=1990-03-22
xiaoming column=info:company, timestamp=1479808315230, value=tecent
xiaoming column=info:favorite, timestamp=1479808315283, value=movie
xueba column=address:city, timestamp=1479808315123, value=shenzhen
xueba column=address:contry, timestamp=1479808315056, value=china
xueba column=address:province, timestamp=1479808315089, value=guangdong
xueba column=info:age, timestamp=1479808314937, value=25
xueba column=info:birthday, timestamp=1479808314975, value=1989-06-19
xueba column=info:company, timestamp=1479808315013, value=tecent
2 row(s) in 0.1080 seconds

hbase(main):059:0> get 'member','xueba'
COLUMN CELL
address:city timestamp=1479808315123, value=shenzhen
address:contry timestamp=1479808315056, value=china
address:province timestamp=1479808315089, value=guangdong
info:age timestamp=1479808314937, value=25
info:birthday timestamp=1479808314975, value=1989-06-19
info:company timestamp=1479808315013, value=tecent
6 row(s) in 0.0330 seconds

hbase(main):060:0> list
TABLE
member
scores
tmp_member
3 row(s) in 0.0230 seconds

=> ["member", "scores", "tmp_member"]
hbase(main):061:0> get 'member','xueba'
COLUMN CELL
address:city timestamp=1479808315123, value=shenzhen
address:contry timestamp=1479808315056, value=china
address:province timestamp=1479808315089, value=guangdong
info:age timestamp=1479808314937, value=25
info:birthday timestamp=1479808314975, value=1989-06-19
info:company timestamp=1479808315013, value=tecent
6 row(s) in 0.0350 seconds

hbase(main):062:0> get 'member','xueba','info'
COLUMN CELL
info:age timestamp=1479808314937, value=25
info:birthday timestamp=1479808314975, value=1989-06-19
info:company timestamp=1479808315013, value=tecent
3 row(s) in 0.0240 seconds

hbase(main):063:0> get 'member','xueba','info:birthday'
COLUMN CELL
info:birthday timestamp=1479808314975, value=1989-06-19
1 row(s) in 0.0130 seconds

hbase(main):064:0> put 'member','xueba','info:age','26'
0 row(s) in 0.0120 seconds

hbase(main):065:0> put 'member','xueba','info:birthday','1990-06-22'
0 row(s) in 0.0140 seconds

hbase(main):066:0> get 'member','xueba','info:birthday'
COLUMN CELL
info:birthday timestamp=1479808994771, value=1990-06-22
1 row(s) in 0.0160 seconds

hbase(main):067:0> delete 'member','xueba','info:age'
0 row(s) in 0.0390 seconds

hbase(main):068:0> get 'member','xueba','info:age'
COLUMN CELL
0 row(s) in 0.0170 seconds

hbase(main):069:0> get 'member','xueba','info'
COLUMN CELL
info:birthday timestamp=1479808994771, value=1990-06-22
info:company timestamp=1479808315013, value=tecent
2 row(s) in 0.0190 seconds

hbase(main):070:0> get 'member','xueba'
COLUMN CELL
address:city timestamp=1479808315123, value=shenzhen
address:contry timestamp=1479808315056, value=china
address:province timestamp=1479808315089, value=guangdong
info:birthday timestamp=1479808994771, value=1990-06-22
info:company timestamp=1479808315013, value=tecent
5 row(s) in 0.0330 seconds

hbase(main):071:0> deleteall 'member','xiaoming'
0 row(s) in 0.0150 seconds

hbase(main):072:0> get 'member','xiaoming'
COLUMN CELL
0 row(s) in 0.0180 seconds

hbase(main):073:0> scan

ERROR: wrong number of arguments (0 for 1)

Here is some help for this command:
Scan a table; pass table name and optionally a dictionary of scanner
specifications. Scanner specifications may include one or more of:
TIMERANGE, FILTER, LIMIT, STARTROW, STOPROW, ROWPREFIXFILTER, TIMESTAMP,
MAXLENGTH or COLUMNS, CACHE or RAW, VERSIONS, ALL_METRICS or METRICS

If no columns are specified, all columns will be scanned.
To scan all members of a column family, leave the qualifier empty as in
'col_family'.

The filter can be specified in two ways:
1. Using a filterString - more information on this is available in the
Filter Language document attached to the HBASE-4176 JIRA
2. Using the entire package name of the filter.

If you wish to see metrics regarding the execution of the scan, the
ALL_METRICS boolean should be set to true. Alternatively, if you would
prefer to see only a subset of the metrics, the METRICS array can be
defined to include the names of only the metrics you care about.

Some examples:

hbase> scan 'hbase:meta'
hbase> scan 'hbase:meta', {COLUMNS => 'info:regioninfo'}
hbase> scan 'ns1:t1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
hbase> scan 't1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
hbase> scan 't1', {COLUMNS => 'c1', TIMERANGE => [1303668804, 1303668904]}
hbase> scan 't1', {REVERSED => true}
hbase> scan 't1', {ALL_METRICS => true}
hbase> scan 't1', {METRICS => ['RPC_RETRIES', 'ROWS_FILTERED']}
hbase> scan 't1', {ROWPREFIXFILTER => 'row2', FILTER => "
(QualifierFilter (>=, 'binary:xyz')) AND (TimestampsFilter ( 123, 456))"}
hbase> scan 't1', {FILTER =>
org.apache.hadoop.hbase.filter.ColumnPaginationFilter.new(1, 0)}
hbase> scan 't1', {CONSISTENCY => 'TIMELINE'}
For setting the Operation Attributes
hbase> scan 't1', { COLUMNS => ['c1', 'c2'], ATTRIBUTES => {'mykey' => 'myvalue'}}
hbase> scan 't1', { COLUMNS => ['c1', 'c2'], AUTHORIZATIONS => ['PRIVATE','SECRET']}
For experts, there is an additional option -- CACHE_BLOCKS -- which
switches block caching for the scanner on (true) or off (false). By
default it is enabled. Examples:

hbase> scan 't1', {COLUMNS => ['c1', 'c2'], CACHE_BLOCKS => false}

Also for experts, there is an advanced option -- RAW -- which instructs the
scanner to return all cells (including delete markers and uncollected deleted
cells). This option cannot be combined with requesting specific COLUMNS.
Disabled by default. Example:

hbase> scan 't1', {RAW => true, VERSIONS => 10}

Besides the default 'toStringBinary' format, 'scan' supports custom formatting
by column. A user can define a FORMATTER by adding it to the column name in
the scan specification. The FORMATTER can be stipulated:

1. either as a org.apache.hadoop.hbase.util.Bytes method name (e.g, toInt, toString)
2. or as a custom class followed by method name: e.g. 'c(MyFormatterClass).format'.

Example formatting cf:qualifier1 and cf:qualifier2 both as Integers:
hbase> scan 't1', {COLUMNS => ['cf:qualifier1:toInt',
'cf:qualifier2:c(org.apache.hadoop.hbase.util.Bytes).toInt'] }

Note that you can specify a FORMATTER by column only (cf:qualifier). You cannot
specify a FORMATTER for all columns of a column family.

Scan can also be used directly from a table, by first getting a reference to a
table, like such:

hbase> t = get_table 't'
hbase> t.scan

Note in the above situation, you can still provide all the filtering, columns,
options, etc as described above.

hbase(main):074:0> count 'member'
1 row(s) in 0.0410 seconds

=> 1
hbase(main):075:0> scan 'member'
ROW COLUMN+CELL
xueba column=address:city, timestamp=1479808315123, value=shenzhen
xueba column=address:contry, timestamp=1479808315056, value=china
xueba column=address:province, timestamp=1479808315089, value=guangdong
xueba column=info:birthday, timestamp=1479808994771, value=1990-06-22
xueba column=info:company, timestamp=1479808315013, value=tecent
1 row(s) in 0.0410 seconds

hbase(main):076:0> truncate 'member'
Truncating 'member' table (it may take a while):
- Disabling table...
- Truncating table...
0 row(s) in 4.0050 seconds

hbase(main):077:0> scan 'member'
ROW COLUMN+CELL
0 row(s) in 0.1440 seconds

hbase(main):078:0> lisyt
NameError: undefined local variable or method `lisyt' for #<Object:0x64d4ebfb>

hbase(main):079:0> list
TABLE
member
scores
tmp_member
3 row(s) in 0.0200 seconds

=> ["member", "scores", "tmp_member"]
hbase(main):080:0> disabled 'score'
NoMethodError: undefined method `disabled' for #<Object:0x64d4ebfb>

hbase(main):081:0> drop 'score'

ERROR: Table score does not exist.

Here is some help for this command:
Drop the named table. Table must first be disabled:
hbase> drop 't1'
hbase> drop 'ns1:t1'


hbase(main):082:0> list
TABLE
member
scores
tmp_member
3 row(s) in 0.0220 seconds

=> ["member", "scores", "tmp_member"]
hbase(main):083:0> drop 'scores'

ERROR: Table scores is enabled. Disable it first.

Here is some help for this command:
Drop the named table. Table must first be disabled:
hbase> drop 't1'
hbase> drop 'ns1:t1'


hbase(main):084:0> list
TABLE
member
scores
tmp_member
3 row(s) in 0.0190 seconds

=> ["member", "scores", "tmp_member"]
hbase(main):085:0> disabled 'scores'
NoMethodError: undefined method `disabled' for #<Object:0x64d4ebfb>

hbase(main):086:0> dis
disable disable_all disable_peer disable_table_replication display
hbase(main):086:0> disable 'scores'
0 row(s) in 2.3300 seconds

hbase(main):087:0> drop 'scores'
0 row(s) in 1.3060 seconds

hbase(main):088:0> list
TABLE
member
tmp_member
2 row(s) in 0.0200 seconds

=> ["member", "tmp_member"]
hbase(main):089:0> create 'scores','grad','course'
0 row(s) in 1.3220 seconds

=> Hbase::Table - scores
hbase(main):090:0> put 'scores','Tom','grade','5'

ERROR: Unknown column family! Valid column names: course:*, grad:*

Here is some help for this command:
Put a cell 'value' at specified table/row/column and optionally
timestamp coordinates. To put a cell value into table 'ns1:t1' or 't1'
at row 'r1' under column 'c1' marked with the time 'ts1', do:

hbase> put 'ns1:t1', 'r1', 'c1', 'value'
hbase> put 't1', 'r1', 'c1', 'value'
hbase> put 't1', 'r1', 'c1', 'value', ts1
hbase> put 't1', 'r1', 'c1', 'value', {ATTRIBUTES=>{'mykey'=>'myvalue'}}
hbase> put 't1', 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}
hbase> put 't1', 'r1', 'c1', 'value', ts1, {VISIBILITY=>'PRIVATE|SECRET'}

The same commands also can be run on a table reference. Suppose you had a reference
t to table 't1', the corresponding command would be:

hbase> t.put 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}


hbase(main):091:0> put 'scores','Tom','course:math','97'
0 row(s) in 0.0210 seconds

hbase(main):092:0> put 'scores','Tom','course:art','87'
0 row(s) in 0.0060 seconds

hbase(main):093:0> put 'scores','Jim','grade','4'

ERROR: Unknown column family! Valid column names: course:*, grad:*

Here is some help for this command:
Put a cell 'value' at specified table/row/column and optionally
timestamp coordinates. To put a cell value into table 'ns1:t1' or 't1'
at row 'r1' under column 'c1' marked with the time 'ts1', do:

hbase> put 'ns1:t1', 'r1', 'c1', 'value'
hbase> put 't1', 'r1', 'c1', 'value'
hbase> put 't1', 'r1', 'c1', 'value', ts1
hbase> put 't1', 'r1', 'c1', 'value', {ATTRIBUTES=>{'mykey'=>'myvalue'}}
hbase> put 't1', 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}
hbase> put 't1', 'r1', 'c1', 'value', ts1, {VISIBILITY=>'PRIVATE|SECRET'}

The same commands also can be run on a table reference. Suppose you had a reference
t to table 't1', the corresponding command would be:

hbase> t.put 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}


hbase(main):094:0> put 'scores','Jim','course:english','88'
0 row(s) in 0.0110 seconds

hbase(main):095:0> put 'scores','Jim','course:chinese','88'
0 row(s) in 0.0050 seconds

hbase(main):096:0> scan 'scores'
ROW COLUMN+CELL
Jim column=course:chinese, timestamp=1479809653115, value=88
Jim column=course:english, timestamp=1479809653084, value=88
Tom column=course:art, timestamp=1479809653019, value=87
Tom column=course:math, timestamp=1479809652982, value=97
2 row(s) in 0.0270 seconds

hbase(main):097:0> get 'scores','Tom'
COLUMN CELL
course:art timestamp=1479809653019, value=87
course:math timestamp=1479809652982, value=97
2 row(s) in 0.0210 seconds

hbase(main):098:0> put 'scores','Tom','course:math','100'
0 row(s) in 0.0100 seconds

hbase(main):099:0> get 'scores','Tom'
COLUMN CELL
course:art timestamp=1479809653019, value=87
course:math timestamp=1479809698610, value=100
2 row(s) in 0.0210 seconds

hbase(main):100:0> get 'scores','Tom',{COLUMN=>'course:math',VERSIONS=>3}
COLUMN CELL
course:math timestamp=1479809698610, value=100
1 row(s) in 0.0250 seconds

hbase(main):101:0> quit
[hadoop@master soft]$

[root@slave1 ~]# su hadoop
[hadoop@slave1 root]$ cd ~
[hadoop@slave1 ~]$ zkServer.sh start

[hadoop@slave1 ~]$ start-all.sh

[hadoop@slave1 ~]$ start-hbase.sh

[hadoop@slave1 ~]$ hbase shell

hbase(main):002:0> version
1.2.4, r67592f3d062743907f8c5ae00dbbe1ae4f69e5af, Tue Oct 25 18:10:20 CDT 2016

hbase(main):003:0> exit

2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:QTINC=/usr/lib64/qt-3.3/include
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:USER=hadoop
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:QTLIB=/usr/lib64/qt-3.3/lib
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:KDE_IS_PRELINKED=1
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:HBASE_SECURITY_LOGGER=INFO,RFAS
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:HOME=/home/hadoop
2016-11-22 17:33:08,101 INFO [main] util.ServerCommandLine: env:LESSOPEN=||/usr/bin/lesspipe.sh %s
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_START_FILE=/tmp/hbase-hadoop-regionserver.autorestart
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_LOGFILE=hbase-hadoop-regionserver-master.log
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:LANG=zh_CN.UTF-8
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:SSH_ASKPASS=/usr/libexec/openssh/gnome-ssh-askpass
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_LOG_PREFIX=hbase-hadoop-regionserver-master
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: env:HBASE_IDENT_STRING=hadoop
2016-11-22 17:33:08,102 INFO [main] util.ServerCommandLine: vmName=Java HotSpot(TM) 64-Bit Server VM, vmVendor=Oracle Corporation, vmVersion=24.65-b04
2016-11-22 17:33:08,111 INFO [main] util.ServerCommandLine: vmInputArguments=[-Dproc_regionserver, -XX:OnOutOfMemoryError=kill -9 %p, -Xmx8G, -XX:+UseConcMarkSweepGC, -XX:PermSize=128m, -XX:MaxPermSize=128m, -Dhbase.log.dir=/home/hadoop/soft/hbase-1.2.4/logs, -Dhbase.log.file=hbase-hadoop-regionserver-master.log, -Dhbase.home.dir=/home/hadoop/soft/hbase-1.2.4, -Dhbase.id.str=hadoop, -Dhbase.root.logger=INFO,RFA, -Djava.library.path=/home/hadoop/soft/hadoop-2.6.4/lib/native, -Dhbase.security.logger=INFO,RFAS]
2016-11-22 17:33:08,594 WARN [main] util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
2016-11-22 17:33:09,651 INFO [main] regionserver.RSRpcServices: regionserver/master/192.168.150.3:16020 server-side HConnection retries=350
2016-11-22 17:33:11,094 INFO [main] ipc.SimpleRpcScheduler: Using deadline as user call queue, count=3
2016-11-22 17:33:11,261 INFO [main] ipc.RpcServer: regionserver/master/192.168.150.3:16020: started 10 reader(s) listening on port=16020
2016-11-22 17:33:12,096 INFO [main] impl.MetricsConfig: loaded properties from hadoop-metrics2-hbase.properties
2016-11-22 17:33:12,567 INFO [main] impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
2016-11-22 17:33:12,567 INFO [main] impl.MetricsSystemImpl: HBase metrics system started
2016-11-22 17:33:16,916 INFO [main] fs.HFileSystem: Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks
2016-11-22 17:33:17,793 INFO [main] zookeeper.RecoverableZooKeeper: Process identifier=regionserver:16020 connecting to ZooKeeper ensemble=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.6-1569965, built on 02/20/2014 09:09 GMT
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:host.name=master
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.version=1.7.0_67
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.home=/usr/java/jdk1.7.0_67/jre
2016-11-22 17:33:17,835 INFO [main] zookeeper.ZooKeeper: Client environment:java.class.path=/home/hadoop/soft/hbase-1.2.4/conf:/usr/java/jdk1.7.0_67/lib/tools.jar:/home/hadoop/soft/hbase-1.2.4:/home/hadoop/soft/hbase-1.2.4/lib/activation-1.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/aopalliance-1.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/apacheds-i18n-2.0.0-M15.jar:/home/hadoop/soft/hbase-1.2.4/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/home/hadoop/soft/hbase-1.2.4/lib/api-asn1-api-1.0.0-M20.jar:/home/hadoop/soft/hbase-1.2.4/lib/api-util-1.0.0-M20.jar:/home/hadoop/soft/hbase-1.2.4/lib/asm-3.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/avro-1.7.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/aws-java-sdk-core-1.10.77.jar:/home/hadoop/soft/hbase-1.2.4/lib/aws-java-sdk-s3-1.11.34.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-beanutils-1.7.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-cli-1.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-codec-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-collections-3.2.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-configuration-1.6.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-daemon-1.0.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-digester-1.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-el-1.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-httpclient-3.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-io-2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-lang-2.6.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-logging-1.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-math-2.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-math3-3.1.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/commons-net-3.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/disruptor-3.3.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/findbugs-annotations-1.3.9-1.jar:/home/hadoop/soft/hbase-1.2.4/lib/guava-12.0.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/guice-3.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/guice-servlet-3.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-annotations-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-ant-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-archives-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-archives-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-archives-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-auth-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-aws-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-datajoin-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-datajoin-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-datajoin-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-distcp-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-distcp-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-distcp-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-extras-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-extras-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-extras-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-gridmix-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-gridmix-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-gridmix-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-hdfs-nfs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-kms-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-app-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-app-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-app-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-core-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-core-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-core-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-plugins-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-plugins-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-hs-plugins-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-jobclient-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-shuffle-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-shuffle-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-client-shuffle-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-examples-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-examples-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-mapreduce-examples-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-nfs-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-openstack-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-rumen-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-rumen-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-rumen-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-sls-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-sls-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-sls-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-streaming-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-streaming-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-streaming-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-api-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-api-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-distributedshell-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-distributedshell-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-distributedshell-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-client-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-client-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-client-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-registry-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-applicationhistoryservice-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-applicationhistoryservice-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-applicationhistoryservice-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-common-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-common-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-common-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-nodemanager-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-nodemanager-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-nodemanager-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-resourcemanager-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-resourcemanager-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-resourcemanager-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-tests-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-web-proxy-2.6.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-web-proxy-2.6.4-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hadoop-yarn-server-web-proxy-2.6.4-test-sources.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-annotations-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-annotations-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-client-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-common-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-common-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-examples-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-external-blockcache-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-hadoop2-compat-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-hadoop-compat-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-it-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-it-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-prefix-tree-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-procedure-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-protocol-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-resource-bundle-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-rest-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-server-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-server-1.2.4-tests.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-shell-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/hbase-thrift-1.2.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/htrace-core-3.1.0-incubating.jar:/home/hadoop/soft/hbase-1.2.4/lib/httpclient-4.2.5.jar:/home/hadoop/soft/hbase-1.2.4/lib/httpcore-4.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-jaxrs-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jackson-xc-1.9.13.jar:/home/hadoop/soft/hbase-1.2.4/lib/jamon-runtime-2.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/jasper-compiler-5.5.23.jar:/home/hadoop/soft/hbase-1.2.4/lib/jasper-runtime-5.5.23.jar:/home/hadoop/soft/hbase-1.2.4/lib/javax.inject-1.jar:/home/hadoop/soft/hbase-1.2.4/lib/java-xmlbuilder-0.4.jar:/home/hadoop/soft/hbase-1.2.4/lib/jaxb-api-2.2.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/jaxb-impl-2.2.3-1.jar:/home/hadoop/soft/hbase-1.2.4/lib/jcodings-1.0.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-client-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-core-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-guice-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-json-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jersey-server-1.9.jar:/home/hadoop/soft/hbase-1.2.4/lib/jets3t-0.9.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/jettison-1.3.3.jar:/home/hadoop/soft/hbase-1.2.4/lib/jetty-6.1.26.jar:/home/hadoop/soft/hbase-1.2.4/lib/jetty-sslengine-6.1.26.jar:/home/hadoop/soft/hbase-1.2.4/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hbase-1.2.4/lib/joni-2.1.2.jar:/home/hadoop/soft/hbase-1.2.4/lib/jruby-complete-1.6.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/jsch-0.1.42.jar:/home/hadoop/soft/hbase-1.2.4/lib/jsp-2.1-6.1.14.jar:/home/hadoop/soft/hbase-1.2.4/lib/jsp-api-2.1-6.1.14.jar:/home/hadoop/soft/hbase-1.2.4/lib/junit-4.12.jar:/home/hadoop/soft/hbase-1.2.4/lib/leveldbjni-all-1.8.jar:/home/hadoop/soft/hbase-1.2.4/lib/libthrift-0.9.3.jar:/home/hadoop/soft/hbase-1.2.4/lib/log4j-1.2.17.jar:/home/hadoop/soft/hbase-1.2.4/lib/metrics-core-2.2.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/netty-all-4.0.23.Final.jar:/home/hadoop/soft/hbase-1.2.4/lib/paranamer-2.3.jar:/home/hadoop/soft/hbase-1.2.4/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/servlet-api-2.5-6.1.14.jar:/home/hadoop/soft/hbase-1.2.4/lib/servlet-api-2.5.jar:/home/hadoop/soft/hbase-1.2.4/lib/slf4j-api-1.7.7.jar:/home/hadoop/soft/hbase-1.2.4/lib/slf4j-log4j12-1.7.5.jar:/home/hadoop/soft/hbase-1.2.4/lib/snappy-java-1.0.4.1.jar:/home/hadoop/soft/hbase-1.2.4/lib/spymemcached-2.11.6.jar:/home/hadoop/soft/hbase-1.2.4/lib/xmlenc-0.52.jar:/home/hadoop/soft/hbase-1.2.4/lib/xz-1.0.jar:/home/hadoop/soft/hbase-1.2.4/lib/zookeeper-3.4.6.jar:/home/hadoop/soft/hadoop-2.6.4/etc/hadoop:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/curator-recipes-2.6.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jettison-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jsr305-1.3.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jsp-api-2.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-lang-2.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-collections-3.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-configuration-1.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/curator-client-2.6.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-jaxrs-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/gson-2.2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/htrace-core-3.0.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/hadoop-annotations-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/httpclient-4.2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/mockito-all-1.8.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/api-asn1-api-1.0.0-M20.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/hadoop-auth-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/apacheds-i18n-2.0.0-M15.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-httpclient-3.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jersey-json-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jsch-0.1.42.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/activation-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-cli-1.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/paranamer-2.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/zookeeper-3.4.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/httpcore-4.2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-el-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/xmlenc-0.52.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/hamcrest-core-1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/guava-11.0.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-digester-1.8.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jets3t-0.9.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/curator-framework-2.6.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/api-util-1.0.0-M20.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-codec-1.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-logging-1.1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/junit-4.11.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/stax-api-1.0-2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/avro-1.7.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/servlet-api-2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/jetty-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-math3-3.1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/commons-net-3.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/lib/xz-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/hadoop-nfs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/hadoop-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/common/hadoop-common-2.6.4-tests.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/htrace-core-3.0.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/xml-apis-1.3.04.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/xercesImpl-2.9.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-el-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/guava-11.0.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/hadoop-hdfs-nfs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/hadoop-hdfs-2.6.4-tests.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/hdfs/hadoop-hdfs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jettison-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-lang-2.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-collections-3.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-client-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-xc-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-jaxrs-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/javax.inject-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/aopalliance-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-json-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/activation-1.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-cli-1.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/zookeeper-3.4.6.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/guice-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/guava-11.0.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-codec-1.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jline-0.9.94.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/servlet-api-2.5.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/jetty-6.1.26.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/lib/xz-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-registry-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-client-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-tests-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-api-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/yarn/hadoop-yarn-server-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/leveldbjni-all-1.8.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/asm-3.2.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/javax.inject-1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/hadoop-annotations-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/guice-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/hamcrest-core-1.3.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jackson-core-asl-1.9.13.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/junit-4.11.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/lib/xz-1.0.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.6.4-tests.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.6.4.jar:/home/hadoop/soft/hadoop-2.6.4/contrib/capacity-scheduler/*.jar
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:java.library.path=/home/hadoop/soft/hadoop-2.6.4/lib/native
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:java.compiler=<NA>
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:os.name=Linux
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:os.arch=amd64
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:os.version=2.6.32-642.el6.x86_64
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:user.name=hadoop
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:user.home=/home/hadoop
2016-11-22 17:33:17,836 INFO [main] zookeeper.ZooKeeper: Client environment:user.dir=/home/hadoop/soft/hbase-1.2.4
2016-11-22 17:33:17,844 INFO [main] zookeeper.ZooKeeper: Initiating client connection, connectString=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181 sessionTimeout=90000 watcher=regionserver:160200x0, quorum=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181, baseZNode=/hbase
2016-11-22 17:33:17,953 INFO [main-SendThread(192.168.150.5:2181)] zookeeper.ClientCnxn: Opening socket connection to server 192.168.150.5/192.168.150.5:2181. Will not attempt to authenticate using SASL (unknown error)
2016-11-22 17:33:17,998 INFO [main-SendThread(192.168.150.5:2181)] zookeeper.ClientCnxn: Socket connection established to 192.168.150.5/192.168.150.5:2181, initiating session
2016-11-22 17:33:18,035 INFO [main-SendThread(192.168.150.5:2181)] zookeeper.ClientCnxn: Session establishment complete on server 192.168.150.5/192.168.150.5:2181, sessionid = 0x3588b8bd1a70002, negotiated timeout = 40000
2016-11-22 17:33:18,201 INFO [RpcServer.responder] ipc.RpcServer: RpcServer.responder: starting
2016-11-22 17:33:18,202 INFO [RpcServer.listener,port=16020] ipc.RpcServer: RpcServer.listener,port=16020: starting
2016-11-22 17:33:18,659 INFO [main] mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2016-11-22 17:33:18,670 INFO [main] http.HttpRequestLog: Http request log for http.requests.regionserver is not defined
2016-11-22 17:33:18,717 INFO [main] http.HttpServer: Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter)
2016-11-22 17:33:18,718 INFO [main] http.HttpServer: Added global filter 'clickjackingprevention' (class=org.apache.hadoop.hbase.http.ClickjackingPreventionFilter)
2016-11-22 17:33:18,726 INFO [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context regionserver
2016-11-22 17:33:18,726 INFO [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
2016-11-22 17:33:18,726 INFO [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
2016-11-22 17:33:18,810 INFO [main] http.HttpServer: Jetty bound to port 16030
2016-11-22 17:33:18,810 INFO [main] mortbay.log: jetty-6.1.26
2016-11-22 17:33:21,264 INFO [main] mortbay.log: Started SelectChannelConnector@0.0.0.0:16030
2016-11-22 17:33:21,631 INFO [regionserver/master/192.168.150.3:16020] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x66bc473e connecting to ZooKeeper ensemble=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181
2016-11-22 17:33:21,632 INFO [regionserver/master/192.168.150.3:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181 sessionTimeout=90000 watcher=hconnection-0x66bc473e0x0, quorum=192.168.150.3:2181,192.168.150.4:2181,192.168.150.5:2181, baseZNode=/hbase
2016-11-22 17:33:21,637 INFO [regionserver/master/192.168.150.3:16020-SendThread(192.168.150.4:2181)] zookeeper.ClientCnxn: Opening socket connection to server 192.168.150.4/192.168.150.4:2181. Will not attempt to authenticate using SASL (unknown error)
2016-11-22 17:33:21,641 INFO [regionserver/master/192.168.150.3:16020-SendThread(192.168.150.4:2181)] zookeeper.ClientCnxn: Socket connection established to 192.168.150.4/192.168.150.4:2181, initiating session
2016-11-22 17:33:21,648 INFO [regionserver/master/192.168.150.3:16020-SendThread(192.168.150.4:2181)] zookeeper.ClientCnxn: Session establishment complete on server 192.168.150.4/192.168.150.4:2181, sessionid = 0x2588b290c690004, negotiated timeout = 40000
2016-11-22 17:33:22,008 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: ClusterId : 3c1b7721-de8f-42a6-90a9-8c0a694da768
2016-11-22 17:33:22,216 INFO [regionserver/master/192.168.150.3:16020] regionserver.MemStoreFlusher: globalMemStoreLimit=3.2 G, globalMemStoreLimitLowMark=3.0 G, maxHeap=8.0 G
2016-11-22 17:33:22,241 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: CompactionChecker runs every 10sec
2016-11-22 17:33:22,362 INFO [regionserver/master/192.168.150.3:16020] regionserver.RegionServerCoprocessorHost: System coprocessor loading is enabled
2016-11-22 17:33:22,362 INFO [regionserver/master/192.168.150.3:16020] regionserver.RegionServerCoprocessorHost: Table coprocessor loading is enabled
2016-11-22 17:33:22,376 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: reportForDuty to master=master,16000,1479807189530 with port=16020, startcode=1479807192653
2016-11-22 17:33:22,996 WARN [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: reportForDuty failed; sleeping and then retrying.
2016-11-22 17:33:26,000 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: reportForDuty to master=master,16000,1479807189530 with port=16020, startcode=1479807192653
2016-11-22 17:33:26,084 INFO [regionserver/master/192.168.150.3:16020] hfile.CacheConfig: Allocating LruBlockCache size=3.20 GB, blockSize=64 KB
2016-11-22 17:33:26,103 INFO [regionserver/master/192.168.150.3:16020] hfile.CacheConfig: Created cacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=3520936, freeSize=3428966488, maxSize=3432487424, heapSize=3520936, minSize=3260862976, minFactor=0.95, multiSize=1630431488, multiFactor=0.5, singleSize=815215744, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-11-22 17:33:26,336 INFO [regionserver/master/192.168.150.3:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-11-22 17:33:26,390 INFO [regionserver/master/192.168.150.3:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=master%2C16020%2C1479807192653.default, suffix=, logDir=hdfs://192.168.150.3:9000/hbase/WALs/master,16020,1479807192653, archiveDir=hdfs://192.168.150.3:9000/hbase/oldWALs
2016-11-22 17:33:26,655 INFO [regionserver/master/192.168.150.3:16020] wal.FSHLog: Slow sync cost: 199 ms, current pipeline: []
2016-11-22 17:33:26,656 INFO [regionserver/master/192.168.150.3:16020] wal.FSHLog: New WAL /hbase/WALs/master,16020,1479807192653/master%2C16020%2C1479807192653.default.1479807206391
2016-11-22 17:33:26,742 INFO [regionserver/master/192.168.150.3:16020] regionserver.MetricsRegionServerWrapperImpl: Computing regionserver metrics every 5000 milliseconds
2016-11-22 17:33:26,761 INFO [regionserver/master/192.168.150.3:16020] regionserver.ReplicationSourceManager: Current list of replicators: [master,16020,1479807192653, slave1,16020,1479712512545, master,16020,1479712420627] other RSs: [master,16020,1479807192653, slave2,16020,1479813638869, slave1,16020,1479807188273]
2016-11-22 17:33:27,468 INFO [main-EventThread] replication.ReplicationTrackerZKImpl: /hbase/rs/slave2,16020,1479813638869 znode expired, triggering replicatorRemoved event
2016-11-22 17:33:27,490 INFO [SplitLogWorker-master:16020] regionserver.SplitLogWorker: SplitLogWorker master,16020,1479807192653 starting
2016-11-22 17:33:27,495 INFO [regionserver/master/192.168.150.3:16020] regionserver.HeapMemoryManager: Starting HeapMemoryTuner chore.
2016-11-22 17:33:27,518 INFO [regionserver/master/192.168.150.3:16020] regionserver.HRegionServer: Serving as master,16020,1479807192653, RpcServer on master/192.168.150.3:16020, sessionid=0x3588b8bd1a70002
2016-11-22 17:33:27,542 INFO [regionserver/master/192.168.150.3:16020] quotas.RegionServerQuotaManager: Quota support disabled

11.26.2016

1.新建项目

2.导包;

3.建项目内容

package abcd;

import java.io.IOException;
import java.util.ArrayList;
import java.util.LinkedList;
import java.util.List;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.Cell;
import org.apache.hadoop.hbase.CellUtil;
import org.apache.hadoop.hbase.HBaseConfiguration;
import org.apache.hadoop.hbase.HColumnDescriptor;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.MasterNotRunningException;
import org.apache.hadoop.hbase.TableName;
import org.apache.hadoop.hbase.ZooKeeperConnectionException;
import org.apache.hadoop.hbase.client.Admin;
import org.apache.hadoop.hbase.client.Connection;
import org.apache.hadoop.hbase.client.ConnectionFactory;
import org.apache.hadoop.hbase.client.Get;
import org.apache.hadoop.hbase.client.HConnection;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.client.Result;
import org.apache.hadoop.hbase.client.ResultScanner;
import org.apache.hadoop.hbase.client.Scan;
import org.apache.hadoop.hbase.client.Table;
import org.apache.hadoop.hbase.io.compress.Compression.Algorithm;
import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding;
import org.apache.hadoop.hbase.util.Bytes;

public class HbaseConnection {

private Configuration conf;
//private HConnection hConn = null;
private Connection connection = null;

public HbaseConnection(String zkServer) throws IOException {

conf = HBaseConfiguration.create();
conf.set("hbase.zookeeper.quorum", zkServer);
//hConn = HConnectionManager.createConnection(conf); //过时的方法
connection = ConnectionFactory.createConnection(conf);

/*
*
*/
}

public void createTable(String tableName, List<String> cols) {
try {
//HBaseAdmin admin = new HBaseAdmin(conf);
Admin admin = connection.getAdmin();
//if (admin.tableExists(tableName)) {
if (admin.tableExists(TableName.valueOf(tableName))) {
throw new IOException("table exists");
} else {
//HTableDescriptor tableDesc = new HTableDescriptor(tableName);
HTableDescriptor tableDesc = new HTableDescriptor(TableName.valueOf(tableName));
for (String col : cols) {
HColumnDescriptor colDesc = new HColumnDescriptor(col);
colDesc.setCompressionType(Algorithm.GZ);//进行compaction的时候使用压缩算法
colDesc.setDataBlockEncoding(DataBlockEncoding.DIFF);////压缩内存和存储中的数据,内存紧张的时候设置
tableDesc.addFamily(colDesc);
}
admin.createTable(tableDesc);
}
} catch (MasterNotRunningException e) {
// TODO Auto-generated catch block
e.printStackTrace();
} catch (ZooKeeperConnectionException e) {
// TODO Auto-generated catch block
e.printStackTrace();
} catch (IOException e) {
// TODO Auto-generated catch block
e.printStackTrace();
}
}
public void saveData(String tableName, List<Put> puts) {
try {
//HTableInterface table = hConn.getTable(tableName);
//Admin admin = connection.getAdmin();

Table table = connection.getTable(TableName.valueOf(tableName));
table.put(puts);

//table.setAutoFlush(false);
//table.flushCommits();

} catch (IOException e) {
// TODO Auto-generated catch block
e.printStackTrace();
}
}
public Result getData(String tablename, String rowkey) {
try {
// HTableInterface table = hConn.getTable(tablename);
Table table = connection.getTable(TableName.valueOf(tablename));
Get get = new Get(Bytes.toBytes(rowkey));
return table.get(get);
} catch (IOException e) {
// TODO Auto-generated catch block
e.printStackTrace();
}
return null;
}
public void format(Result result) {



for (Cell rowKV : result.rawCells()) {
System.out.print("Row Name: " + new String(CellUtil.cloneRow(rowKV)) + " ");
System.out.print("Timestamp: " + rowKV.getTimestamp() + " ");
System.out.print("column Family: " + new String(CellUtil.cloneFamily(rowKV)) + " ");
System.out.print("column qualifier: " + new String(CellUtil.cloneQualifier(rowKV)) + " ");
System.out.println("Value: " + new String(CellUtil.cloneValue(rowKV)) + " ");
}
}
public void hbaseScan(String tableName){
Scan scan = new Scan();
scan.setCaching(1000);
try {
//HTableInterface table = hConn.getTable(tableName);
Table table = connection.getTable(TableName.valueOf(tableName));
ResultScanner scanner = table.getScanner(scan);
for (Result res : scanner) {
format(res);
}
table.close();
} catch (IOException e) {
// TODO Auto-generated catch block
e.printStackTrace();
}
}

public static void main(String[] args) throws IOException {
// HbaseConnection conn = new HbaseConnection("slave1,slave2");
HbaseConnection conn = new HbaseConnection("192.168.150.3,192.168.150.4,192.168.150.5");
ArrayList<String> list = new ArrayList<String>();//列族名列表
list.add("basicInfo");
list.add("moreInfo");
conn.createTable("person", list);

// List<Put> puts = new LinkedList<Put>();
// Put put1 = new Put("Tom".getBytes());
// // put1.add(Bytes.toBytes("basicInfo"), Bytes.toBytes("age"),Bytes.toBytes("27"));
// // put1.add(Bytes.toBytes("moreInfo"), Bytes.toBytes("tel"),Bytes.toBytes("111"));
// put1.addColumn(Bytes.toBytes("basicInfo"), Bytes.toBytes("age"),
// Bytes.toBytes("27"));
// put1.addColumn(Bytes.toBytes("moreInfo"), Bytes.toBytes("tel"),
// Bytes.toBytes("111"));
//
// Put put2 = new Put("Jim".getBytes());
// // put2.add(Bytes.toBytes("basicInfo"), Bytes.toBytes("age"),Bytes.toBytes("28"));
// // put2.add(Bytes.toBytes("moreInfo"), Bytes.toBytes("tel"),Bytes.toBytes("112"));
// put2.addColumn(Bytes.toBytes("basicInfo"), Bytes.toBytes("age"),
// Bytes.toBytes("28"));
// put2.addColumn(Bytes.toBytes("moreInfo"), Bytes.toBytes("tel"),
// Bytes.toBytes("112"));
// puts.add(put1);
// puts.add(put2);
// conn.saveData("person", puts);
//
//
// Result result = conn.getData("person", "Tom");
// conn.format(result);
//
// conn.hbaseScan("person");
}

}

 

# Copyright (c) 1993-2009 Microsoft Corp.
#
# This is a sample HOSTS file used by Microsoft TCP/IP for Windows.
#
# This file contains the mappings of IP addresses to host names. Each
# entry should be kept on an individual line. The IP address should
# be placed in the first column followed by the corresponding host name.
# The IP address and the host name should be separated by at least one
# space.
#
# Additionally, comments (such as these) may be inserted on individual
# lines or following the machine name denoted by a '#' symbol.
#
# For example:
#
# 102.54.94.97 rhino.acme.com # source server
# 38.25.63.10 x.acme.com # x client host

# localhost name resolution is handled within DNS itself.
# 127.0.0.1 localhost
# ::1 localhost
192.168.150.3 master
192.168.150.4 slave1
192.168.150.5 slave2

 在windows  的C:WindowsSystem32driversetchosts目录下 在文档末尾上加上

 这是两个步骤

[hadoop@master ~]$ start-all.sh
This script is Deprecated. Instead use start-dfs.sh and start-yarn.sh

[hadoop@master ~]$ hbase shell

[hadoop@master ~]$ jps
2675 SecondaryNameNode
2375 QuorumPeerMain
3462 Jps
2845 ResourceManager
2517 NameNode
3380 HRegionServer
3246 HMaster

http://www.cnblogs.com/shishanyuan/p/4172806.html

http://blog.csdn.net/whuqin/article/details/6728603

http://wenku.baidu.com/link?url=iTigQgQkM6lgpf3HKPuJcS4_X2_WPNoUoUF0jEQnTQzhgjKYmyrDJs-gCwzPDdOPawmqGzqGCNZSie3lcZqBaQcSSxvDcF4HqrWGgkCtCbW

沙耶加是最棒的也是最强的,今天换我来拯救世界!
原文地址:https://www.cnblogs.com/wanghui626/p/6012311.html