Support Questions
Find answers, ask questions, and share your expertise
Announcements
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

Super Collaborator

@dengke li

Can you attach master log ?

You should find it under /grid/0/log/hbase/

Consider increasing hbase.master.namespace.init.timeout

Default is 300000 ms

Highlighted

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

New Contributor

hello,

i use hdp 2.4, and the hbase master log is under /var/log/hbase, not under /

grid/0/log/hbase/, and in hbase master log is :

2016-09-01 14:19:49,843 ERROR [Thread-68] master.HMaster: Master failed to complete initialization after 900000ms. Please consider submitting a bug report including a thread dump of this process.

how to increase

hbase.master.namespace.init.timeout,

in which configuration file?

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

New Contributor

and in hbase log there is also

2016-09-01 14:38:37,876 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics

but i think it is not problem of hbase

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

Super Collaborator

Can you attach the master log so that we can better help you ?

You can add hbase.master.namespace.init.timeout to hbase-site.xml by using Ambari.

Still, finding the root cause is desirable.

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

New Contributor

here is the log of hbase master:

/var/log/hbase$ tail -f hbase-hbase-master-ec2-13-26-256-150.eu-central-1.compute.amazonaws.com.log 2016-09-01 16:17:34,702 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:17:34,703 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:14,705 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:14,706 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:34,702 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:34,703 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:34,705 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:34,709 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:44,702 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics 2016-09-01 16:18:54,707 WARN [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://http://ec2-52-29-246-252.eu-central-1.compute.amazonaws.com:6188/ws/v1/timeline/metrics

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

New Contributor

in fact, i firstly use command :

sudo -u hdfs hdfs dfsadmin -safemode forceExit

to let namenode leave safemode, i don't know if it will affect hbase or not

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

Super Collaborator

We need more of the master log beyond what tail showed us.

Please attach master log.

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

New Contributor

hello,

when i install hdp, there is error for hbase check, the log is :

can some one help me?

2016-09-10 12:34:21,313 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
2016-09-10 12:34:21,320 - File['/var/lib/ambari-agent/tmp/hbaseSmokeVerify.sh'] {'content': StaticFile('hbaseSmokeVerify.sh'), 'mode': 0755}
2016-09-10 12:34:21,326 - File['/var/lib/ambari-agent/tmp/hbase-smoke.sh'] {'content': Template('hbase-smoke.sh.j2'), 'mode': 0755}
2016-09-10 12:34:21,327 - Writing File['/var/lib/ambari-agent/tmp/hbase-smoke.sh'] because contents don't match
2016-09-10 12:34:21,327 - Execute[' /usr/hdp/current/hbase-client/bin/hbase --config /usr/hdp/current/hbase-client/conf shell /var/lib/ambari-agent/tmp/hbase-smoke.sh && /var/lib/ambari-agent/tmp/hbaseSmokeVerify.sh /usr/hdp/current/hbase-client/conf id1fac351e_date341016 /usr/hdp/current/hbase-client/bin/hbase'] {'logoutput': True, 'tries': 6, 'user': 'ambari-qa', 'try_sleep': 5}
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]

ERROR: Table ambarismoketest does not exist.

Here is some help for this command:
Start disable of named table:
  hbase> disable 't1'
  hbase> disable 'ns1:t1'
ERROR: Table ambarismoketest does not exist.

Here is some help for this command:
Drop the named table. Table must first be disabled:
  hbase> drop 't1'
  hbase> drop 'ns1:t1'
0 row(s) in 28.2970 seconds

2016-09-10 12:34:58,865 ERROR [main] client.AsyncProcess: Failed to get region location 
org.apache.hadoop.hbase.client.NoServerForRegionException: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row row01
	at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1299)
	at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1162)
	at org.apache.hadoop.hbase.client.AsyncProcess.submit(AsyncProcess.java:370)
	at org.apache.hadoop.hbase.client.AsyncProcess.submit(AsyncProcess.java:321)
	at org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:206)
	at org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:183)
	at org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1449)
	at org.apache.hadoop.hbase.client.HTable.put(HTable.java:1040)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:497)
	at org.jruby.javasupport.JavaMethod.invokeDirectWithExceptionHandling(JavaMethod.java:450)
	at org.jruby.javasupport.JavaMethod.invokeDirect(JavaMethod.java:311)
	at org.jruby.java.invokers.InstanceMethodInvoker.call(InstanceMethodInvoker.java:59)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169)
	at org.jruby.ast.CallOneArgNode.interpret(CallOneArgNode.java:57)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.CallManyArgsNode.interpret(CallManyArgsNode.java:59)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_BLOCK(ASTInterpreter.java:111)
	at org.jruby.runtime.InterpretedBlock.evalBlockBody(InterpretedBlock.java:374)
	at org.jruby.runtime.InterpretedBlock.yield(InterpretedBlock.java:295)
	at org.jruby.runtime.InterpretedBlock.yieldSpecific(InterpretedBlock.java:229)
	at org.jruby.runtime.Block.yieldSpecific(Block.java:99)
	at org.jruby.ast.ZYieldNode.interpret(ZYieldNode.java:25)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:169)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:191)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:302)
	at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:144)
	at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:153)
	at org.jruby.ast.FCallNoArgBlockNode.interpret(FCallNoArgBlockNode.java:32)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.FCallManyArgsNode.interpret(FCallManyArgsNode.java:60)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:165)
	at org.jruby.RubyClass.finvoke(RubyClass.java:573)
	at org.jruby.RubyBasicObject.send(RubyBasicObject.java:2801)
	at org.jruby.RubyKernel.send(RubyKernel.java:2117)
	at org.jruby.RubyKernel$s$send.call(RubyKernel$s$send.gen:65535)
	at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:181)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.FCallSpecialArgNode.interpret(FCallSpecialArgNode.java:45)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_BLOCK(ASTInterpreter.java:111)
	at org.jruby.runtime.InterpretedBlock.evalBlockBody(InterpretedBlock.java:374)
	at org.jruby.runtime.InterpretedBlock.yield(InterpretedBlock.java:295)
	at org.jruby.runtime.InterpretedBlock.yieldSpecific(InterpretedBlock.java:229)
	at org.jruby.runtime.Block.yieldSpecific(Block.java:99)
	at org.jruby.ast.ZYieldNode.interpret(ZYieldNode.java:25)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.RescueNode.executeBody(RescueNode.java:216)
	at org.jruby.ast.RescueNode.interpretWithJavaExceptions(RescueNode.java:120)
	at org.jruby.ast.RescueNode.interpret(RescueNode.java:110)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:165)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:272)
	at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:80)
	at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:89)
	at org.jruby.ast.FCallSpecialArgBlockNode.interpret(FCallSpecialArgBlockNode.java:42)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.RescueNode.executeBody(RescueNode.java:216)
	at org.jruby.ast.RescueNode.interpretWithJavaExceptions(RescueNode.java:120)
	at org.jruby.ast.RescueNode.interpret(RescueNode.java:110)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.CallSpecialArgNode.interpret(CallSpecialArgNode.java:73)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:69)
	at org.jruby.ast.FCallSpecialArgNode.interpret(FCallSpecialArgNode.java:45)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.CallSpecialArgNode.interpret(CallSpecialArgNode.java:73)
	at org.jruby.ast.LocalAsgnNode.interpret(LocalAsgnNode.java:123)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.FCallManyArgsNode.interpret(FCallManyArgsNode.java:60)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.ast.RootNode.interpret(RootNode.java:129)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:119)
	at org.jruby.Ruby.runInterpreter(Ruby.java:724)
	at org.jruby.Ruby.loadFile(Ruby.java:2489)
	at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66)
	at org.jruby.runtime.load.LoadService.load(LoadService.java:270)
	at org.jruby.RubyKernel.loadCommon(RubyKernel.java:1105)
	at org.jruby.RubyKernel.load(RubyKernel.java:1087)
	at org.jruby.RubyKernel$s$0$1$load.call(RubyKernel$s$0$1$load.gen:65535)
	at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:211)
	at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:207)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169)
	at usr.hdp.$2_dot_4_dot_2_dot_0_minus_258.hbase.bin.hirb.__file__(/usr/hdp/2.4.2.0-258/hbase/bin/hirb.rb:177)
	at usr.hdp.$2_dot_4_dot_2_dot_0_minus_258.hbase.bin.hirb.load(/usr/hdp/2.4.2.0-258/hbase/bin/hirb.rb)
	at org.jruby.Ruby.runScript(Ruby.java:697)
	at org.jruby.Ruby.runScript(Ruby.java:690)
	at org.jruby.Ruby.runNormally(Ruby.java:597)
	at org.jruby.Ruby.runFromMain(Ruby.java:446)
	at org.jruby.Main.doRunFromMain(Main.java:369)
	at org.jruby.Main.internalRun(Main.java:258)
	at org.jruby.Main.run(Main.java:224)
	at org.jruby.Main.run(Main.java:208)
	at org.jruby.Main.main(Main.java:188)

ERROR: Failed 1 action: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row row01: 1 time, 

Here is some help for this command:
Put a cell 'value' at specified table/row/column and optionally
timestamp coordinates.  To put a cell value into table 'ns1:t1' or 't1'
at row 'r1' under column 'c1' marked with the time 'ts1', do:

  hbase> put 'ns1:t1', 'r1', 'c1', 'value'
  hbase> put 't1', 'r1', 'c1', 'value'
  hbase> put 't1', 'r1', 'c1', 'value', ts1
  hbase> put 't1', 'r1', 'c1', 'value', {ATTRIBUTES=>{'mykey'=>'myvalue'}}
  hbase> put 't1', 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}
  hbase> put 't1', 'r1', 'c1', 'value', ts1, {VISIBILITY=>'PRIVATE|SECRET'}

The same commands also can be run on a table reference. Suppose you had a reference
t to table 't1', the corresponding command would be:

  hbase> t.put 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}
ROW  COLUMN+CELL

ERROR: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row 

Here is some help for this command:
Scan a table; pass table name and optionally a dictionary of scanner
specifications.  Scanner specifications may include one or more of:
TIMERANGE, FILTER, LIMIT, STARTROW, STOPROW, ROWPREFIXFILTER, TIMESTAMP,
MAXLENGTH or COLUMNS, CACHE or RAW, VERSIONS

If no columns are specified, all columns will be scanned.
To scan all members of a column family, leave the qualifier empty as in
'col_family:'.

The filter can be specified in two ways:
1. Using a filterString - more information on this is available in the
Filter Language document attached to the HBASE-4176 JIRA
2. Using the entire package name of the filter.

Some examples:

  hbase> scan 'hbase:meta'
  hbase> scan 'hbase:meta', {COLUMNS => 'info:regioninfo'}
  hbase> scan 'ns1:t1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => 'c1', TIMERANGE => [1303668804, 1303668904]}
  hbase> scan 't1', {REVERSED => true}
  hbase> scan 't1', {ROWPREFIXFILTER => 'row2', FILTER => "
    (QualifierFilter (>=, 'binary:xyz')) AND (TimestampsFilter ( 123, 456))"}
  hbase> scan 't1', {FILTER =>
    org.apache.hadoop.hbase.filter.ColumnPaginationFilter.new(1, 0)}
  hbase> scan 't1', {CONSISTENCY => 'TIMELINE'}
For setting the Operation Attributes 
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], ATTRIBUTES => {'mykey' => 'myvalue'}}
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], AUTHORIZATIONS => ['PRIVATE','SECRET']}
For experts, there is an additional option -- CACHE_BLOCKS -- which
switches block caching for the scanner on (true) or off (false).  By
default it is enabled.  Examples:

  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], CACHE_BLOCKS => false}

Also for experts, there is an advanced option -- RAW -- which instructs the
scanner to return all cells (including delete markers and uncollected deleted
cells). This option cannot be combined with requesting specific COLUMNS.
Disabled by default.  Example:

  hbase> scan 't1', {RAW => true, VERSIONS => 10}

Besides the default 'toStringBinary' format, 'scan' supports custom formatting
by column.  A user can define a FORMATTER by adding it to the column name in
the scan specification.  The FORMATTER can be stipulated: 

 1. either as a org.apache.hadoop.hbase.util.Bytes method name (e.g, toInt, toString)
 2. or as a custom class followed by method name: e.g. 'c(MyFormatterClass).format'.

Example formatting cf:qualifier1 and cf:qualifier2 both as Integers: 
  hbase> scan 't1', {COLUMNS => ['cf:qualifier1:toInt',
    'cf:qualifier2:c(org.apache.hadoop.hbase.util.Bytes).toInt'] } 

Note that you can specify a FORMATTER by column only (cf:qualifier).  You cannot
specify a FORMATTER for all columns of a column family.

Scan can also be used directly from a table, by first getting a reference to a
table, like such:

  hbase> t = get_table 't'
  hbase> t.scan

Note in the above situation, you can still provide all the filtering, columns,
options, etc as described above.
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
HBase Shell; enter 'help<RETURN>' for list of supported commands.
Type "exit<RETURN>" to leave the HBase Shell
Version 1.1.2.2.4.2.0-258, rUnknown, Mon Apr 25 06:07:29 UTC 2016

scan 'ambarismoketest'
ROW  COLUMN+CELL

ERROR: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row 

Here is some help for this command:
Scan a table; pass table name and optionally a dictionary of scanner
specifications.  Scanner specifications may include one or more of:
TIMERANGE, FILTER, LIMIT, STARTROW, STOPROW, ROWPREFIXFILTER, TIMESTAMP,
MAXLENGTH or COLUMNS, CACHE or RAW, VERSIONS

If no columns are specified, all columns will be scanned.
To scan all members of a column family, leave the qualifier empty as in
'col_family:'.

The filter can be specified in two ways:
1. Using a filterString - more information on this is available in the
Filter Language document attached to the HBASE-4176 JIRA
2. Using the entire package name of the filter.

Some examples:

  hbase> scan 'hbase:meta'
  hbase> scan 'hbase:meta', {COLUMNS => 'info:regioninfo'}
  hbase> scan 'ns1:t1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => 'c1', TIMERANGE => [1303668804, 1303668904]}
  hbase> scan 't1', {REVERSED => true}
  hbase> scan 't1', {ROWPREFIXFILTER => 'row2', FILTER => "
    (QualifierFilter (>=, 'binary:xyz')) AND (TimestampsFilter ( 123, 456))"}
  hbase> scan 't1', {FILTER =>
    org.apache.hadoop.hbase.filter.ColumnPaginationFilter.new(1, 0)}
  hbase> scan 't1', {CONSISTENCY => 'TIMELINE'}
For setting the Operation Attributes 
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], ATTRIBUTES => {'mykey' => 'myvalue'}}
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], AUTHORIZATIONS => ['PRIVATE','SECRET']}
For experts, there is an additional option -- CACHE_BLOCKS -- which
switches block caching for the scanner on (true) or off (false).  By
default it is enabled.  Examples:

  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], CACHE_BLOCKS => false}

Also for experts, there is an advanced option -- RAW -- which instructs the
scanner to return all cells (including delete markers and uncollected deleted
cells). This option cannot be combined with requesting specific COLUMNS.
Disabled by default.  Example:

  hbase> scan 't1', {RAW => true, VERSIONS => 10}

Besides the default 'toStringBinary' format, 'scan' supports custom formatting
by column.  A user can define a FORMATTER by adding it to the column name in
the scan specification.  The FORMATTER can be stipulated: 

 1. either as a org.apache.hadoop.hbase.util.Bytes method name (e.g, toInt, toString)
 2. or as a custom class followed by method name: e.g. 'c(MyFormatterClass).format'.

Example formatting cf:qualifier1 and cf:qualifier2 both as Integers: 
  hbase> scan 't1', {COLUMNS => ['cf:qualifier1:toInt',
    'cf:qualifier2:c(org.apache.hadoop.hbase.util.Bytes).toInt'] } 

Note that you can specify a FORMATTER by column only (cf:qualifier).  You cannot
specify a FORMATTER for all columns of a column family.

Scan can also be used directly from a table, by first getting a reference to a
table, like such:

  hbase> t = get_table 't'
  hbase> t.scan

Note in the above situation, you can still provide all the filtering, columns,
options, etc as described above.
Looking for id1fac351e_date341016
2016-09-10 12:35:12,865 - Retrying after 5 seconds. Reason: Execution of ' /usr/hdp/current/hbase-client/bin/hbase --config /usr/hdp/current/hbase-client/conf shell /var/lib/ambari-agent/tmp/hbase-smoke.sh && /var/lib/ambari-agent/tmp/hbaseSmokeVerify.sh /usr/hdp/current/hbase-client/conf id1fac351e_date341016 /usr/hdp/current/hbase-client/bin/hbase' returned 1. SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]

ERROR: Table ambarismoketest does not exist.

Here is some help for this command:
Start disable of named table:
  hbase> disable 't1'
  hbase> disable 'ns1:t1'
ERROR: Table ambarismoketest does not exist.

Here is some help for this command:
Drop the named table. Table must first be disabled:
  hbase> drop 't1'
  hbase> drop 'ns1:t1'
0 row(s) in 28.2970 seconds

2016-09-10 12:34:58,865 ERROR [main] client.AsyncProcess: Failed to get region location 
org.apache.hadoop.hbase.client.NoServerForRegionException: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row row01
	at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1299)
	at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1162)
	at org.apache.hadoop.hbase.client.AsyncProcess.submit(AsyncProcess.java:370)
	at org.apache.hadoop.hbase.client.AsyncProcess.submit(AsyncProcess.java:321)
	at org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:206)
	at org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:183)
	at org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1449)
	at org.apache.hadoop.hbase.client.HTable.put(HTable.java:1040)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:497)
	at org.jruby.javasupport.JavaMethod.invokeDirectWithExceptionHandling(JavaMethod.java:450)
	at org.jruby.javasupport.JavaMethod.invokeDirect(JavaMethod.java:311)
	at org.jruby.java.invokers.InstanceMethodInvoker.call(InstanceMethodInvoker.java:59)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169)
	at org.jruby.ast.CallOneArgNode.interpret(CallOneArgNode.java:57)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.CallManyArgsNode.interpret(CallManyArgsNode.java:59)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_BLOCK(ASTInterpreter.java:111)
	at org.jruby.runtime.InterpretedBlock.evalBlockBody(InterpretedBlock.java:374)
	at org.jruby.runtime.InterpretedBlock.yield(InterpretedBlock.java:295)
	at org.jruby.runtime.InterpretedBlock.yieldSpecific(InterpretedBlock.java:229)
	at org.jruby.runtime.Block.yieldSpecific(Block.java:99)
	at org.jruby.ast.ZYieldNode.interpret(ZYieldNode.java:25)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:169)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:191)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:302)
	at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:144)
	at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:153)
	at org.jruby.ast.FCallNoArgBlockNode.interpret(FCallNoArgBlockNode.java:32)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.FCallManyArgsNode.interpret(FCallManyArgsNode.java:60)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:165)
	at org.jruby.RubyClass.finvoke(RubyClass.java:573)
	at org.jruby.RubyBasicObject.send(RubyBasicObject.java:2801)
	at org.jruby.RubyKernel.send(RubyKernel.java:2117)
	at org.jruby.RubyKernel$s$send.call(RubyKernel$s$send.gen:65535)
	at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:181)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.FCallSpecialArgNode.interpret(FCallSpecialArgNode.java:45)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_BLOCK(ASTInterpreter.java:111)
	at org.jruby.runtime.InterpretedBlock.evalBlockBody(InterpretedBlock.java:374)
	at org.jruby.runtime.InterpretedBlock.yield(InterpretedBlock.java:295)
	at org.jruby.runtime.InterpretedBlock.yieldSpecific(InterpretedBlock.java:229)
	at org.jruby.runtime.Block.yieldSpecific(Block.java:99)
	at org.jruby.ast.ZYieldNode.interpret(ZYieldNode.java:25)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.RescueNode.executeBody(RescueNode.java:216)
	at org.jruby.ast.RescueNode.interpretWithJavaExceptions(RescueNode.java:120)
	at org.jruby.ast.RescueNode.interpret(RescueNode.java:110)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:165)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:272)
	at org.jruby.runtime.callsite.CachingCallSite.callBlock(CachingCallSite.java:80)
	at org.jruby.runtime.callsite.CachingCallSite.callIter(CachingCallSite.java:89)
	at org.jruby.ast.FCallSpecialArgBlockNode.interpret(FCallSpecialArgBlockNode.java:42)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.RescueNode.executeBody(RescueNode.java:216)
	at org.jruby.ast.RescueNode.interpretWithJavaExceptions(RescueNode.java:120)
	at org.jruby.ast.RescueNode.interpret(RescueNode.java:110)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.CallSpecialArgNode.interpret(CallSpecialArgNode.java:73)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:69)
	at org.jruby.ast.FCallSpecialArgNode.interpret(FCallSpecialArgNode.java:45)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.CallSpecialArgNode.interpret(CallSpecialArgNode.java:73)
	at org.jruby.ast.LocalAsgnNode.interpret(LocalAsgnNode.java:123)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_METHOD(ASTInterpreter.java:74)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:120)
	at org.jruby.internal.runtime.methods.InterpretedMethod.call(InterpretedMethod.java:134)
	at org.jruby.internal.runtime.methods.DefaultMethod.call(DefaultMethod.java:174)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:282)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:71)
	at org.jruby.ast.FCallManyArgsNode.interpret(FCallManyArgsNode.java:60)
	at org.jruby.ast.NewlineNode.interpret(NewlineNode.java:104)
	at org.jruby.ast.BlockNode.interpret(BlockNode.java:71)
	at org.jruby.ast.RootNode.interpret(RootNode.java:129)
	at org.jruby.evaluator.ASTInterpreter.INTERPRET_ROOT(ASTInterpreter.java:119)
	at org.jruby.Ruby.runInterpreter(Ruby.java:724)
	at org.jruby.Ruby.loadFile(Ruby.java:2489)
	at org.jruby.runtime.load.ExternalScript.load(ExternalScript.java:66)
	at org.jruby.runtime.load.LoadService.load(LoadService.java:270)
	at org.jruby.RubyKernel.loadCommon(RubyKernel.java:1105)
	at org.jruby.RubyKernel.load(RubyKernel.java:1087)
	at org.jruby.RubyKernel$s$0$1$load.call(RubyKernel$s$0$1$load.gen:65535)
	at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:211)
	at org.jruby.internal.runtime.methods.DynamicMethod.call(DynamicMethod.java:207)
	at org.jruby.runtime.callsite.CachingCallSite.cacheAndCall(CachingCallSite.java:312)
	at org.jruby.runtime.callsite.CachingCallSite.call(CachingCallSite.java:169)
	at usr.hdp.$2_dot_4_dot_2_dot_0_minus_258.hbase.bin.hirb.__file__(/usr/hdp/2.4.2.0-258/hbase/bin/hirb.rb:177)
	at usr.hdp.$2_dot_4_dot_2_dot_0_minus_258.hbase.bin.hirb.load(/usr/hdp/2.4.2.0-258/hbase/bin/hirb.rb)
	at org.jruby.Ruby.runScript(Ruby.java:697)
	at org.jruby.Ruby.runScript(Ruby.java:690)
	at org.jruby.Ruby.runNormally(Ruby.java:597)
	at org.jruby.Ruby.runFromMain(Ruby.java:446)
	at org.jruby.Main.doRunFromMain(Main.java:369)
	at org.jruby.Main.internalRun(Main.java:258)
	at org.jruby.Main.run(Main.java:224)
	at org.jruby.Main.run(Main.java:208)
	at org.jruby.Main.main(Main.java:188)

ERROR: Failed 1 action: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row row01: 1 time, 

Here is some help for this command:
Put a cell 'value' at specified table/row/column and optionally
timestamp coordinates.  To put a cell value into table 'ns1:t1' or 't1'
at row 'r1' under column 'c1' marked with the time 'ts1', do:

  hbase> put 'ns1:t1', 'r1', 'c1', 'value'
  hbase> put 't1', 'r1', 'c1', 'value'
  hbase> put 't1', 'r1', 'c1', 'value', ts1
  hbase> put 't1', 'r1', 'c1', 'value', {ATTRIBUTES=>{'mykey'=>'myvalue'}}
  hbase> put 't1', 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}
  hbase> put 't1', 'r1', 'c1', 'value', ts1, {VISIBILITY=>'PRIVATE|SECRET'}

The same commands also can be run on a table reference. Suppose you had a reference
t to table 't1', the corresponding command would be:

  hbase> t.put 'r1', 'c1', 'value', ts1, {ATTRIBUTES=>{'mykey'=>'myvalue'}}
ROW  COLUMN+CELL

ERROR: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row 

Here is some help for this command:
Scan a table; pass table name and optionally a dictionary of scanner
specifications.  Scanner specifications may include one or more of:
TIMERANGE, FILTER, LIMIT, STARTROW, STOPROW, ROWPREFIXFILTER, TIMESTAMP,
MAXLENGTH or COLUMNS, CACHE or RAW, VERSIONS

If no columns are specified, all columns will be scanned.
To scan all members of a column family, leave the qualifier empty as in
'col_family:'.

The filter can be specified in two ways:
1. Using a filterString - more information on this is available in the
Filter Language document attached to the HBASE-4176 JIRA
2. Using the entire package name of the filter.

Some examples:

  hbase> scan 'hbase:meta'
  hbase> scan 'hbase:meta', {COLUMNS => 'info:regioninfo'}
  hbase> scan 'ns1:t1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => 'c1', TIMERANGE => [1303668804, 1303668904]}
  hbase> scan 't1', {REVERSED => true}
  hbase> scan 't1', {ROWPREFIXFILTER => 'row2', FILTER => "
    (QualifierFilter (>=, 'binary:xyz')) AND (TimestampsFilter ( 123, 456))"}
  hbase> scan 't1', {FILTER =>
    org.apache.hadoop.hbase.filter.ColumnPaginationFilter.new(1, 0)}
  hbase> scan 't1', {CONSISTENCY => 'TIMELINE'}
For setting the Operation Attributes 
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], ATTRIBUTES => {'mykey' => 'myvalue'}}
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], AUTHORIZATIONS => ['PRIVATE','SECRET']}
For experts, there is an additional option -- CACHE_BLOCKS -- which
switches block caching for the scanner on (true) or off (false).  By
default it is enabled.  Examples:

  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], CACHE_BLOCKS => false}

Also for experts, there is an advanced option -- RAW -- which instructs the
scanner to return all cells (including delete markers and uncollected deleted
cells). This option cannot be combined with requesting specific COLUMNS.
Disabled by default.  Example:

  hbase> scan 't1', {RAW => true, VERSIONS => 10}

Besides the default 'toStringBinary' format, 'scan' supports custom formatting
by column.  A user can define a FORMATTER by adding it to the column name in
the scan specification.  The FORMATTER can be stipulated: 

 1. either as a org.apache.hadoop.hbase.util.Bytes method name (e.g, toInt, toString)
 2. or as a custom class followed by method name: e.g. 'c(MyFormatterClass).format'.

Example formatting cf:qualifier1 and cf:qualifier2 both as Integers: 
  hbase> scan 't1', {COLUMNS => ['cf:qualifier1:toInt',
    'cf:qualifier2:c(org.apache.hadoop.hbase.util.Bytes).toInt'] } 

Note that you can specify a FORMATTER by column only (cf:qualifier).  You cannot
specify a FORMATTER for all columns of a column family.

Scan can also be used directly from a table, by first getting a reference to a
table, like such:

  hbase> t = get_table 't'
  hbase> t.scan

Note in the above situation, you can still provide all the filtering, columns,
options, etc as described above.
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
HBase Shell; enter 'help<RETURN>' for list of supported commands.
Type "exit<RETURN>" to leave the HBase Shell
Version 1.1.2.2.4.2.0-258, rUnknown, Mon Apr 25 06:07:29 UTC 2016

scan 'ambarismoketest'
ROW  COLUMN+CELL

ERROR: No server address listed in hbase:meta for region ambarismoketest,,1473510866427.7b288aad9960fe5563740fc3a901985d. containing row 

Here is some help for this command:
Scan a table; pass table name and optionally a dictionary of scanner
specifications.  Scanner specifications may include one or more of:
TIMERANGE, FILTER, LIMIT, STARTROW, STOPROW, ROWPREFIXFILTER, TIMESTAMP,
MAXLENGTH or COLUMNS, CACHE or RAW, VERSIONS

If no columns are specified, all columns will be scanned.
To scan all members of a column family, leave the qualifier empty as in
'col_family:'.

The filter can be specified in two ways:
1. Using a filterString - more information on this is available in the
Filter Language document attached to the HBASE-4176 JIRA
2. Using the entire package name of the filter.

Some examples:

  hbase> scan 'hbase:meta'
  hbase> scan 'hbase:meta', {COLUMNS => 'info:regioninfo'}
  hbase> scan 'ns1:t1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], LIMIT => 10, STARTROW => 'xyz'}
  hbase> scan 't1', {COLUMNS => 'c1', TIMERANGE => [1303668804, 1303668904]}
  hbase> scan 't1', {REVERSED => true}
  hbase> scan 't1', {ROWPREFIXFILTER => 'row2', FILTER => "
    (QualifierFilter (>=, 'binary:xyz')) AND (TimestampsFilter ( 123, 456))"}
  hbase> scan 't1', {FILTER =>
    org.apache.hadoop.hbase.filter.ColumnPaginationFilter.new(1, 0)}
  hbase> scan 't1', {CONSISTENCY => 'TIMELINE'}
For setting the Operation Attributes 
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], ATTRIBUTES => {'mykey' => 'myvalue'}}
  hbase> scan 't1', { COLUMNS => ['c1', 'c2'], AUTHORIZATIONS => ['PRIVATE','SECRET']}
For experts, there is an additional option -- CACHE_BLOCKS -- which
switches block caching for the scanner on (true) or off (false).  By
default it is enabled.  Examples:

  hbase> scan 't1', {COLUMNS => ['c1', 'c2'], CACHE_BLOCKS => false}

Also for experts, there is an advanced option -- RAW -- which instructs the
scanner to return all cells (including delete markers and uncollected deleted
cells). This option cannot be combined with requesting specific COLUMNS.
Disabled by default.  Example:

  hbase> scan 't1', {RAW => true, VERSIONS => 10}

Besides the default 'toStringBinary' format, 'scan' supports custom formatting
by column.  A user can define a FORMATTER by adding it to the column name in
the scan specification.  The FORMATTER can be stipulated: 

 1. either as a org.apache.hadoop.hbase.util.Bytes method name (e.g, toInt, toString)
 2. or as a custom class followed by method name: e.g. 'c(MyFormatterClass).format'.

Example formatting cf:qualifier1 and cf:qualifier2 both as Integers: 
  hbase> scan 't1', {COLUMNS => ['cf:qualifier1:toInt',
    'cf:qualifier2:c(org.apache.hadoop.hbase.util.Bytes).toInt'] } 

Note that you can specify a FORMATTER by column only (cf:qualifier).  You cannot
specify a FORMATTER for all columns of a column family.

Scan can also be used directly from a table, by first getting a reference to a
table, like such:

  hbase> t = get_table 't'
  hbase> t.scan

Note in the above situation, you can still provide all the filtering, columns,
options, etc as described above.
Looking for id1fac351e_date341016
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
0 row(s) in 1.5310 seconds


Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

New Contributor

This worked for me. (Do this at your own risk)

Restart Zookeeper and HBASE

Clean Wal and zookeeper hbase-unsecure

$ sudo -u hdfs hdfs dfs -rm -r /apps/hbase/data/WALs/

$ zookeeper-client rmr /hbase-unsecure/rs

Restart HBase

Re: org.apache.hadoop.hbase.PleaseHoldException: Master is initializing error

Explorer

Hi @linuslukia or anyone else... can you PLEASE tell us what is the solution for this problem?

I tried and last advice with deleteing the WALS and rm from /hbase-unsecure/rs and restarted zookeeper and hbase and it didn't work.

 

Don't have an account?
Coming from Hortonworks? Activate your account here