Member since
03-15-2017
5
Posts
0
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
860 | 05-15-2017 07:19 AM |
05-11-2018
08:45 AM
Directory was created by full backup couple rows above.
Remote debugging hbase-release from hortonworks github repo gives that org.apache.hadoop.fs.FileSystem.checkPath comparing authority points from uri. Compare takes uri authority from arguments and Hbase Master conf, which gives
"dedrain-workstation:8020" and 172.25.10.22:8020. Obviously comparing gives false.)
... View more
05-10-2018
10:08 AM
Hi all! I'm trying to create hbase backups to another remote hdfs cluster. Full backup done successfully, but incremental fails: [hbase@dedrain-workstation dedrain]$ hbase backup create full hdfs://172.25.10.22:8020/backup_fr/10052018
2018-05-10 12:49:37,850 INFO [main] util.BackupClientUtil: Backup root dir hdfs://172.25.10.22:8020/backup_fr/10052018 does not exist. Will be created.
Backup session backup_1525945778016 finished. Status: SUCCESS
[hbase@dedrain-workstation dedrain]$ hbase backup create incremental hdfs://172.25.10.22:8020/backup_fr/10052018
2018-05-10 12:51:00,116 INFO [main] util.BackupClientUtil: Using existing backup root dir: hdfs://172.25.10.22:8020/backup_fr/10052018
Backup session finished. Status: FAILURE
2018-05-10 12:52:48,772 ERROR [main] util.AbstractHBaseTool: Error running command-line tool
org.apache.hadoop.ipc.RemoteException(java.lang.IllegalArgumentException): Wrong FS: hdfs://172.25.10.22:8020/backup_fr/10052018/.tmp/backup_1525945860331, expected: hdfs://dedrain-workstation:8020
at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:666)
at org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:214)
at org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:816)
at org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:812)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.delete(DistributedFileSystem.java:823)
at org.apache.hadoop.hbase.backup.master.IncrementalTableBackupProcedure.deleteBulkLoadDirectory(IncrementalTableBackupProcedure.java:487)
at org.apache.hadoop.hbase.backup.master.IncrementalTableBackupProcedure.incrementalCopyHFiles(IncrementalTableBackupProcedure.java:478)
at org.apache.hadoop.hbase.backup.master.IncrementalTableBackupProcedure.executeFromState(IncrementalTableBackupProcedure.java:287)
at org.apache.hadoop.hbase.backup.master.IncrementalTableBackupProcedure.executeFromState(IncrementalTableBackupProcedure.java:71)
at org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:107)
at org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:500)
at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1086)
at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:888)
at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:841)
at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$200(ProcedureExecutor.java:77)
at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$1.run(ProcedureExecutor.java:443)
Host dedrain-workstation and 172.25.10.22 is different hosts and clusters. What am i doing wrong?
... View more
- Tags:
- HBase
Labels:
05-15-2017
07:19 AM
The problem was in Hive transaction timeouts. In Flume agent conf: hive_sink.heartBeatInterval = 10000 In Hive conf: hive.txn.timeout = 300 Hive close transactions before any heartbeat was received.
... View more
04-28-2017
12:31 PM
Context: Flume takes files from local dir and writes them to transactional hive table. Problem: Sometime flume agent fails with exception. For example: pr 2017 15:48:02,126 ERROR [hive-f_sink-call-runner-0] (org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.markDead:755) - Fatal error on TxnIds=[841192...841291] on endPoint = {metaStoreUri='thrift://node02.hdp:9083', database='default', table='f', partitionVals=[] }; cause Unable to abort invalid transaction id : 841253: No such transaction txnid:841253
org.apache.hive.hcatalog.streaming.TransactionError: Unable to abort invalid transaction id : 841253: No such transaction txnid:841253
at org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.abortImpl(HiveEndPoint.java:934)
at org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.abort(HiveEndPoint.java:893)
at org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.markDead(HiveEndPoint.java:752)
at org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.commit(HiveEndPoint.java:852)
at org.apache.flume.sink.hive.HiveWriter$6.call(HiveWriter.java:345)
at org.apache.flume.sink.hive.HiveWriter$6.call(HiveWriter.java:342)
at org.apache.flume.sink.hive.HiveWriter$11.call(HiveWriter.java:429)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Caused by: NoSuchTxnException(message:No such transaction txnid:841253)
at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$abort_txn_result$abort_txn_resultStandardScheme.read(ThriftHiveMetastore.java)
at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$abort_txn_result$abort_txn_resultStandardScheme.read(ThriftHiveMetastore.java)
at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$abort_txn_result.read(ThriftHiveMetastore.java)
at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:86)
at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.recv_abort_txn(ThriftHiveMetastore.java:3898)
at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Client.abort_txn(ThriftHiveMetastore.java:3885)
at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.rollbackTxn(HiveMetaStoreClient.java:1885)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:155)
at com.sun.proxy.$Proxy8.rollbackTxn(Unknown Source)
at org.apache.hive.hcatalog.streaming.HiveEndPoint$TransactionBatchImpl.abortImpl(HiveEndPoint.java:922)
... 10 more
Can anyone explain why HCatalog trying to abort transaction?
Setup: HDP-2.5 Flume conf: f.sources = f_dir
f.channels = f_channel
f.sinks = f_sink
f.sources.f_dir.type = spooldir
f.sources.f_dir.spoolDir = /var/data/f_data/History
f.sources.f_dir.deletePolicy = immediate
f.sources.f_dir.channels = f_channel
f.sources.f_dir.deserializer.maxLineLength = 150000
f.channels.f_channel.type = file
f.channels.f_channel.capacity = 20000
f.channels.f_channel.transactionCapacity = 15000
f.sinks.f_sink.type = hive
f.sinks.f_sink.hive.metastore = thrift://node02.hdp:9083
f.sinks.f_sink.hive.database = default
f.sinks.f_sink.hive.table = f
f.sinks.f_sink.channel = f_channel
f.sinks.f_sink.serializer = JSON
f.sinks.f_sink.HDFSEventSink.batchSize = 2000
f.sinks.f_sink.callTimeout = 80000
f.sinks.f_sink.heartBeatInterval = 10000
Table - it's just some json fields, clustered by monotonically increasing timestamp into 30 buckets. Timestamp definitely not repeating. Hive table props: inputFormat:org.apache.hadoop.hive.ql.io.orc.OrcInputFormat,
outputFormat:org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat,
compressed:false,
numBuckets:30,
serdeInfo:SerDeInfo(name:null,
serializationLib:org.apache.hadoop.hive.ql.io.orc.OrcSerde,
parameters:{serialization.format=1}),
bucketCols:[firsttimestamp],
sortCols:[],
parameters:{},
skewedInfo:SkewedInfo(skewedColNames:[],
skewedColValues:[],
skewedColValueLocationMaps:{}),
storedAsSubDirectories:false
partitionKeys:[],
parameters:{totalSize=5498078799,
last_modified_time=1490710346,
numRows=1027277,
rawDataSize=28307100383,
compactorthreshold.hive.compactor.delta.num.threshold=1,
numFiles=270,
transient_lastDdlTime=1490710346,
last_modified_by=root,
transactional=true},
viewOriginalText:null,
viewExpandedText:null,
tableType:MANAGED_TABLE Sorry for my english 😃
... View more
03-15-2017
08:53 AM
Hi all! We've come to strange situation with transactional tables. For several dropped tables we've see many compactions with "attempted" state. There is no files, no tables, but compactions still in compaction queue. As i understand this compactions never will be completed. Is there any way to remove this compactions from queue? Will be pleased any help. P.S. Sorry for my english.
... View more