Member since
09-28-2016
59
Posts
15
Kudos Received
0
Solutions
11-04-2016
08:31 PM
I am trying to remove duplicate json records from json array using Jolt transformation . Here is an example i tried :
Input :
[
{
"id": 1,
"name": "jeorge",
"age": 25
},
{
"id": 2,
"name": "manhan",
"age": 25
},
{
"id": 1,
"name": "george",
"age": 225
}
] Jolt script :
[
{
"operation": "shift",
"spec": {
"*": {
"id": "[&1].id"
}
}
}
] Output :
[ {
"id" : 1
}, {
"id" : 2
}, {
"id" : 1
} ] getting only selected records . along with that , i would like to remove duplicates .
Desired Output :
[ {
"id" : 1
}, {
"id" : 2
} ] Please provide the necessary script which will help me . Thanks in advance .
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache NiFi
10-31-2016
04:12 PM
means , it is length of row , not key length . for better understanding , please take a look at below ex , row8 is the row-key value of a row , age=88,name=srini88,no=8 are the values of a row .
so , here what will be the value which compares against the MAX_ROW_LENGTH . is it row8 length or whole row length(which is the combination of row8,88,srini88,8) ? and Is MAX_ROW_LENGTH possible to modify in hbase-site or for specific table ? if yes , please let us know how it is ? ex : row8 column=0:age, timestamp=1475378868472, value=88 row8 column=0:name, timestamp=1475378868438, value=srini8 row8 column=0:no, timestamp=1475378868384, value=8
... View more
10-31-2016
03:33 PM
1 Kudo
Hi , In one my application , i am getting illegalargument exception 38121 is > 32767 . so i found that MAX_ROW_LENGTH is having constant value of 32767 . Here my question is MAX_ROW_LENGTH values is belongs to row key or only key length ? as per this link https://hbase.apache.org/apidocs/org/apache/hadoop/hbase/HConstants.html#MAX_ROW_LENGTH , I unsrtood that it is belong to row length not key length ?? Please correct me if i am wrong .
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache Ranger
10-31-2016
03:29 PM
Thank you . I am refering this link https://hbase.apache.org/apidocs/org/apache/hadoop/hbase/HConstants.html#MAX_ROW_LENGTH what i understood is MAX_ROW_LENGTH is about whole row not only key . Here is my confusion , how can i take this property is only belongs to key ?
... View more
10-31-2016
02:41 PM
Is MAX_ROW_LENGTH belong to to only key or whole row which consists of all columns ??
... View more
10-28-2016
07:28 PM
Hi , We have followed the same method . It is working successfully most of the time . But sometimes we are getting the below error .
2016-10-28 18:43:03,603 ERROR [Timer-Driven Process Thread-70] o.apache.nifi.processors.standard.PutSQL PutSQL[id=df59f4c8-f60c-4eb3-7fda-882f7ece2d2a] PutSQL[id=df59f4c8-f60c-4eb3-7fda-882f7ece2d2a] failed to process session due to java.lang.IllegalArgumentException: Row length 37812 is > 32767: java.lang.IllegalArgumentException: Row length 37812 is > 32767 2016-10-28 18:43:03,611 ERROR [Timer-Driven Process Thread-70] o.apache.nifi.processors.standard.PutSQL java.lang.IllegalArgumentException: Row length 37812 is > 32767 at org.apache.hadoop.hbase.client.Mutation.checkRow(Mutation.java:545) ~[na:na] at org.apache.hadoop.hbase.client.Put.<init>(Put.java:110) ~[na:na] at org.apache.hadoop.hbase.client.Put.<init>(Put.java:68) ~[na:na] at org.apache.hadoop.hbase.client.Put.<init>(Put.java:58) ~[na:na] at org.apache.phoenix.index.IndexMaintainer.buildUpdateMutation(IndexMaintainer.java:779) ~[na:na] at org.apache.phoenix.util.IndexUtil.generateIndexData(IndexUtil.java:263) ~[na:na] at org.apache.phoenix.execute.MutationState$1.next(MutationState.java:221) ~[na:na] at org.apache.phoenix.execute.MutationState$1.next(MutationState.java:204) ~[na:na] at org.apache.phoenix.execute.MutationState.commit(MutationState.java:370) ~[na:na] at org.apache.phoenix.jdbc.PhoenixConnection$3.call(PhoenixConnection.java:459) ~[na:na] at org.apache.phoenix.jdbc.PhoenixConnection$3.call(PhoenixConnection.java:456) ~[na:na] at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) ~[na:na] at org.apache.phoenix.jdbc.PhoenixConnection.commit(PhoenixConnection.java:456) ~[na:na] at org.apache.commons.dbcp.DelegatingConnection.commit(DelegatingConnection.java:334) ~[na:na] at org.apache.commons.dbcp.PoolingDataSource$PoolGuardConnectionWrapper.commit(PoolingDataSource.java:211) ~[na:na] at org.apache.nifi.processors.standard.PutSQL.onTrigger(PutSQL.java:371) ~[na:na] at org.apache.nifi.processor.AbstractProcessor.onTrigger(AbstractProcessor.java:27) ~[nifi-api-1.0.0.2.0.0.0-579.jar:1.0.0.2.0.0.0-579] at org.apache.nifi.controller.StandardProcessorNode.onTrigger(StandardProcessorNode.java:1064) ~[nifi-framework-core-1.0.0.2.0.0.0-579.jar:1.0.0.2.0.0.0-579] at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:136) [nifi-framework-core-1.0.0.2.0.0.0-579.jar:1.0.0.2.0.0.0-579] at org.apache.nifi.controller.tasks.ContinuallyRunProcessorTask.call(ContinuallyRunProcessorTask.java:47) [nifi-framework-core-1.0.0.2.0.0.0-579.jar:1.0.0.2.0.0.0-579] at org.apache.nifi.controller.scheduling.TimerDrivenSchedulingAgent$1.run(TimerDrivenSchedulingAgent.java:132) [nifi-framework-core-1.0.0.2.0.0.0-579.jar:1.0.0.2.0.0.0-579] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [na:1.8.0_91] at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) [na:1.8.0_91] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180) [na:1.8.0_91] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294) [na:1.8.0_91] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_91] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_91] at java.lang.Thread.run(Thread.java:745) [na:1.8.0_91]
FYI - we are using Nifi1.0 and we dont have each rowlength more than 500 bytes .
firsttime when we got the error , we just cleaned the queue and restarted . its working fine later . once again we got th error . restarting i snot the good solution , and we are loosing the data if we do that .
PFA for more information .
... View more
10-25-2016
03:08 AM
1 Kudo
nifi - how to configure a logback.xml in nifi to capture seperate log files for each nifi processor in terms of debug and error mode ?? Thanks .
... View more
Labels:
- Labels:
-
Apache NiFi
10-25-2016
03:04 AM
1 Kudo
How to debug each nifi processor interms of how much RAM, network , how many flowfiles ..etc . Or is it possible to write seperate log files for each nifi processor ? Please let us know the process . thanks
... View more
Labels:
- Labels:
-
Apache NiFi
10-25-2016
02:59 AM
Nifi -- Without writing to disk , is it possible to send flowfile from one peocessor to another ? Ex: i have 3 processors in order- splitjson , evaluatejsonpath , updateattribute processors . I would like to process only these 3 processors in-memory (from the output of splitjson to update attribute o/p). if it is possible only selected processors in-memory . let us know the process ??
... View more
Labels:
- Labels:
-
Apache Falcon
-
Apache NiFi
10-21-2016
05:27 PM
we got same issue , i removed the WALs files and restarted the cluster , still the issue is persistent . what might be the reason ??
... View more