Member since
04-25-2016
579
Posts
609
Kudos Received
111
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2946 | 02-12-2020 03:17 PM | |
| 2147 | 08-10-2017 09:42 AM | |
| 12510 | 07-28-2017 03:57 AM | |
| 3455 | 07-19-2017 02:43 AM | |
| 2544 | 07-13-2017 11:42 AM |
12-13-2016
02:40 PM
@Joshua Adeleke did you see any exception/error in application master log
... View more
12-13-2016
12:28 PM
2 Kudos
@akshay singh The error indicates that your replication is falling behind,this exception doesn't really imply that a broker is down, it can likewise mean that an excessive number of copies for that topic partition have fallen behind the leader with some reason.
... View more
12-10-2016
07:41 PM
2 Kudos
looks you are having this problem on kerberized cluster and user dont have access on kafka topic, please try to set kafka topic acl using following link. https://community.hortonworks.com/articles/58370/kafka-topic-creation-and-acl-configuration-for-atl.html
... View more
12-10-2016
08:45 AM
4 Kudos
@soumyabrata kole this is often a problem with multiline json document where during read spark read it as corrupt record http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrameReader.json if you create a json file with a json document single line it will able to get the schema right. [spark@rkk1 ~]$ cat sample.json{"employees":[{"firstName":"John", "lastName":"Doe"},{"firstName":"Anna", "lastName":"Smith"},{"firstName":"Peter", "lastName":"Jones"}]}
scala> val dfs = spark.sqlContext.read.json("file:///home/spark/sample.json")dfs: org.apache.spark.sql.DataFrame = [employees: array<struct<firstName:string,lastName:string>>]scala> dfs.printSchemaroot |-- employees: array (nullable = true) | |-- element: struct (containsNull = true) | | |-- firstName: string (nullable = true) | | |-- lastName: string (nullable = true)
... View more
12-10-2016
04:52 AM
3 Kudos
@Akhil Bansal could you please try this https://community.hortonworks.com/repos/62212/sparkorcwriter.html
... View more
12-10-2016
04:49 AM
3 Kudos
@Joshua Adeleke for Causedby: org.apache.tez.dag.api.SessionNotRunning:TezSession has already shutdown.No cluster diagnostics found.at org.apache.tez.client.TezClient.waitTillReady(TezClient.java:730) at org.apache.hadoop.hive.ql.exec.tez.TezSessionState.open(TezSessionState.java:217) at org.apache.hadoop.hive.ql.exec.tez.TezSessionState.open(TezSessionState.java:117) it seems that your application master/cluster is busy or did not get enough resources to process DAG over RPC request thats why was not able to copy the dag from client.if there are EOFExceptions in Appliction master logs then do check your network, for a workaround you can increase tez.session.am.dag.submit.timeout.secs which is currently 300 seconds to 900 seconds.
... View more
12-10-2016
04:38 AM
2 Kudos
@Lawrence Lau could you please try building hive after adding hortonworks repo in you pom <repositories>
<repository> <id>public</id> <url>http://nexus-private.hortonworks.com/nexus/content/groups/public
</url> </repository> </repositories>
... View more
12-09-2016
12:19 PM
3 Kudos
you can set following property in storm.yaml, if you are using ambari then you need to create custom property as follows Key: nimbus.impersonation.acl
Value: {'stormdev':{'hosts':['host1','host2'],'groups':['group1','group2']}}
restart cluster after setting these property
... View more
12-09-2016
12:01 PM
@akshay singh check if kafka process is running with these parameters -Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.port=9999 -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.authenticate=false see a similar issue when jmx is not bound to port in HDP distribution so you need to set JMX_PORT env variable and restart kafka process
... View more
12-09-2016
11:54 AM
3 Kudos
what exception you are getting here?
... View more