Member since
04-25-2016
579
Posts
609
Kudos Received
111
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1118 | 02-12-2020 03:17 PM | |
883 | 08-10-2017 09:42 AM | |
7140 | 07-28-2017 03:57 AM | |
1321 | 07-19-2017 02:43 AM | |
1021 | 07-13-2017 11:42 AM |
07-10-2017
06:15 AM
@Arnault Droz since your file not too big, you can use HDFS java api to write on HDFS, similar to this http://hadoopinrealworld.com/writing-a-file-to-hdfs-java-program/
... View more
07-10-2017
05:52 AM
@oula.alshiekh@gmail.com alshiekh possibly your retry policy is set too high, could you please check what the value for dfs.client.retry.policy.spec is?
... View more
07-10-2017
05:39 AM
1 Kudo
@Triffids G if you have hive installed on the sandbox you can copy /usr/hdp/2.6.0.3-8/hive/lib/parquet-hadoop-bundle-1.8.1.jar and paste it to /usr/hdp/2.6.0.3-8/hadoop/lib and try running the job again and let me know how it goes.
... View more
07-10-2017
05:27 AM
@Tomomichi Hirano LLAP cache data for several queries running on it so doAs is not supported in LLAP, every query run on LLAP run as hive user instead of the user who submitted the query. to get a fine grain control over accessibility you can user Ranger security policies along with hive.server2.enable.doAs=false
... View more
07-10-2017
04:44 AM
weird, could you please post both the url?
... View more
05-16-2017
12:33 PM
@Rajesh Reddy please set hive.execution.engine=mr in hive-site.xml under /etc/spark/conf on the node where you are launching spark job.
... View more
05-05-2017
01:53 PM
this can happen when client tgt got corrupt or windows client sends an NTLM based ticket not a kerberos based ticket. please refer https://community.hortonworks.com/questions/2580/accessing-hdp-web-ui-from-windows-pc-causes-gsshea.html
... View more
05-02-2017
12:23 PM
@Jatin Kheradiya try this way /usr/hdp/current/zookeeper-server/bin/zkCli.sh -server localhost:2181 -c create "/hello 1" "hvalue"
/usr/hdp/current/zookeeper-server/bin/zkCli.sh -server localhost:2181 -c get "/hello 1"
... View more
04-20-2017
08:36 AM
no need to store the result in RDBMS, these libraries are javascript library, you can pull the result using java jdbc program https://github.com/rajkrrsingh/HiveServer2JDBCSample, set in some variable and plot using these libraries.
... View more
04-20-2017
07:14 AM
@heta desai you can pull the query result from hive using hive-jdbc and plot using d3js (https://d3js.org/) or c3js (http://c3js.org/examples.html)
... View more
01-24-2017
05:41 AM
@Hajime did you try setting debug args for executors spark.driver.extraJavaOptions?
... View more
12-27-2016
05:29 AM
@Ye Jun this is the hortonworks public repo for maven artifects http://nexus-private.hortonworks.com/nexus/content/groups/public/ add this to you pom.xml <repository>
<id>public</id>
<url>http://nexus-private.hortonworks.com/nexus/content/groups/public</url>
</repository> get your dependecy from the repo link and add in your maven dependencies.
... View more
12-26-2016
05:16 PM
@Maria Khan possibly you are hitting https://issues.apache.org/jira/browse/KAFKA-3042
... View more
12-26-2016
05:12 PM
@vamsi valiveti could you please try this? hive> create table x(yy string);
OK
Time taken: 2.155 seconds
hive> show create table x;
OK
CREATE TABLE `x`(
`yy` string)
ROW FORMAT SERDE
'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
STORED AS INPUTFORMAT
'org.apache.hadoop.mapred.TextInputFormat'
OUTPUTFORMAT
'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
LOCATION
'hdfs://rkk1.hdp.local:8020/apps/hive/warehouse/x'
TBLPROPERTIES (
'COLUMN_STATS_ACCURATE'='{\"BASIC_STATS\":\"true\"}',
'numFiles'='0',
'numRows'='0',
'rawDataSize'='0',
'totalSize'='0',
'transient_lastDdlTime'='1482771845')
Time taken: 0.359 seconds, Fetched: 17 row(s)
hive> alter table x set fileformat inputformat "org.apache.hadoop.hive.ql.io.orc.OrcInputFormat" outputformat "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat" serde "org.apache.hadoop.hive.ql.io.orc.OrcSerde";
OK
Time taken: 0.429 seconds
hive> show create table x;
OK
CREATE TABLE `x`(
`yy` string)
ROW FORMAT SERDE
'org.apache.hadoop.hive.ql.io.orc.OrcSerde'
STORED AS INPUTFORMAT
'org.apache.hadoop.hive.ql.io.orc.OrcInputFormat'
OUTPUTFORMAT
'org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat'
LOCATION
'hdfs://rkk1.hdp.local:8020/apps/hive/warehouse/x'
TBLPROPERTIES (
'last_modified_by'='hive',
'last_modified_time'='1482772074',
'numFiles'='0',
'numRows'='0',
'rawDataSize'='0',
'totalSize'='0',
'transient_lastDdlTime'='1482772074')
Time taken: 0.087 seconds, Fetched: 18 row(s)
hive>
... View more
12-26-2016
04:48 PM
did you check the availability of Nimbus, is it up?
... View more
12-26-2016
04:42 PM
@Sergey Paramoshkin possibly you are hitting this bug in spark codegen https://issues.apache.org/jira/browse/SPARK-18528
... View more
12-26-2016
04:34 PM
@Mustafa El-badry could you please try the suggestions mentioned in the following post and see if it helps? http://stackoverflow.com/questions/40567454/sqoop-imports-empty-strings-for-oracle-9i-table
... View more
12-26-2016
04:31 PM
@Mon key could please try to submit the job again after setting the following property in you configuration? conf.set("mapreduce.app-submission.cross-platform", "true");
... View more
12-26-2016
04:22 PM
@Viraj Vekaria could you please try setting these properties in conf and see if it helps? dfs.nameservices=HadoopTestHA dfs.ha.namenodes.HadoopTestHA=nn1,nn2 dfs.namenode.rpc-address.HadoopTestHA.namenode1=hadoop4ind.india:8020 dfs.namenode.rpc-address.HadoopTestHA.namenode2=hadoop5ind.india:8020 dfs.client.failover.proxy.provider.HadoopTestHA=org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider
... View more
12-26-2016
03:55 PM
1 Kudo
@Santhosh B Gowda if you are using hive-cli/hiveserver2 then get the process id and check lsof -p <pid> | grep jets3t it will tell you what jets3t jar available on the classpath
... View more
12-26-2016
01:06 PM
@Santhosh B Gowda I can see hive is picking right jar from these locations, are you seeing different jar version on classpath?
java 25940 hive mem REG 252,1 539735 1180054 /usr/hdp/2.5.0.0-1133/hadoop-mapreduce/jets3t-0.9.0.jar
java 25940 hive mem REG 252,1 539735 1179933 /usr/hdp/2.5.0.0-1133/hadoop-yarn/lib/jets3t-0.9.0.jar
java 25940 hive mem REG 252,1 539735 1053479 /usr/hdp/2.5.0.0-1133/hadoop/lib/jets3t-0.9.0.jar
java 25940 hive 183r REG 252,1 539735 1053479 /usr/hdp/2.5.0.0-1133/hadoop/lib/jets3t-0.9.0.jar
java 25940 hive 297r REG 252,1 539735 1179933 /usr/hdp/2.5.0.0-1133/hadoop-yarn/lib/jets3t-0.9.0.jar
java 25940 hive 415r REG 252,1 539735 1180054 /usr/hdp/2.5.0.0-1133/hadoop-mapreduce/jets3t-0.9.0.jar
... View more
12-26-2016
12:32 PM
how many of this type(#google=google) of property you are going to have in your applicaiton, if these are fews then you can put in your ReadConfiguration while building propertyMap.
... View more
12-26-2016
12:30 PM
what version of jet2st library is on your classpath? Jet3st 0.9.0 has introduced ServiceException, if you have older library then you need to upgrade lib.
... View more
12-26-2016
10:20 AM
@radhika mantri instead of reading # keys from properties file why don you put null check in your producer or consumer and set this value there like this if (properties.getProperty("#google") == null) {
properties.setProperty("#google", "google");
}
... View more
12-26-2016
07:45 AM
1 Kudo
@Bramantya Anggriawan ideally you should install the kafka broker on the nodes where logs should be collected, topic partition is the unit of parallelism in Kafka. On both the producer and the broker side, writes to different partitions can be done fully in parallel.if you dont have many topics on which you need to produce data to then you can have 1-2 kafka server
... View more
12-25-2016
04:24 PM
@jiang zhixingwith the logs Unregistering application from RM, exitStatus=SUCCEEDED, exitMessage=Session stats:submittedDAGs=0, successfulDAGs=0, failedDAGs=0, killedDAGs=0 it seems that your dag was completed successfully so need not to worry, for interrupt yarn community(https://issues.apache.org/jira/browse/YARN-1022) decided to to change logging level for it to DEBUG but the issue is reproducible
... View more
12-25-2016
03:50 PM
@Aditya Mamidala could you please share your properties file
... View more
12-25-2016
10:56 AM
the counter suggest that dag submission took 34 secs while whole execution time was 51 sec, see if there was resource issue at RM side.
... View more
12-25-2016
10:52 AM
@Huahua Wei lsof output suggest that your zk log location is /var/log/zookeeper/zookeeper-zookeeper-server-insightcluster132.out
... View more