Member since
05-22-2016
32
Posts
4
Kudos Received
0
Solutions
06-20-2016
05:10 PM
Just after this discussion I noticed I am using Amabari 1.7 where there is no Spark instalation
... View more
08-21-2018
08:37 AM
we can change the parameter in session level by making the entry in hive-site.xml. example: for property hive.exec.scratchdir hive-site.xml ==ADD <property> <name>hive.security.authorization.sqlstd.confwhitelist.append</name> <value>hive\.exec\.scratchdir|</value> <description>append conf property in white list followed by pipeline</description> </property> == Restart metastore and hiveserver beeline> set hive.exec.scratchdir=/user/hive/scratch; beeline> set hive.exec.scratchdir; Hope it helps, if someone is looking for the same. Thanks, Manjil
... View more
03-13-2019
07:45 AM
Edit the /etc/hive/conf/beeline-log4j.properties you created by copying from beeline-log4j.properties.template and specify log4j.rootLogger=DEBUG, console
... View more
07-05-2016
09:24 AM
1 Kudo
Below our findings: As shown in
the DDL above, bucketing is used in the problematic tables. Bucket number gets
decided according to hashing algorithm, out of 10 buckets for each insert 1
bucket will have actual data file and other 9 buckets will have same file name
with zero size. During this hash calculation race condition is happening when inserting
a new row into the bucketed table via multiple different threads/processes, due
to which 2 or more threads/processes are trying to create the same bucket file. In addition,
as discussed here, the current architecture is not really recommended as over the period of time there would be millions of files on HDFS,
which would create extra overhead on the Namenode. Also select * statement
would take lot of time as it will have to merge all the files from bucket. Solutions which solved both issues: Removed buckets from the two
problematic tables, hece the probability of race conditions will be very less Added hive.support.concurrency=true before the insert statements Weekly Oozie workflow that uses implicit Hive concatenate command on both tables to mitigate the small file problem FYI @Ravi Mutyala
... View more
12-15-2017
06:02 AM
Did you solved the problem? I'm facing the same problem with you. And I tried all methods mentioned above, but it didn't work Here is my exceptions [root@cos1 ~]# sqoop list-tables --connect jdbc:mysql://192.168.2.190/experiment3 --username scott -P 17/12/15 19:17:36 INFO sqoop.Sqoop: Running Sqoop version: 1.4.6
Enter password:
17/12/15 19:17:41 INFO manager.MySQLManager: Preparing to use a MySQL streaming resultset.
17/12/15 19:17:41 ERROR sqoop.Sqoop: Got exception running Sqoop: java.lang.RuntimeException: Could not load db driver class: com.mysql.jdbc.Driver java.lang.RuntimeException: Could not load db driver class: com.mysql.jdbc.Driver at org.apache.sqoop.manager.SqlManager.makeConnection(SqlManager.java:856) at org.apache.sqoop.manager.GenericJdbcManager.getConnection(GenericJdbcManager.java:52) at org.apache.sqoop.manager.CatalogQueryManager.listTables(CatalogQueryManager.java:102) at org.apache.sqoop.tool.ListTablesTool.run(ListTablesTool.java:49)
at org.apache.sqoop.Sqoop.run(Sqoop.java:143) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:179) at org.apache.sqoop.Sqoop.runTool(Sqoop.java:218)
at org.apache.sqoop.Sqoop.runTool(Sqoop.java:227) at org.apache.sqoop.Sqoop.main(Sqoop.java:236) Here is my screenshot I'm using sqoop-1.4.6, hadoop-2.7.4 and mysql-connecctor-java-5.1.38.jar @rajdip chaudhuri @Artem Ervits
... View more