Member since
08-01-2016
14
Posts
1
Kudos Received
0
Solutions
01-31-2018
07:14 PM
Thanks @Naresh P R. Can you please let me know where can I track all the issues like this? Thanks in advance!
... View more
01-30-2018
03:22 PM
@rtrivedi Attached the screenshot FYR. greatest.jpg
... View more
01-29-2018
11:21 PM
I have a requirement to calculate greatest value in multiple columns. Input: name sub1 sub2 sub3 st1 10 12 15 st2 50 30 35 st3 15 18 13 expected output: name max_sub st1 15 st2 50 st3 18 When I try to use greatest function as below, I'm getting Failed with exception java.io.IOException:org.apache.hadoop.hive.ql.metadata.HiveException: Error evaluating greatest(sub1,sub2,sub3) select name,greatest(sub1,sub2,sub3) from students; any other functions or simple udf available? Note: I cant' use CASE statement as I have to consider around 15 columns.
... View more
Labels:
- Labels:
-
Apache Hive
05-26-2017
01:30 PM
@mclark Thank you for providing such good information improve the performance of NiFi. Can you please provide some best practices for developing NiFi templates, flows and custom processors?
... View more
10-20-2016
05:11 PM
Hi Mani, Below are my observations 1. specifying schema in your DDL and also the Avro schema. i.e column names as part of DDL and Avro URL. use only one. 2. Is the schema in provided URL accessible from HDFS? you can move it HDFS and test 3. The serde is pointing to hcatalog, try using org.apache.hadoop.hive.serde2.avro.AvroSerDe
... View more
09-26-2016
06:52 PM
1 Kudo
This problem occurs primarily due to the lack of a fast random number generation device on the host where the map tasks execute. Pass below argument with sqoop import command export HADOOP_OPTS=-Djava.security.egd=file:/dev/../dev/urandom
sqoop import -D mapred.child.java.opts="-Djava.security.egd=file:/dev/../dev/urandom"
References: http://stackoverflow.com/questions/2327220/oracle-jdbc-intermittent-connection-issue/ https://community.oracle.com/thread/943911?tstart=0&messageID=3793101 https://sqoop.apache.org/docs/1.4.6/SqoopUserGuide.html#_oracle_connection_reset_errors
... View more
08-25-2016
07:13 PM
It is not working for me. Can you let me know if i'm doing anything wrong?? test4 is a table partitioned on lname and is ORC format. the partition I'm trying to merge has just 2 small files. ALTER TABLE test4 PARTITION (lname='vr') CONCATENATE;
... View more
08-25-2016
06:35 PM
Ideally the number of files should be equal to number of buckets declared in a table DDL. It is working fine whenever an initial insert or every insert overwrite is performed. But, insert into hive bucketed table is creating extra files from next insert on wards. ex:
# of Buckets = 4
No. of files after Initial insert --> 4 No. of files after 2nd insert --> 8 No. of files after 3rd insert --> 12 No. of files after n insert --> n* # of Buckets. First insert :
-rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000000_0
-rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000001_0 -rwxrwxrwx 3 hvallur hdfs 308 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000002_0 -rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000003_0 2nd Insert:
-rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000000_0 -rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:47 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000000_0_copy_1
-rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000001_0 -rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:47 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000001_0_copy_1 -rwxrwxrwx 3 hvallur hdfs 308 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000002_0 -rwxrwxrwx 3 hvallur hdfs 302 2016-08-25 12:47 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000002_0_copy_1 -rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:42 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000003_0 -rwxrwxrwx 3 hvallur hdfs 49 2016-08-25 12:47 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=vr/000003_0_copy_1 Does any one know how to resolve this issue ? I thought of writing a program to consolidate all the copy files to related bucket file (i.e before _copy). But, I doubt it might effect Hive meta store mapping because Hive stores the file name related to each bucket#. below query gives the details. SELECT input__file__name,block__offset__inside__file,id from test2 where id%4==3 order by block__offset__inside__file; (Trying to display all the files related to 3rd bucket) hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=v/000003_0_copy_1 163 3 hdfs://dshdp-dev-cluster/apps/hive/warehouse/upsert_testing.db/test3/lname=v/000003_0 166 3 Thank you!
... View more
Labels:
- Labels:
-
Apache Hive
08-17-2016
02:39 PM
Please help me with below scenarios. I have a flow file with a single record as data. Is there a way to apply expression language on content of flow file if I want to add a new flow file attribute using update attribute? in my current approach I'm using extract text to assign the entire row as an attribute and then using update attribute to add new attribute with EL. I feel like it is not a correct approach. Please suggest a better way to do this. Also, Can we add one flow file's attribute to another flow file? ex: I want to extract some part of filename attribute of flow file A and use that to add an attribute for flow file B.
... View more
Labels:
- Labels:
-
Apache NiFi
08-15-2016
04:51 PM
it is same NiFi instance with multiple users.
... View more