Member since
10-19-2014
58
Posts
6
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
5868 | 03-20-2016 10:41 PM | |
10790 | 04-26-2015 02:30 AM |
12-02-2022
09:49 AM
I have one doubt that somehow push down predicates are not working have quick read into this article https://engineering.dynatrace.com/blog/optimizing-spark-queries-with-filter-pushdown/
... View more
10-28-2020
04:48 AM
Hi What i have seen is tht the share option only gives you read or read+modify permisson. There is nothing as such execute? If i give read+modify other users will be ale to run the oozie workflow. I have seen it does not happen. As the permission on the underlying hdfs folder for the workflow is only for my user and it does not get modified. drwxrwx--- - kuaksha hue 0 2020-10-28 10:42 /user/hue/oozie/workspaces/hue-oozie-1520605312.96 Please elaborate and help. Regards Akshay
... View more
04-12-2019
06:14 AM
this workflow is working? Just workflow and it will send to email?
... View more
09-18-2017
08:44 AM
@SINGHabhijeet wrote: Hi Romainr Thanks for linking relevant JIRAs and highlighting Hue 4.0 features. We are using Hue 3.12 and it does not have workflow.xml Import/Export feature. > If some XML is missing you could tell us what that way we could improve the editor and remove the need to manually edit the XML. I was looking at adding <arg> </arg> in the workflow.xml as shown here using Hue, but was unable to do so. Can you please point out how we can achieve this? I ran into this error and resolved it by just using Arguments. Earlier was trying to combine <command> and <arg> and since it's a xs:choice in DTD, I just used <arg> to get it working.
... View more
08-14-2016
11:35 AM
1 Kudo
The whole support around Parquet is documented at http://www.cloudera.com/documentation/enterprise/latest/topics/cdh_ig_parquet.html Impala's support for Parquet is ahead of Hive at this moment, while https://issues.apache.org/jira/browse/HIVE-8950 will help it catch up in future. In Hive you will still need to manually specify a column, but you may alternatively create the table in Impala and use it then in Hive. Parquet's loader in Pig supports reading the schema off the file [1] [2], as does Spark's Parquet support [3]. None of the eco system approaches use an external schema file as was the case with Avro storages. [1] - https://github.com/Parquet/parquet-mr/blob/master/parquet-pig/src/main/java/parquet/pig/ParquetLoader.java#L90-L95 [2] - https://github.com/Parquet/parquet-mr/blob/master/parquet-pig/src/test/java/parquet/pig/TestParquetLoader.java#L94-L97 [3] - http://spark.apache.org/docs/latest/sql-programming-guide.html#parquet-files
... View more
08-11-2016
01:45 AM
@sgiri - This has already been answered above. If you use un-secure YARN, all your containers and the commands the container runs (such as in shell action) will run as the "yarn" user (i.e. the user your NodeManager daemon runs as). If you want to run containers in YARN as the actual submitting user, use the LinuxContainerExecutor instead. The caveat of using it is that you need your user accounts available on all NodeManagers so a setuid can be done against them when spawning the container processes.
... View more
07-27-2016
01:23 AM
Hi sairamvj, I would suggest you open a new thread for your question, as it is not related to this topic of this thread. Martin
... View more
05-17-2016
08:28 AM
1 Kudo
What is the right way to pass the -no_multiquery option to Pig from Oozie workflow developed in Hue?
The simple way is to set the option -M / -no_multiquery as parameter in the hue>oozie editor>your pig action>parameters.
Open Hue>Oozie Workflow Editor>click on your job>edit your job> set in parameters -M or -no_multiquery > save
... View more
03-20-2016
10:41 PM
Answering my own question, found this: http://www.cloudera.com/documentation/enterprise/5-5-x/topics/spark_avro.html dzimka, hope this works for you too.
... View more