Member since
05-30-2018
1322
Posts
715
Kudos Received
148
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 4045 | 08-20-2018 08:26 PM | |
| 1943 | 08-15-2018 01:59 PM | |
| 2372 | 08-13-2018 02:20 PM | |
| 4104 | 07-23-2018 04:37 PM | |
| 5010 | 07-19-2018 12:52 PM |
07-14-2016
03:56 PM
Can NiFi execute MDX queries? Which processor would I use?
... View more
Labels:
- Labels:
-
Apache NiFi
07-14-2016
03:51 PM
Can NiFi connect to SAP BW? I want to move data from SAP BW/Hana to HDP. Can NiFi do this? which processor would I use?
... View more
Labels:
- Labels:
-
Apache NiFi
07-14-2016
01:56 PM
@Raghu Udiyar you can manage your cluster through puppet/chef which would make call to ambari rest api for changes. this is used a ton in the field with success. I believe you are suggest simply upload new blueprint and cluster should detect changes and perform. its an interested idea. however the current methods work very well with success.
... View more
07-14-2016
04:11 AM
@Emily Sharpe thanks for the insights.
... View more
07-14-2016
03:29 AM
@Raghu Udiyar the blueprint is managed by ambari. So if you are changing your cluster, the blueprint should reflect the change. for example I just added storm to my cluster. I did a fetch on the blueprint and the storm now exist in the blueprint. Does that help or did i misunderstand your questions. cheers
... View more
07-14-2016
03:24 AM
@Binu Mathew do you have any thoughts?
... View more
07-14-2016
03:16 AM
1 Kudo
@ANSARI FAHEEM AHMED I am not sure if I follow your question. You should be able to increase yarn memory independent of node node heap size. Ambari may make recommendations based on the memory available to cluster what the name node heap size should be. So if yarn memory is increase it now believes more memory is available to the cluster and may make a increase/decrease recommendation Run the yarn util scipt available here to play with different configurations.
... View more
07-14-2016
03:12 AM
@Saravanan Ramaraj have you looked into apache knox? The Knox API Gateway is designed as a reverse proxy with consideration for pluggability in the areas of
policy enforcement, through providers and the backend services for which it proxies requests. The Apache Knox Gateway is a REST API Gateway for interacting with Apache Hadoop clusters. The Knox Gateway provides a single access point for all REST interactions with Apache Hadoop clusters. In this capacity, the Knox Gateway is able to provide valuable functionality to aid in the control,
integration, monitoring and automation of critical administrative and analytical needs of the enterprise.
Authentication (LDAP and Active Directory Authentication Provider) Federation/SSO (HTTP Header Based Identity Federation) Authorization (Service Level Authorization) Auditing And then for authorization you can use Apache Ranger which offers a centralized security framework to manage fine-grained access control over Hadoop data access components coupled with kerberos you cluster will be secured and the links shall be authenticed using kerberos and ranger will provide authorization on what services the user has access to. Finally knox will be your perimeter security.
... View more
07-12-2016
04:38 AM
@Kit Menke can you verify your KDC is using udp and not tcp?
... View more
07-12-2016
04:09 AM
@jestin ma found a similar solution here. "You can use date processing functions which have been introduced in Spark 1.5. Assuming you have following data: val df =Seq((1L,"05/26/2016 01:01:01"),(2L,"#$@#@#")).toDF("id","dts") You can use unix_timestamp to parse strings and cast it to timestamp import org.apache.spark.sql.functions.unix_timestamp
val ts = unix_timestamp($"dts","MM/dd/yyyy HH:mm:ss").cast("timestamp")
df.withColumn("ts", ts).show(2,false)// +---+-------------------+---------------------+// |id |dts |ts |// +---+-------------------+---------------------+// |1 |05/26/2016 01:01:01|2016-05-26 01:01:01.0|// |2 |#$@#@# |null |// +---+-------------------+---------------------+ As you can see it covers both parsing and error handling. In Spark < 1.6 you'll have to use use something like this: unix_timestamp($"dts","MM/dd/yyyy HH:mm:ss").cast("double").cast("timestamp") or (unix_timestamp($"dts","MM/dd/yyyy HH:mm:ss")*1000).cast("timestamp") due to SPARK-11724. In Spark < 1.5 you should be able to use these with expr and HiveContext ."
... View more