Member since
04-20-2019
9
Posts
0
Kudos Received
0
Solutions
04-25-2019
12:43 AM
Let's try to rule out various types of problems. 1. Are you able to read/write to Kerberos-enabled HDFS with PySpark? Is Kudu the only Kerberos-enabled service that is not working from within PySpark? 2. Have you checked to ensure that the Spark driver is running on the host and shell you kinited from instead of being started in a YARN container? If it's running in YARN you have to give YARN access to the keytab to run as. 3. Have you tried connecting to Kudu with the regular Spark shell? Does it work? For examples see https://kudu.apache.org/docs/developing.html#_kudu_integration_with_spark
... View more
04-22-2019
04:07 PM
Thanks Will try that,do you have any suggestion on best way to implementing dimension with scd2 type in hadoop, our dimension table has several sources and all should be able to load /update concurrently in dimension table. #- Please type your reply above this line -##
... View more