Support Questions

Find answers, ask questions, and share your expertise
Announcements
Check out our newest addition to the community, the Cloudera Data Analytics (CDA) group hub.

Spark Application on YARN .

Contributor

Hi All,

When executing the spark application on YARN cluster can I access the local file system (Underlying OS FS).

Though YARN is pointing to HDFS .

Thanks ,

Param.

1 ACCEPTED SOLUTION

@Param NC

Yes, you can access the local file. Here is the sample:-

spark-shell --master yarn-client

scala> sc.textFile("file:///etc/passwd").count()

res0: Long = 40

View solution in original post

3 REPLIES 3

@Param NC

Yes, you can access the local file. Here is the sample:-

spark-shell --master yarn-client

scala> sc.textFile("file:///etc/passwd").count()

res0: Long = 40

Contributor

@Sandeep Nemuri

Thanks its worked ..I tried this already but forgot to create the file on each node ,now its fine.

And I just got one more question here :

If I run in spark app YARN mode I can set the memory parameter through sparkconfiguration using spark.yarn.driver.memoryOverhead properties ,

Is something similar available for the standalone and local mode ?

Thanks in advance ,

Param.

@Param NC , Please close this thread by accepting the answer and consider asking new question.

Take a Tour of the Community
Don't have an account?
Your experience may be limited. Sign in to explore more.