Member since
02-01-2019
650
Posts
143
Kudos Received
117
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2818 | 04-01-2019 09:53 AM | |
1460 | 04-01-2019 09:34 AM | |
6905 | 01-28-2019 03:50 PM | |
1574 | 11-08-2018 09:26 AM | |
3801 | 11-08-2018 08:55 AM |
01-22-2019
11:04 PM
Thanks, @Sandeep Nemuri
... View more
12-25-2018
12:53 PM
Hi @Michael Mester, Can you please see if this commend helps you. login and accept this answer if it did. 🙂
... View more
11-08-2018
08:39 AM
@Muhammad Taimoor Yousaf You'd need to have a valid ticket in in the machine you are launching pyspark shell (which is test-m1). Looking at the exception it is clearly saying 'No valid credentials provided'. Hope this helps.
... View more
10-03-2018
07:05 AM
@Anurag Mishra If the response answered your question can you take time an login and "Accept" the answer and close the thread so other members can use it as a solution
... View more
09-25-2018
12:02 PM
@Sandeep Nemuri Thanks
... View more
09-14-2018
09:28 AM
@Ronnie 10 , Do consider accepting the answer if it helped you 🙂
... View more
08-13-2018
12:59 PM
@rinu shrivastav The split size is calculated by the formula:- max(mapred.min.split.size, min(mapred.max.split.size, dfs.block.size))
Say, HDFS block size is 64 MB and min.input.size is set to 128MB, then there will be split size would be 128MB. To read 256MB of data, there will be two mappers. To increase the number of mappers, then you could decrease min.input.size till the HDFS block size. split size=max(128,min(256,64))
... View more