- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
I need to create a spark streaming application which should monitor a hdfs path continuously and read only the new incoming csv data in csv tabular format.
- Labels:
-
Apache Spark
Created ‎08-07-2018 09:02 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Created ‎08-09-2018 11:20 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Thanks @Felix Albani. I found it. For Scala code pls check this git hub link https://github.com/phatak-dev/spark2.0-examples/blob/master/src/main/scala/com/madhukaraphatak/examp.... This is exactly what i needed.
Created ‎08-07-2018 01:13 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Out of the box spark provides the fileStream. You can read more here:
https://spark.apache.org/docs/latest/streaming-programming-guide.html
HTH
*** If you found this answer addressed your question, please take a moment to login and click the "accept" link on the answer.
Created ‎08-09-2018 05:41 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@Sudharsan Ganeshkumar if my answer has helped you please remember to login and mark it as accepted.
Created ‎08-09-2018 11:20 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Thanks @Felix Albani. I found it. For Scala code pls check this git hub link https://github.com/phatak-dev/spark2.0-examples/blob/master/src/main/scala/com/madhukaraphatak/examp.... This is exactly what i needed.
