Archives of Support Questions (Read Only)

This is an archived board for historical reference. Information and links may no longer be available or relevant
Announcements
This board is archived and read-only for historical reference. To ask a new question, please post a new topic on the appropriate active board.

How to upload/download file from webserver(Windows) to Hortonworks sandbox hdfs ?

avatar
New Member

I want to make a software that let users upload a video file , process with Spark and send back the processed file. Until now i can upload files to windows webserver (tomcat). After the file is uploaded i want to send it to hdfs , processed it, and after processing send it back to hdfs and webserver .

1) It's possible to do such thing ? I mean is there a webservice to send files from windows to hdfs hosted on linux?

Is there a shorter path i can follow?

2) Suppose i can upload files to hdfs , how can i trigger the MapReduce job to run , process the video and send it back to webserver ?

1 ACCEPTED SOLUTION

avatar
Master Mentor
@Mihai Mihai

for the file upload - you can leverage webhdfs https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-hdfs/WebHDFS.html

Oozie can help to schedule the workflow to do the processing

View solution in original post

2 REPLIES 2

avatar
Master Mentor

@Mihai Mihai take a look at webhdfs and webhcat APIs. Webhdfs Webhcat

avatar
Master Mentor
@Mihai Mihai

for the file upload - you can leverage webhdfs https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-hdfs/WebHDFS.html

Oozie can help to schedule the workflow to do the processing