Created 02-08-2017 09:48 PM
Hi Guys,
I am following this tutorial:
$>hbase org.apache.hadoop.hbase.mapreduce.ImportTsv-Dimporttsv.separator=,-Dimporttsv.columns="HBASE_ROW_KEY,events:driverId,events:driverName,events:eventTime,events:eventType,events:latitudeColumn,events:longitudeColumn,events:routeId,events:routeName,events:truckId" driver_dangerous_event hdfs://sandbox.hortonworks.com:/tmp/data.csv
While trying to run the above command in hbase shell it get stuck at following line: INFO [main] mapreduce.Job: Running job: job_1485012840284_0025
I am doing this tutorial on Azure Sandbox not sure if I need to change something in the command line above?
Thanks
Created 02-08-2017 09:55 PM
Did you verify on the YARN ResourceManager UI that your job was running? It sounds like it was started..
Created 02-08-2017 10:05 PM
It may happen if your YARN cluster has limited resources and they are occupied by other applications. Check the YARN UI for the application status and free resources. Use yarn application -kill appID to remove those that were stuck by some reason.
Created on 02-09-2017 09:20 PM - edited 08-18-2019 04:10 AM
@Josh Elser @Sergey Soldatov Thanks for getting back on this. I am just starting out really so I am struggling to resolve this. I can see that YARN memory consumption is high but I am not sure how to verify that it is running or where to kill the application? My cluster have around 6.9GB of memory, from the screenshot below I can see that YARN Cluster Memory is 2GB. Should I be increasing this? If so by how much and how to do this?
-------------Updated------------------------------
Should I kill all of these?
For example: yarn application -kill application_1485012840284_0017
Thanks for your time.
Created 02-09-2017 09:33 PM
"Should I kill all of these? For example: yarn application -kill application_1485012840284_0017"
If you are not expecting them to complete, yes. Kill them.