Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3365 | 05-03-2017 05:13 PM | |
2797 | 05-02-2017 08:38 AM | |
3076 | 05-02-2017 08:13 AM | |
3006 | 04-10-2017 10:51 PM | |
1518 | 03-28-2017 02:27 AM |
01-13-2016
07:41 PM
@Ram D usually intermediate data is in user's staging directory so if you're submitting jobs as guest look in /user/guest/.staging, notice the dot in front it's a hidden directory.
... View more
01-13-2016
07:36 PM
you add more disks by adding new nodes called datanodes. If you want to add disk to existing nodes, you will need to change the property in ambari for datanode dir to include the newly created directories. @vijaya inturi
... View more
01-13-2016
11:16 AM
does /usr/hdp/2.x exist @Gerd Koenig? Do you have enough space on /usr?
... View more
01-13-2016
11:05 AM
1 Kudo
you can earn points and build a great reputation on this site if you can write a short article and post it. I am sure there are a lot of customers that would be interested in the same @Mehdi TAZI
... View more
01-13-2016
05:42 AM
just to add, most likely adding the node did not complete as hdp-select statement executed last to symlink /usr/hdp/2.x directory to /usr/hdp/current for client and server libs.
... View more
01-13-2016
05:40 AM
@Gerd Koenig you need to run hdp-select tool with the full version of the HDP you're running on the other servers. Check if you have /usr/hdp/2.x directory that matches your current version on other nodes, once that's confirmed, refer to the docs to execute hdp-select statement associated with your version, then you restart all services on the node and run service checks.
... View more
01-13-2016
05:24 AM
I think you have bigger issues than having your Spark application not running. You need to fix the HA failover first before you can do anything else as it's the core of your environment. Go to the /var/log/hadoop/hdfs/ directory on the 2nd node where namenode is not coming up and review the logs, paste them here and we may help you out. @Benoit Perroud
... View more
01-13-2016
05:20 AM
@Vidya SK here's my tutorial for telnet source and logger sink and additionally hdfs sink. # tested on HDP 2.3.2 Sandbox # Example, single-node Flume configuration using netcat source, memory channel and logger sink # install telnet yum install -y telnet # start flume with this configuration ****************************************************************************** # example.conf: A single-node Flume configuration # Name the components on this agent a1.sources = r1 a1.sinks = k1 a1.channels = c1 # Describe/configure the source a1.sources.r1.type = netcat a1.sources.r1.bind = localhost a1.sources.r1.port = 44444 # Describe the sink a1.sinks.k1.type = logger # Use a channel which buffers events in memory a1.channels.c1.type = memory a1.channels.c1.capacity = 1000 a1.channels.c1.transactionCapacity = 100 # Bind the source and sink to the channel a1.sources.r1.channels = c1 a1.sinks.k1.channel = c1 ****************************************************************************** # in another terminal telnet localhost 44444 # type anything # then in the original terminal tail -f /var/log/flume/flume-a1.log # Exampe netcat source, hdfs sink as DataStream # create hdfs flume directory sudo -u hdfs hdfs dfs -mkdir /flume sudo -u hdfs hdfs dfs -mkdir /flume/events sudo -u hdfs hdfs dfs -chown -R flume:hdfs /flume/events ****************************************************************************** # example.conf: A single-node Flume configuration # Name the components on this agent a1.sources = r1 a1.sinks = k1 a1.channels = c1 # Describe/configure the source a1.sources.r1.type = netcat a1.sources.r1.bind = localhost a1.sources.r1.port = 44444 # Describe the sink a1.sinks.k1.type = hdfs a1.sinks.k1.hdfs.path = /flume/events/%y-%m-%d/%H%M/%S a1.sinks.k1.hdfs.filePrefix = events- a1.sinks.k1.hdfs.round = true a1.sinks.k1.hdfs.roundValue = 10 a1.sinks.k1.hdfs.roundUnit = minute a1.sinks.k1.hdfs.useLocalTimeStamp = true a1.sinks.k1.hdfs.fileType = DataStream # Use a channel which buffers events in memory a1.channels.c1.type = memory a1.channels.c1.capacity = 1000 a1.channels.c1.transactionCapacity = 100 # Bind the source and sink to the channel a1.sources.r1.channels = c1 a1.sinks.k1.channel = c1 ****************************************************************************** # show the output in hdfs sudo -u flume hdfs dfs -ls /flume/events/ sudo -u flume hdfs dfs -ls /flume/events/ sudo -u flume hdfs dfs -cat /flume/events/*/*/*/*
... View more
01-13-2016
05:10 AM
1 Kudo
@Hefei Li we would recommend you remove the new nodes and re-add them again with new FQDNs. Cleaning up all of the issues from adding FQDN after provisioning is going to be messy.
... View more
01-12-2016
03:12 PM
1 Kudo
is the machine with agent still up? Check the uptime. @Anilkumar Panda
... View more