Member since
06-20-2018
6
Posts
0
Kudos Received
0
Solutions
06-21-2018
02:28 AM
Surprisingly, I closed the vm and opened it again. then It launch fine. I REALLY REALLY THANK YOU
... View more
06-20-2018
04:59 AM
I've been trying to launch Hadoop by typing any command, for example, entering " Hadoop fs -ls" in the terminal and I get this error WARN ipc.Client: Failed to connect to server: quickstart.cloudera/XXX.XXX.XX.128:8020: try once and fail.
java.net.NoRouteToHostException: No route to host
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:494)
at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:648)
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:744)
at org.apache.hadoop.ipc.Client$Connection.access$3000(Client.java:396)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1557)
at org.apache.hadoop.ipc.Client.call(Client.java:1480)
at org.apache.hadoop.ipc.Client.call(Client.java:1441)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230)
at com.sun.proxy.$Proxy16.getFileInfo(Unknown Source)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:786)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:260)
at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104)
at com.sun.proxy.$Proxy17.getFileInfo(Unknown Source)
at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:2131)
at org.apache.hadoop.hdfs.DistributedFileSystem$20.doCall(DistributedFileSystem.java:1265)
at org.apache.hadoop.hdfs.DistributedFileSystem$20.doCall(DistributedFileSystem.java:1261)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1261)
at org.apache.hadoop.fs.Globber.getFileStatus(Globber.java:64)
at org.apache.hadoop.fs.Globber.doGlob(Globber.java:272)
at org.apache.hadoop.fs.Globber.glob(Globber.java:151)
at org.apache.hadoop.fs.FileSystem.globStatus(FileSystem.java:1715)
at org.apache.hadoop.fs.shell.PathData.expandAsGlob(PathData.java:326)
at org.apache.hadoop.fs.shell.Command.expandArgument(Command.java:235)
at org.apache.hadoop.fs.shell.Command.expandArguments(Command.java:218)
at org.apache.hadoop.fs.shell.FsCommand.processRawArguments(FsCommand.java:102)
at org.apache.hadoop.fs.shell.Command.run(Command.java:165)
at org.apache.hadoop.fs.FsShell.run(FsShell.java:315)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
at org.apache.hadoop.fs.FsShell.main(FsShell.java:372)
ls: No Route to Host from quickstart.cloudera/XXX.XXX.XX.128 to quickstart.cloudera:8020 failed on socket timeout exception: java.net.NoRouteToHostException: No route to host; For more details see: http://wiki.apache.org/hadoop/NoRouteToHost Notice Hadoop was running perfectly after my IP suddenly changed to XXX.XXX.XX.129. this error accrued. I don't know how to make/configure Hadoop launch on the same IP address Thank you
... View more
Labels:
- Labels:
-
Apache HBase
-
Cloudera Search
-
HDFS
-
Quickstart VM
06-04-2018
12:52 PM
hello, I Want to make a demo which includes kafka, spark and Hbase. For example having files store it in kafka topic, then spark streaming will read and navigate the files from kafka topic and performs some filtering analysis and stores the records into the Hbase. before saving it hbase, first I need to check the resulted record in Hbase whether if it exists or not based on its' date. also I want to ask something else, can I join table from hive with the data files before performing spark analysis from spark. can anyone please provide a step by step guide in java. I am new to hadoop. Thank you
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache Kafka
-
Apache Spark
01-24-2018
03:07 PM
It works, I really can't thank you enough
... View more
01-24-2018
01:28 PM
My problem is that I've been following a lot of documents that can help me with installing, and after finishing all the steps,I've found that krb5kdc or kadmin is not running / enabling. for example when i type " service kadmin start " or " service krb5kdc start" command. its doesnt do /show me anything. i tried many other commands " systemctl start krb5kdc" or " systemctl enable krb5kdc". So does anyone knows what is the problem. PS. the last document that I've been following is this video : https://www.youtube.com/watch?v=-pBh7fgV6w4&t=434s
... View more
Labels: