Member since
08-10-2017
26
Posts
1
Kudos Received
0
Solutions
08-21-2017
12:59 PM
Thanks csguna for the reply. I'm not exactly sure how to do that. Do I need to make a modification to the code that is published on exercise 3? https://www.cloudera.com/developers/get-started-with-hadoop-tutorial/exercise-3.html
... View more
08-18-2017
12:05 PM
Thanks!! - that seemed to correct the java HDFS errors, but still getting some others. Any ideas on why it is giving so many value not founds? scala> // First we're going to import the classes we need scala> import org.apache.hadoop.mapreduce.Job import org.apache.hadoop.mapreduce.Job scala> import org.apache.hadoop.mapreduce.lib.input.FileInputFormat import org.apache.hadoop.mapreduce.lib.input.FileInputFormat scala> import org.apache.avro.generic.GenericRecord import org.apache.avro.generic.GenericRecord scala> import parquet.hadoop.ParquetInputFormat import parquet.hadoop.ParquetInputFormat scala> import parquet.avro.AvroReadSupport import parquet.avro.AvroReadSupport scala> import org.apache.spark.rdd.RDD import org.apache.spark.rdd.RDD scala> // Then we create RDD's for 2 of the files we imported from MySQL with Sqoop scala> // RDD's are Spark's data structures for working with distributed datasets scala> def rddFromParquetHdfsFile(path: String): RDD[GenericRecord] = { | val job = new Job() | FileInputFormat.setInputPaths(job, path) | ParquetInputFormat.setReadSupportClass(job, | classOf[AvroReadSupport[GenericRecord]]) | return sc.newAPIHadoopRDD(job.getConfiguration, | classOf[ParquetInputFormat[GenericRecord]], | classOf[Void], | classOf[GenericRecord]).map(x => x._2) | } <console>:37: error: not found: value sc return sc.newAPIHadoopRDD(job.getConfiguration, ^ scala> //Changed to line below - val warehouse = "hdfs://{{cluster_data.manager_node_hostname}}/user/hive/warehouse/" scala> val warehouse = "hdfs://quickstart.cloudera/user/hive/warehouse/" warehouse: String = hdfs://quickstart.cloudera/user/hive/warehouse/ scala> val order_items = rddFromParquetHdfsFile(warehouse + "order_items"); <console>:34: error: not found: value rddFromParquetHdfsFile val order_items = rddFromParquetHdfsFile(warehouse + "order_items"); ^ scala> val products = rddFromParquetHdfsFile(warehouse + "products"); <console>:34: error: not found: value rddFromParquetHdfsFile val products = rddFromParquetHdfsFile(warehouse + "products"); ^ scala> // Next, we extract the fields from order_items and products that we care about scala> // and get a list of every product, its name and quantity, grouped by order scala> val orders = order_items.map { x => ( | x.get("order_item_product_id"), | (x.get("order_item_order_id"), x.get("order_item_quantity"))) | }.join( | products.map { x => ( | x.get("product_id"), | (x.get("product_name"))) | } | ).map(x => ( | scala.Int.unbox(x._2._1._1), // order_id | ( | scala.Int.unbox(x._2._1._2), // quantity | x._2._2.toString // product_name | ) | )).groupByKey() <console>:32: error: not found: value order_items val orders = order_items.map { x => ( ^ scala> // Finally, we tally how many times each combination of products appears scala> // together in an order, then we sort them and take the 10 most common scala> val cooccurrences = orders.map(order => | ( | order._1, | order._2.toList.combinations(2).map(order_pair => | ( | if (order_pair(0)._2 < order_pair(1)._2) | (order_pair(0)._2, order_pair(1)._2) | else | (order_pair(1)._2, order_pair(0)._2), | order_pair(0)._1 * order_pair(1)._1 | ) | ) | ) | ) <console>:32: error: not found: value orders val cooccurrences = orders.map(order => ^ scala> val combos = cooccurrences.flatMap(x => x._2).reduceByKey((a, b) => a + b) <console>:32: error: not found: value cooccurrences val combos = cooccurrences.flatMap(x => x._2).reduceByKey((a, b) => a + b) ^ scala> val mostCommon = combos.map(x => (x._2, x._1)).sortByKey(false).take(10) <console>:32: error: not found: value combos val mostCommon = combos.map(x => (x._2, x._1)).sortByKey(false).take(10) ^ scala> // We print our results, 1 per line, and exit the Spark shell scala> println(mostCommon.deep.mkString("\n")) <console>:33: error: not found: value mostCommon println(mostCommon.deep.mkString("\n"))
... View more
08-17-2017
08:37 AM
I'm getting several java.io. - errors when trying to run the code in exercise 3. Any Ideas? scala> // First we're going to import the classes we need scala> import org.apache.hadoop.mapreduce.Job import org.apache.hadoop.mapreduce.Job scala> import org.apache.hadoop.mapreduce.lib.input.FileInputFormat import org.apache.hadoop.mapreduce.lib.input.FileInputFormat scala> import org.apache.avro.generic.GenericRecord import org.apache.avro.generic.GenericRecord scala> import parquet.hadoop.ParquetInputFormat import parquet.hadoop.ParquetInputFormat scala> import parquet.avro.AvroReadSupport import parquet.avro.AvroReadSupport scala> import org.apache.spark.rdd.RDD import org.apache.spark.rdd.RDD scala> // Then we create RDD's for 2 of the files we imported from MySQL with Sqoop scala> // RDD's are Spark's data structures for working with distributed datasets scala> def rddFromParquetHdfsFile(path: String): RDD[GenericRecord] = { | val job = new Job() | FileInputFormat.setInputPaths(job, path) | ParquetInputFormat.setReadSupportClass(job, | classOf[AvroReadSupport[GenericRecord]]) | return sc.newAPIHadoopRDD(job.getConfiguration, | classOf[ParquetInputFormat[GenericRecord]], | classOf[Void], | classOf[GenericRecord]).map(x => x._2) | } warning: there were 1 deprecation warning(s); re-run with -deprecation for details rddFromParquetHdfsFile: (path: String)org.apache.spark.rdd.RDD[org.apache.avro.generic.GenericRecord] scala> scala> val warehouse = "hdfs://{{cluster_data.manager_node_hostname}}/user/hive/warehouse/" warehouse: String = hdfs://{{cluster_data.manager_node_hostname}}/user/hive/warehouse/ scala> val order_items = rddFromParquetHdfsFile(warehouse + "order_items"); java.io.IOException: Incomplete HDFS URI, no host: hdfs://%7B%7Bcluster_data.manager_node_hostname%7D%7D/user/hive/warehouse/order_items at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2800) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:98) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2837) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2819) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:387) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296) at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.setInputPaths(FileInputFormat.java:507) at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.setInputPaths(FileInputFormat.java:476) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.rddFromParquetHdfsFile(<console>:42) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:44) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:49) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:51) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:53) at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:55) at $iwC$$iwC$$iwC$$iwC.<init>(<console>:57) at $iwC$$iwC$$iwC.<init>(<console>:59) at $iwC$$iwC.<init>(<console>:61) at $iwC.<init>(<console>:63) at <init>(<console>:65) at .<init>(<console>:69) at .<clinit>(<console>) at .<init>(<console>:7) at .<clinit>(<console>) at $print(<console>) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1045) at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1326) at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:821) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:852) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:800) at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857) at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902) at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814) at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657) at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1064) at org.apache.spark.repl.Main$.main(Main.scala:35) at org.apache.spark.repl.Main.main(Main.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) scala> val products = rddFromParquetHdfsFile(warehouse + "products"); java.io.IOException: Incomplete HDFS URI, no host: hdfs://%7B%7Bcluster_data.manager_node_hostname%7D%7D/user/hive/warehouse/products at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:149) at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2800) at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:98) at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2837) at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2819) at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:387) at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296) at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.setInputPaths(FileInputFormat.java:507) at org.apache.hadoop.mapreduce.lib.input.FileInputFormat.setInputPaths(FileInputFormat.java:476) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.rddFromParquetHdfsFile(<console>:42) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:44) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:49) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:51) at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:53) at $iwC$$iwC$$iwC$$iwC$$iwC.<init>(<console>:55) at $iwC$$iwC$$iwC$$iwC.<init>(<console>:57) at $iwC$$iwC$$iwC.<init>(<console>:59) at $iwC$$iwC.<init>(<console>:61) at $iwC.<init>(<console>:63) at <init>(<console>:65) at .<init>(<console>:69) at .<clinit>(<console>) at .<init>(<console>:7) at .<clinit>(<console>) at $print(<console>) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1045) at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1326) at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:821) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:852) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:800) at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857) at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902) at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814) at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657) at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945) at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1064) at org.apache.spark.repl.Main$.main(Main.scala:35) at org.apache.spark.repl.Main.main(Main.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) scala> // Next, we extract the fields from order_items and products that we care about scala> // and get a list of every product, its name and quantity, grouped by order scala> val orders = order_items.map { x => ( | x.get("order_item_product_id"), | (x.get("order_item_order_id"), x.get("order_item_quantity"))) | }.join( | products.map { x => ( | x.get("product_id"), | (x.get("product_name"))) | } | ).map(x => ( | scala.Int.unbox(x._2._1._1), // order_id | ( | scala.Int.unbox(x._2._1._2), // quantity | x._2._2.toString // product_name | ) | )).groupByKey() <console>:38: error: not found: value order_items val orders = order_items.map { x => ( ^ scala> // Finally, we tally how many times each combination of products appears scala> // together in an order, then we sort them and take the 10 most common scala> val cooccurrences = orders.map(order => | ( | order._1, | order._2.toList.combinations(2).map(order_pair => | ( | if (order_pair(0)._2 < order_pair(1)._2) | (order_pair(0)._2, order_pair(1)._2) | else | (order_pair(1)._2, order_pair(0)._2), | order_pair(0)._1 * order_pair(1)._1 | ) | ) | ) | ) <console>:38: error: not found: value orders val cooccurrences = orders.map(order => ^ scala> val combos = cooccurrences.flatMap(x => x._2).reduceByKey((a, b) => a + b) <console>:38: error: not found: value cooccurrences val combos = cooccurrences.flatMap(x => x._2).reduceByKey((a, b) => a + b) ^ scala> val mostCommon = combos.map(x => (x._2, x._1)).sortByKey(false).take(10) <console>:38: error: not found: value combos val mostCommon = combos.map(x => (x._2, x._1)).sortByKey(false).take(10) ^ scala> // We print our results, 1 per line, and exit the Spark shell scala> println(mostCommon.deep.mkString("\n")) <console>:39: error: not found: value mostCommon println(mostCommon.deep.mkString("\n"))
... View more
Labels:
- Labels:
-
Apache Spark
-
Quickstart VM
08-17-2017
05:00 AM
I found the file hive-contrib.jar the command ls /usr/lib/have/lib via the terminal window. I ran ADD JAR /usr/lib/hive/lib/hive-contrib.jar; from the HUE Hive and it worked! I also was able to run the last statement successfully INSERT OVERWRITE TABLE tokenized_access_logs SELECT * FROM intermediate_access_logs; I have now been able to complete exercise 2!! Big thanks for the help!
... View more
08-16-2017
03:12 PM
I am running this in HUE Hive ADD JAR /opt/cloudera/parcels/CDH/lib/hive/lib/hive-contrib.jar Error while processing statement: /opt/cloudera/parcels/CDH/lib/hive/lib/hive-contrib.jar does not exist Hopefully the screen shot will come though, but when I do ls from terminal window I see the folder "parcels" but it is highlighted green. If I try ls /parcels its says No Such File or Directory
... View more
08-16-2017
01:42 PM
Which code/script is running? How can I tell? I ran this but got similar error. ADD JAR /lib/hive/lib/hive-contrib.jar; Error while processing statement: /lib/hive/lib/hive-contrib.jar does not exist Really appreciate you work with me on this!!
... View more
08-16-2017
10:24 AM
Based on another article I have tried the following combinations but all issue and error that the hive-contrib.jar does not exist. ADD JAR /hive/lib/hive-contrib.jar; ADD JAR /home/cloudera/lib/hive-contrib.jar; ADD JAR /opt/cloudera/parcels/CDH/lib/hive/hive-contrib.jar;
... View more
08-16-2017
09:53 AM
In the Cloudera Management Service Actions under the CDH 5 Packages dashboard I restarted the service. This fixed the issue and I was able to complete this step successfully as well, but get another error described below. CREATE EXTERNAL TABLE tokenized_access_logs ( ip STRING, date STRING, method STRING, url STRING, http_version STRING, code1 STRING, code2 STRING, dash STRING, user_agent STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' LOCATION '/user/hive/warehouse/tokenized_access_logs'; I did however get the following error when I hit this line of code. Any ideas? ADD JAR {{lib_dir}}/hive/lib/hive-contrib.jar; Error while processing statement: java.net.URISyntaxException: Illegal character in path at index 0: {{lib_dir}}/hive/lib/hive-contrib.jar
... View more
08-11-2017
09:15 AM
1 Kudo
There are a few other post on the same topic however none of the suggested fixes have worked. This query (which I am using HUE Hive) just hangs. I also started the Zookeeper as suggested in another post, but it did not help. I did let it run over night and it produced the following error:
Error while processing statement: FAILED: Error in acquiring locks: Locks on the underlying objects cannot be acquired. retry after some time
Anyone know a solution?
CREATE EXTERNAL TABLE intermediate_access_logs ( ip STRING, date STRING, method STRING, url STRING, http_version STRING, code1 STRING, code2 STRING, dash STRING, user_agent STRING) ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe' WITH SERDEPROPERTIES ( 'input.regex' = '([^ ]*) - - \\[([^\\]]*)\\] "([^\ ]*) ([^\ ]*) ([^\ ]*)" (\\d*) (\\d*) "([^"]*)" "([^"]*)"', 'output.format.string' = "%1$$s %2$$s %3$$s %4$$s %5$$s %6$$s %7$$s %8$$s %9$$s") LOCATION '/user/hive/warehouse/original_access_logs';
... View more
Labels:
07-17-2017
04:49 PM
Big thanks for this!! I really appreciate the help!!
... View more
07-14-2017
08:31 PM
trying to perform a simple test using HIVE + PIG using the following PIG script. --Load records from populated table
int = LOAD 'table1' USING org.apache.hive.hcatalog.pig.HCatLoader();
--store the records to a new table
store int into 'table2' using org.apache.hive.hcatalog.pig.HCatStorer();
-useHCatalog added Any idea on the error? No records get process into the identical and empty table 2. WARNING: Use "yarn jar" to launch YARN applications.
17/07/14 20:08:09 INFO pig.ExecTypeProvider: Trying ExecType : LOCAL
17/07/14 20:08:09 INFO pig.ExecTypeProvider: Trying ExecType : MAPREDUCE
17/07/14 20:08:09 INFO pig.ExecTypeProvider: Trying ExecType : TEZ_LOCAL
17/07/14 20:08:09 INFO pig.ExecTypeProvider: Trying ExecType : TEZ
17/07/14 20:08:09 INFO pig.ExecTypeProvider: Picked TEZ as the ExecType
2017-07-14 20:08:09,110 [main] INFO org.apache.pig.Main - Apache Pig version 0.16.0.2.5.0.0-1245 (rexported) compiled Aug 26 2016, 02:07:35
2017-07-14 20:08:09,110 [main] INFO org.apache.pig.Main - Logging error messages to: /hadoop/yarn/local/usercache/maria_dev/appcache/application_1499717461125_0015/container_1499717461125_0015_01_000002/pig_1500062889108.log
2017-07-14 20:08:10,228 [main] INFO org.apache.pig.impl.util.Utils - Default bootup file /home/yarn/.pigbootup not found
2017-07-14 20:08:10,418 [main] INFO org.apache.pig.backend.hadoop.executionengine.HExecutionEngine - Connecting to hadoop file system at: hdfs://sandbox.hortonworks.com:8020
2017-07-14 20:08:11,609 [main] INFO org.apache.pig.PigServer - Pig Script ID for the session: PIG-script.pig-edf7096e-5d22-4b5d-a99a-8a1b879aee01
2017-07-14 20:08:12,279 [main] INFO org.apache.hadoop.yarn.client.api.impl.TimelineClientImpl - Timeline service address: http://sandbox.hortonworks.com:8188/ws/v1/timeline/
2017-07-14 20:08:12,493 [main] INFO org.apache.pig.backend.hadoop.PigATSClient - Created ATS Hook
2017-07-14 20:08:13,025 [main] ERROR org.apache.pig.PigServer - exception during parsing: Error during parsing. <file script.pig, line 2, column 0> mismatched input 'int' expecting EOF
Failed to parse: <file script.pig, line 2, column 0> mismatched input 'int' expecting EOF
at org.apache.pig.parser.QueryParserDriver.parse(QueryParserDriver.java:244)
at org.apache.pig.parser.QueryParserDriver.parse(QueryParserDriver.java:182)
at org.apache.pig.PigServer$Graph.parseQuery(PigServer.java:1819)
at org.apache.pig.PigServer$Graph.access$000(PigServer.java:1527)
at org.apache.pig.PigServer.parseAndBuild(PigServer.java:460)
at org.apache.pig.PigServer.executeBatch(PigServer.java:485)
at org.apache.pig.PigServer.executeBatch(PigServer.java:471)
at org.apache.pig.tools.grunt.GruntParser.executeBatch(GruntParser.java:172)
at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:235)
at org.apache.pig.tools.grunt.GruntParser.parseStopOnError(GruntParser.java:206)
at org.apache.pig.tools.grunt.Grunt.exec(Grunt.java:81)
at org.apache.pig.Main.run(Main.java:503)
at org.apache.pig.Main.main(Main.java:178)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:233)
at org.apache.hadoop.util.RunJar.main(RunJar.java:148)
2017-07-14 20:08:13,030 [main] ERROR org.apache.pig.tools.grunt.Grunt - ERROR 1200: <file script.pig, line 2, column 0> mismatched input 'int' expecting EOF
Details at logfile: /hadoop/yarn/local/usercache/maria_dev/appcache/application_1499717461125_0015/container_1499717461125_0015_01_000002/pig_1500062889108.log
2017-07-14 20:08:13,087 [main] INFO org.apache.pig.Main - Pig script completed in 4 seconds and 218 milliseconds (4218 ms)
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Pig
-
Apache YARN
07-07-2017
11:15 AM
I did try changing from %spark2 that is in the tutorial apparently written for HDP 2.6 to % spark %spark
val hiveContext = new org.apache.spark.sql.SparkSession.Builder().getOrCreate() This is the error I received. <console>:27: error: object SparkSession is not a member of package org.apache.spark.sql
val hiveContext = new org.apache.spark.sql.SparkSession.Builder().getOrCreate() I get similar errors when I try and run the tutorial code in Scala logged in as 'root'.
... View more
07-06-2017
07:45 PM
I believe I'm running HDP 2.5 and I'm getting the following errors running these commands in Zeppelin as part of th eSpark Risk Factor Tutorial. When these are run I get Prefix not found. Both Spark and Spark2 show running (not in maintenance mode) Ambari dashboard. Any ideas? %spark2 valhiveContext=neworg.apache.spark.sql.SparkSession.Builder().getOrCreate() %spark2 hiveContext.sql("show tables").show()
... View more
Labels:
- Labels:
-
Apache Spark
07-06-2017
03:44 PM
Big thanks for that!! Now back on track with the tutorial because I was getting an error trying to delete that record!!
... View more
06-30-2017
12:13 PM
Unfortunately I don't have Hive View 2.0 in my drop down menu. I also don't have Workflow Manager as in your screenshot logged in as maria_dev.
... View more
06-29-2017
09:50 PM
where-is-the-first-row-header-checkbox.pngI'm working to complete the Hive - Data ETL tutorial on HDP2.5 and noticed there is not a Table link on the Ribbon bar, but more importantly there is no Is First Row Header check box!
... View more
Labels:
06-29-2017
08:37 PM
I switched to using Chrome instead of IE v 11 and it seems to be correcting all my issues! 🙂
... View more
06-29-2017
08:18 PM
I'm getting the following error when trying top uploading the 2 files for the HDP 2.5 Loading Sensor Data into HDFS tutorial. Server Message Invalid path name /user/maria_dev/data/C:UsersABDocumentsDad's StuffHadoopgeolocation.csv
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2454)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2417)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:729)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:405)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307) Trace Error org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.RemoteException): Invalid path name /user/maria_dev/data/C:UsersABDocumentsDad's StuffHadoopgeolocation.csv
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2454)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2417)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:729)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:405)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307) at org.apache.hadoop.hdfs.web.JsonUtil.toRemoteException(JsonUtil.java:118)
at org.apache.hadoop.hdfs.web.WebHdfsFileSystem.validateResponse(WebHdfsFileSystem.java:477)
at org.apache.hadoop.hdfs.web.WebHdfsFileSystem.access$200(WebHdfsFileSystem.java:113)
at org.apache.hadoop.hdfs.web.WebHdfsFileSystem$FsPathOutputStreamRunner$1.close(WebHdfsFileSystem.java:936)
at org.apache.ambari.view.commons.hdfs.UploadService.uploadFile(UploadService.java:64)
at org.apache.ambari.view.commons.hdfs.UploadService.uploadFile(UploadService.java:87)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at com.sun.jersey.spi.container.JavaMethodInvokerFactory$1.invoke(JavaMethodInvokerFactory.java:60)
at com.sun.jersey.server.impl.model.method.dispatch.AbstractResourceMethodDispatchProvider$ResponseOutInvoker._dispatch(AbstractResourceMethodDispatchProvider.java:205)
at com.sun.jersey.server.impl.model.method.dispatch.ResourceJavaMethodDispatcher.dispatch(ResourceJavaMethodDispatcher.java:75)
at com.sun.jersey.server.impl.uri.rules.HttpMethodRule.accept(HttpMethodRule.java:302)
at com.sun.jersey.server.impl.uri.rules.SubLocatorRule.accept(SubLocatorRule.java:137)
at com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at com.sun.jersey.server.impl.uri.rules.SubLocatorRule.accept(SubLocatorRule.java:137)
at com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at com.sun.jersey.server.impl.uri.rules.SubLocatorRule.accept(SubLocatorRule.java:137)
at com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at com.sun.jersey.server.impl.uri.rules.SubLocatorRule.accept(SubLocatorRule.java:137)
at com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at com.sun.jersey.server.impl.uri.rules.SubLocatorRule.accept(SubLocatorRule.java:137)
at com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at com.sun.jersey.server.impl.uri.rules.ResourceClassRule.accept(ResourceClassRule.java:108)
at com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
at com.sun.jersey.server.impl.uri.rules.RootResourceClassesRule.accept(RootResourceClassesRule.java:84)
at com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1542)
at com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1473)
at com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1419)
at com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1409)
at com.sun.jersey.spi.container.servlet.WebComponent.service(WebComponent.java:409)
at com.sun.jersey.spi.container.servlet.ServletContainer.service(ServletContainer.java:558)
at com.sun.jersey.spi.container.servlet.ServletContainer.service(ServletContainer.java:733)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:684)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1507)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:330)
at org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:118)
at org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:84)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.apache.ambari.server.security.authorization.AmbariAuthorizationFilter.doFilter(AmbariAuthorizationFilter.java:257)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:113)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:103)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:113)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:54)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:45)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.apache.ambari.server.security.authorization.jwt.JwtAuthenticationFilter.doFilter(JwtAuthenticationFilter.java:96)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.authentication.www.BasicAuthenticationFilter.doFilter(BasicAuthenticationFilter.java:150)
at org.apache.ambari.server.security.authentication.AmbariAuthenticationFilter.doFilter(AmbariAuthenticationFilter.java:88)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.apache.ambari.server.security.authorization.AmbariUserAuthorizationFilter.doFilter(AmbariUserAuthorizationFilter.java:91)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.context.SecurityContextPersistenceFilter.doFilter(SecurityContextPersistenceFilter.java:87)
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:342)
at org.springframework.security.web.FilterChainProxy.doFilterInternal(FilterChainProxy.java:192)
at org.springframework.security.web.FilterChainProxy.doFilter(FilterChainProxy.java:160)
at org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:237)
at org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:167)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.apache.ambari.server.api.MethodOverrideFilter.doFilter(MethodOverrideFilter.java:72)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.apache.ambari.server.api.AmbariPersistFilter.doFilter(AmbariPersistFilter.java:47)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.apache.ambari.server.view.AmbariViewsMDCLoggingFilter.doFilter(AmbariViewsMDCLoggingFilter.java:54)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.apache.ambari.server.view.ViewThrottleFilter.doFilter(ViewThrottleFilter.java:161)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.apache.ambari.server.security.AbstractSecurityHeaderFilter.doFilter(AbstractSecurityHeaderFilter.java:109)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.apache.ambari.server.security.AbstractSecurityHeaderFilter.doFilter(AbstractSecurityHeaderFilter.java:109)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.eclipse.jetty.servlets.UserAgentFilter.doFilter(UserAgentFilter.java:82)
at org.eclipse.jetty.servlets.GzipFilter.doFilter(GzipFilter.java:294)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1478)
at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:499)
at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:557)
at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1086)
at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:427)
at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1020)
at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
at org.apache.ambari.server.controller.AmbariHandlerList.processHandlers(AmbariHandlerList.java:212)
at org.apache.ambari.server.controller.AmbariHandlerList.processHandlers(AmbariHandlerList.java:201)
at org.apache.ambari.server.controller.AmbariHandlerList.handle(AmbariHandlerList.java:150)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:116)
at org.eclipse.jetty.server.Server.handle(Server.java:370)
at org.eclipse.jetty.server.AbstractHttpConnection.handleRequest(AbstractHttpConnection.java:494)
at org.eclipse.jetty.server.AbstractHttpConnection.content(AbstractHttpConnection.java:984)
at org.eclipse.jetty.server.AbstractHttpConnection$RequestHandler.content(AbstractHttpConnection.java:1045)
at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:861)
at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:231)
at org.eclipse.jetty.server.AsyncHttpConnection.handle(AsyncHttpConnection.java:82)
at org.eclipse.jetty.io.nio.SelectChannelEndPoint.handle(SelectChannelEndPoint.java:696)
at org.eclipse.jetty.io.nio.SelectChannelEndPoint$1.run(SelectChannelEndPoint.java:53)
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:608)
at org.eclipse.jetty.util.thread.QueuedThreadPool$3.run(QueuedThreadPool.java:543)
at java.lang.Thread.run(Thread.java:745)
... View more
Labels:
06-29-2017
05:53 PM
I'm using HDP 2.5 sandbox and when I try to go into the Hive View its blank. All other views, Pig View, Tez View, Storm View, etc all work fine.
... View more
Labels:
- Labels:
-
Apache Hive
06-29-2017
05:37 PM
I'm trying to complete Loading Sensor Data into HDFS tutorial, but when I try and create the "data" folder in File VIew user > maria_dev it doesn't show. I do get the message it was created successfully.
... View more
Labels: