Member since
04-20-2016
3
Posts
1
Kudos Received
0
Solutions
07-21-2016
09:51 AM
I'm reading from Hbase in a spark Hdp 2.3.4.0-3485 (and I've done reading with lower Hdp releases). Read hbase through a RDD: Configuration hbaseConfiguration = HBaseConfiguration.create();
hbaseConfiguration.set(TableInputFormat.INPUT_TABLE, "sometable");
JavaPairRDD<ImmutableBytesWritable, Result> hbaseRdd = sc.newAPIHadoopRDD(hbaseConfiguration, TableInputFormat.class, ImmutableBytesWritable.class, Result.class);
Or am I missing something?
... View more
07-07-2016
08:40 AM
--driver-java-options "-Dlog4j.debug=true -Dlog4j.configuration=file://${BASEDIR}/src/main/resources/log4j.properties" \ --conf "spark.executor.extraJavaOptions=-Dlog4j.configuration=file:///log4j.properties" \ --files "{${BASEDIR}/src/main/resources}" \ So, for the driver I use the driver-java-options and for the executer I combine the conf and the files.
... View more
06-23-2016
07:40 AM
1 Kudo
I would like to store the HDFS checksums of certain files on HDFS in an other location to detect tampering of the data in those files. Is this a good idea? Will future versions of HDFS deliver the same checksum values? Or should I calculate my own checksums based on the bytes in the raw files?
... View more
Labels:
- Labels:
-
Apache Hadoop