Member since
12-19-2016
149
Posts
15
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3913 | 04-04-2017 03:01 PM | |
1698 | 01-17-2017 10:44 AM |
08-02-2017
05:24 AM
You can create a Schema(which is similar to databases) by using following grammar. https://phoenix.apache.org/language/index.html#create_schema Schema will be mapped to a namespace in HBase so your tables can be segregated logically as well as physically. https://phoenix.apache.org/namspace_mapping.html
... View more
04-04-2017
03:01 PM
1 Kudo
I got the fix. This was all because of docker version mismatching with selinux. I tried to install lesser version of docker and it worked for me
yum -y install docker-engine-1.12.6-1.el7.centos docker-engine-selinux-1.12.6-1.el7.centos
... View more
06-20-2017
03:29 AM
I believe #39 is wrong. Access after expiry shouldn't be Yes for all (public).Can you please tell me what format the date should be , is it fixed or we can change ?
... View more
01-09-2018
06:06 PM
IBM offers free courses in Scala and other languages, they are free. There are tests at the end of the course once successful you can earn badges and showcase them. https://cognitiveclass.ai/
... View more
02-07-2019
03:11 PM
val path = "adl://azuredatalakestore.net/xxx/Budget/*.xlsx"
val sc = spark.sparkContext val data = sc.wholeTextFiles(path) var z: Array[String] = new Array[String](7)
var i=1
val files = data.map { case (filename, content) => filename } files.collect.foreach(filename => { println(i + "->" + filename) z(i) = filename
println(z(i)) i = i + 1})
... View more
02-10-2017
12:00 PM
@Joe Widen Thank you Sir, But I think if we do join for a larger dataset memory issues will happen. So in such case can we use if/else or look up function here . My Aim is to match input_file DFwith gsam DF and if CCKT_NO = ckt_id and SEV_LVL = 3 then print complete row for that ckt_id.
... View more
11-03-2017
09:01 AM
I had the same error. Based on this recomendation I added new properties in Zeppelin Spark2 interpreter as spark.deploy.maxExecutorRetries=10 It worked for me.
... View more
02-10-2017
12:09 PM
@Binu Mathew Thanks for the python code. Am tryin to do it in both scala n python for knowledge purpose. Am using Spark 1.6.2 . Yes I have created SQLContext.
... View more
02-08-2017
04:48 PM
1 Kudo
There is a JDBC RDD function: newJdbcRDD(sc: SparkContext, getConnection: () ⇒ Connection, sql: String, lowerBound: Long, upperBound: Long, numPartitions: Int, mapRow: (ResultSet) ⇒ T = JdbcRDD.resultSetToObjectArray)(implicit arg0: ClassTag[T])
... View more
02-09-2017
12:38 PM
Hi Frank lu Can you provide me the code with the example i didnt exactly what you are saying
... View more