Member since
09-25-2015
37
Posts
45
Kudos Received
4
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
5432 | 07-06-2017 08:46 PM | |
1415 | 10-03-2016 08:51 PM | |
1913 | 06-23-2016 02:24 PM | |
2155 | 05-25-2016 10:04 PM |
11-10-2016
02:36 AM
3 Kudos
Using Ignition version 7.8.4 (b2016082217) | 64-bit Copy dependencies needed to make a connection to the Apache Phoneix Server werecopied into Ignition\lib Edit Igntion\data\ignition.conf to register the dependacy with Ignition's boot strap process Restart Nifi and then begin the process of configuring JDBC drivers by selecting the drivers link in the Databases section of Ignition management console Select create new JDBC Driver on the Drivers Home Page Complete and submit the driver properties page using the following as a guide Now select connections from the data bases menu and create new database connection from the connections home page Complete and submit the connection properties dialog After successfully submitting the configuration the connection should appear on the connections home page If there are any issues with the driver or connection they can be found in the log of the system console With the connection configured we can now added data to the HMI application from queries of the phoenix server Select new query tag define and the SQL statement
... View more
Labels:
10-03-2016
08:51 PM
3 Kudos
When we are requested to retrieve data from an OSI PI System for long term storage or
further processing/analytics. Batch files are the
easiest way to get data out of the OSI PI System. Batch files can be created via Excel or VB
using data link connector or via the PI SDK.Batch data exports from OSI can also be generated via the OSI provided
executable. These various components provide
an interface into the PI data archive storage mechanism and exposes tables,
tags and values. While using batch
files for loading PI Archive data into other systems is reliable it is
cumbersome to maintain, because changes in data model mean changes to scripts,
file formats and client interfaces. OSI provides a JDBC
interface that a programmatic mechanism for interfacing with the OSI data
archive.However there is very little
information available for performance of the JDBC connection for large
datasets. Since the OSI System is usually considered critical infrastructure
and understanding the performance implication of an interface technique is
necessary for production deployments. The last option I will call out is OSI's BI Integrator which can forward data from the PI database into Hive periodically. I have no experience with this particular method.
... View more
06-23-2016
02:24 PM
Found that the options -Pspark-X.X & -Psparkr are required for the build
... View more
06-22-2016
07:51 PM
After building Zeppelin 0.6.0 and I test the %spark.r %r interpreters and receive the error shown in https://issues.apache.org/jira/browse/ZEPPELIN-983 It seems to boil down to class not found for SparkRInterpreter I use the most simple build process as it seems there is no need to use -Pr or Psparkr anymore git clone https://github.com/apache/zeppelin
cd zeppelin
mvn clean package -DskipTests from zeppelin-interpreter-spark-root-hostname.log ERROR [2016-06-17 04:16:18,990] ({pool-1-thread-2} ProcessFunction.java[process]:41) - Internal error processing createInterpreter
org.apache.thrift.TException: java.lang.ClassNotFoundException: org.apache.zeppelin.spark.SparkRInterpreter
at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer.createInterpreter(RemoteInterpreterServer.java:175)
at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Processor$createInterpreter.getResult(RemoteInterpreterService.java:1409)
at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Processor$createInterpreter.getResult(RemoteInterpreterService.java:1394)
at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:285)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.ClassNotFoundException: org.apache.zeppelin.spark.SparkRInterpreter
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:190)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer.createInterpreter(RemoteInterpreterServer.java:148)
... 8 more When I check the dependency jars I find no class SparkRInterpreter though it does exist in the src folder of ~/zeppelin/spark [root@hostname zeppelin]# jar -tf spark/target/zeppelin-spark-0.6.0-SNAPSHOT.jar | grep SparkR
[root@hostname zeppelin]# ls spark/src/main/java/org/apache/zeppelin/spark/ | grep SparkR
SparkRInterpreter.java Every other interpreter works as expected and can be found in the same dependency jar for Spark Interpreter [root@rings hostname]# jar -tf spark/target/zeppelin-spark-0.6.0-SNAPSHOT.jar | grep PySpark
org/apache/zeppelin/spark/PySparkInterpreter.class
org/apache/zeppelin/spark/PySparkInterpreter$PythonInterpretRequest.class
[root@rings hostname]# jar -tf spark/target/zeppelin-spark-0.6.0-SNAPSHOT.jar | grep SparkSql
org/apache/zeppelin/spark/SparkSqlInterpreter.class
[root@rings hostname]# jar -tf spark/target/zeppelin-spark-0.6.0-SNAPSHOT.jar | grep DepInterpreter
org/apache/zeppelin/spark/DepInterpreter.class
... View more
Labels:
- Labels:
-
Apache Spark
-
Apache Zeppelin
05-25-2016
10:04 PM
finally got to the bottom of this one hbase-spark is not in maven central and the error was not "in my face". Once I added the hbase-spark repo to maven everything works as expected. Thanks for your quick replies
... View more
05-25-2016
04:19 PM
@Jitendra Yadav no change with the addtional imports @Ted Yu I am converting this to a stand alone program as we speak, will post later. This is HDP 2.4.2 with Spark 1.6.1 Just FYI I tried val conf = sc.hadoopConfiguration With success but this is essentially the same as the HbaseConfiguration and even receives the same type class. I was thinking these should be interchangeable. Perhaps that is not the case?
... View more
05-25-2016
04:09 PM
I have a Spark application that needs to retrieve data from HBase directly. I provide: import org.apache.spark._
import org.apache.hadoop.hbase.HBaseConfiguration
import org.apache.hadoop.hbase.client.HBaseAdmin
import org.apache.spark.rdd.NewHadoopRDD
val conf = HBaseConfiguration.create()
conf.set(TableInputFormat.INPUT_TABLE, "timeseries") with val hBaseRDD = sc.newAPIHadoopRDD(conf, classOf[TableInputFormat],
classOf[org.apache.hadoop.hbase.io.ImmutableBytesWritable],
classOf[org.apache.hadoop.hbase.client.Result]) I receive the error <console>:113: error: type mismatch;
found : org.apache.hadoop.conf.org.apache.hadoop.conf.org.apache.hadoop.conf.org.apache.hadoop.conf.org.apache.hadoop.conf.Configuration
required: org.apache.hadoop.conf.org.apache.hadoop.conf.org.apache.hadoop.conf.org.apache.hadoop.conf.org.apache.hadoop.conf.Configuration
val hBaseRDD = sc.newAPIHadoopRDD(conf, classOf[TableInputFormat],
^
Which is a very confusing error. Am I missing an import or have the conf types wrong?
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache Spark
02-29-2016
04:01 PM
3 Kudos
I think I am asking a slightly different question than is here https://community.hortonworks.com/questions/6796/how-to-increase-datanode-filesystem-size.html but a solution should help both. SAN issues aside! Is there a method to expand the volume under a datanode directory and have HDFS recognize the new allocated space? For instance if we were to mount a virtual file system, say netapp, in Centos and then expand that filesystem: How would one make the change known to HDFS?
... View more
Labels:
- Labels:
-
Apache Hadoop
11-03-2015
08:24 PM
Thanks Andrew ${now();toNumber():Divide(1000)} got me what I needed.
... View more
10-30-2015
02:41 AM
1 Kudo
According to the docs ${now()} should yeild "current date and time as the number of milliseconds since Midnight GMT on
January 1, 1970." Which sounds like a Unix epoch style timestamp in miliseconds but what I actually receive is Fri Oct 30 02:36:08 UTC 2015 Does now() have options I am missing?
... View more
Labels:
- Labels:
-
Apache NiFi
-
Cloudera DataFlow (CDF)
- « Previous
-
- 1
- 2
- Next »