Member since
03-01-2018
16
Posts
0
Kudos Received
0
Solutions
09-19-2018
08:27 AM
I am using the Join Processor in SAM to join multiple streams of simulated measurement Data from multiple machines into one stream, where each message contains the entire data for one machine at a given time. The problem is: The data is generated 4x per second, and I can not make the window interval for the Join processor less than 1 second, because I cannot enter a "," symbol into the processor, and when entering a "." and then closing and re-opening the processor, it is broken. This leads to every data message being joined to multiple messages from timestamps that are close, but not the same as the one from the previous data. Is there a way to set the Window Interval <1 second? Or is there an efficient workaround?
... View more
Labels:
- Labels:
-
Apache Storm
09-18-2018
07:52 AM
I get the same error. It seems to be related to timestamps. I am transforming timestamps from text into long format with nifi using Schema Registry, and then, using a Schema with the exact same content as the Avro writer in NiFi, SAM cannot parse the incoming data, seemingly expecting a String-Format timestamp.
... View more
09-13-2018
02:00 PM
I have the same issue, trying to connect to a Postgres remote database. I guess the real question is: Where do I put the JDBC jar and how do I specify it in the "Driver Class Name" in the JDBC Sink. Currently, the jar file is located at /usr/hdf/3.1.2.0-7/streamline/libs/postgresql-9.4.1212.jar and the Driver Class Name that I specified is org.postgresql.postgresql.postgresql-9.4.1212. This is obviously wrong, but what is right? The following Driver Class Names did not work:
org.postgresql.Driver /usr/hdf/3.1.2.0-7/streamline/libs/postgresql-9.4.1212 (with or without ".jar" extension) postgresql-9.4.1212.jar org.postgresql.postgresql.postgresql-9.4.1212.jar libs.postgresql-9.4.1212.jar
... View more
09-12-2018
10:56 AM
How would I use this with a literal letter in my timestamp format? My timestamp format looks like this: yyyy-mm-dd'T'HH:mm:ss.SSS
... View more
09-04-2018
03:22 PM
I solved the problem. It was related to Browser Cache Data. Cleared the Browser Cache for the Browser that I use the SAM UI with, and now it works.
... View more
09-04-2018
03:22 PM
When I try to use the Kafka Consumer in SAM and chose the topic named "SHOWCASE", it says "Schema not Found" although a valid Schema named "SHOWCASE" exists within Schema Registry. It seems like SAM doesn't even connect to Schema Registry for some reason. This is the ERROR-Message: ERROR [2018-09-04 08:40:02.058] [dw-1772 - GET /api/v1/schemas/SHOWCASE_IN_TEMP/MASTER/versions] c.h.s.s.s.GenericExceptionMapper - Got exception: [NotFoundException] / message [HTTP 404 Not Found]
javax.ws.rs.NotFoundException: HTTP 404 Not Found
at org.glassfish.jersey.server.ServerRuntime$2.run(ServerRuntime.java:323)
at org.glassfish.jersey.internal.Errors$1.call(Errors.java:271)
at org.glassfish.jersey.internal.Errors$1.call(Errors.java:267)
at org.glassfish.jersey.internal.Errors.process(Errors.java:315)
at org.glassfish.jersey.internal.Errors.process(Errors.java:297)
at org.glassfish.jersey.internal.Errors.process(Errors.java:267)
at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317)
at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:305)
at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:1154)
at org.glassfish.jersey.servlet.WebComponent.serviceImpl(WebComponent.java:473)
at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:427)
at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:388)
at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:341)
at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:228)
at io.dropwizard.jetty.NonblockingServletHolder.handle(NonblockingServletHolder.java:49)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1689)
at io.dropwizard.servlets.ThreadNameFilter.doFilter(ThreadNameFilter.java:34)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1676)
at io.dropwizard.jersey.filter.AllowedMethodsFilter.handle(AllowedMethodsFilter.java:50)
at io.dropwizard.jersey.filter.AllowedMethodsFilter.doFilter(AllowedMethodsFilter.java:44)
at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1676)
at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:581)
at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1174)
at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:511)
at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1106)
at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
at com.codahale.metrics.jetty9.InstrumentedHandler.handle(InstrumentedHandler.java:240)
at io.dropwizard.jetty.RoutingHandler.handle(RoutingHandler.java:51)
at org.eclipse.jetty.server.handler.gzip.GzipHandler.handle(GzipHandler.java:459)
at io.dropwizard.jetty.BiDiGzipHandler.handle(BiDiGzipHandler.java:68)
at org.eclipse.jetty.server.handler.RequestLogHandler.handle(RequestLogHandler.java:56)
at org.eclipse.jetty.server.handler.StatisticsHandler.handle(StatisticsHandler.java:169)
at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:134)
at org.eclipse.jetty.server.Server.handle(Server.java:524)
at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:319)
at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:253)
at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:273)
at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:95)
at org.eclipse.jetty.io.SelectChannelEndPoint$2.run(SelectChannelEndPoint.java:93)
at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.executeProduceConsume(ExecuteProduceConsume.java:303)
at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.produceConsume(ExecuteProduceConsume.java:148)
at org.eclipse.jetty.util.thread.strategy.ExecuteProduceConsume.run(ExecuteProduceConsume.java:136)
at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:671)
at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:589)
at java.lang.Thread.run(Thread.java:748)
This is the Avro-Schema in SR: {
"type":"record",
"name":"SHOWCASE",
"fields":
[
{
"name":"field1",
"type":"string"
}
]
}
This is the streamline.yaml (without annotations): modules:
- name: tag-registry
className: com.hortonworks.streamline.registries.tag.service.TagRegistryModule
- name: dashboard-registry
className: com.hortonworks.streamline.registries.dashboard.service.DashboardRegistryModule
- name: model-registry
className: com.hortonworks.streamline.registries.model.service.MLModelRegistryModule
- name: streams
className: com.hortonworks.streamline.streams.service.StreamsModule
config:
stormHomeDir: "/usr/hdf/current/storm-client"
schemaRegistryUrl: "http://hdfmetastore.devnet.local:7788/api/v1"
mavenRepoUrl: "hwx-public^http://repo.hortonworks.com/content/groups/public/,hwx-private^http://nexus-private.hortonworks.com/nexus/content/groups/public/"
topologyTestRunResultDir: "/hdf/streamline/topology_test_results"
catalogRootUrl: http://hdfsam.devnet.local:7777/api/v1/catalog
fileStorageConfiguration:
className: com.hortonworks.registries.common.util.LocalFileSystemStorage
properties:
directory: /hdf/streamline/jars
dashboardConfiguration:
url: "http://localhost:9088"
storageProviderConfiguration:
properties:
db.properties:
dataSource.password: "streamline"
dataSource.url: "jdbc:postgresql://hdfmetastore.devnet.local:5432/streamline"
dataSource.user: "streamline"
dataSourceClassName: "org.postgresql.ds.PGSimpleDataSource"
db.type: postgresql
queryTimeoutInSecs: 30
providerClass: "com.hortonworks.registries.storage.impl.jdbc.JdbcStorageManager"
server:
rootPath: '/api/*'
applicationConnectors:
- type: http
port: 7777
adminConnectors:
- type: http
port: 7778
logging:
level: INFO
loggers:
com.hortonworks.streamline: DEBUG
appenders:
- type: file
threshold: INFO
logFormat: "%-6level [%d{yyyy-MM-dd HH:mm:ss.SSS}] [%t] %logger{5} - %X{code} %msg %n"
currentLogFilename: /var/log/streamline/streamline.log
archivedLogFilenamePattern: /var/log/streamline/streamline-%d{yyyy-MM-dd}-%i.log.gz
archivedFileCount: 2
timeZone: UTC
maxFileSize: 64MB
Schema Registry is installed on the hdfmetastore-node and listens on port 7788. Thanks in advance!
... View more
Labels:
09-04-2018
02:40 PM
I have solved the problem with a very uncomfortable work-around: Downgrading the cluster to HDF-3.1.2. Since I have time-pressure, this was the only reliable way to solve the problem in time.
... View more
09-03-2018
10:55 AM
Similar Error here. SAM and Schema Registry seem to not work together correctly. Are there known work-arounds?
... View more
08-24-2018
01:58 PM
Similar problem here. I changed the Kafka-Version-numbers in the *component.json files to the Kafka version associated with HDF 3.2.0.0-520 (1.1.1) but it seems like SAM still tries to download an old Kafka-Version (0.10.2.1).
... View more
08-24-2018
10:04 AM
I am getting the same issue even in HDF 3.2.0. I am running HDF on a CentOS cluster with limited internet access. I activated the Proxy for SAM but it doesn't change anything.
... View more
03-02-2018
10:27 AM
I found a work-around using NiFi: I use NiFi to add additional info to the content of the flow file and I updated the Schema Registry to reflect this, to then Join over the new value in the field "SignalID" as I called it.
... View more
03-02-2018
10:16 AM
Sorry, can't really help you with that. But there is a processor for executing SQL statements. Maybe that might help. @Amira khalifa
... View more
03-01-2018
03:09 PM
How big is your avro Schema? There is always the workaround of hard-coding the column names to the SQL query.
... View more
03-01-2018
02:56 PM
I want to use the Projection Processor in my Streaming Analytics App to concatenate 2 fields of a data stream, so I can use both of them in a join processor. When I try to use the Projection processor, though, and I enter the projection fields and CONCAT-function and the right arguments and a field name, and then I press Ok and re-enter the processor, it is completely empty again. So how do I use the processor in a way that it doesn't reset all the time? Or is there a workaround to use 2 fields on a join processor? EDIT: I am using HDF 3.0.2.0-76
... View more