1973
Posts
1225
Kudos Received
124
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1842 | 04-03-2024 06:39 AM | |
| 2863 | 01-12-2024 08:19 AM | |
| 1582 | 12-07-2023 01:49 PM | |
| 2345 | 08-02-2023 07:30 AM | |
| 3233 | 03-29-2023 01:22 PM |
08-15-2016
10:41 PM
Are you using any caching? Have you run the count in hive cli or beeline? Or Spark beeline? Have you looked at it with parquet tools? how many records should there be? https://github.com/Parquet/parquet-mr/tree/master/parquet-tools also try with the default SQL Context try: SET spark.sql.hive.metastorePartitionPruning=true Could be an issue between SparkSQL and HiveMetastore.
... View more
08-15-2016
04:38 PM
1 Kudo
did you do it under user DSN? sometimes it shows up there. sometimes you have to add it again with new data source. ODBC is touchy
... View more
08-15-2016
04:05 PM
2 Kudos
select new data source http://hortonworks.com/hadoop-tutorial/how-to-install-and-configure-the-hortonworks-odbc-driver-on-windows-7/
... View more
08-15-2016
01:19 AM
File HDB-Sandbox-2.0-vmware-2016081107592216.zip Description Single-Node VM with HDB 2.0 and MADlib 1.9 on HDP 2.4 Sandbox with core Hadoop services including Apache Zepplin.
Username/Passwords:
1) root/hadoop
2) gpadmin/gpadmin
System Requirements See VM Prerequisites listed here: http://hortonworks.com/wp-content/uploads/2016/02/Import_on_VMware_3_1_2016.pdf
MD5 910051909c7e94646e0cab5d01bc5abb
Documentation View file documentation https://network.pivotal.io/api/v2/products/pivotal-hdb/releases/1695/product_files/5595/download And don't forget the ODBC drivers https://network.pivotal.io/products/pivotal-hdb#/releases/1695/file_groups/230
... View more
08-15-2016
12:47 AM
1 Kudo
Okay, this is my bad. Make sure you set text/html. And the content must be a full HTML file. You cannot include attributes or attach the file for HTML to work. <html><head><title>Test</title></head><body>Message From Twitter
<br>
<b>Test</b><br>
<br>
FileName: ${filename}<br>
<img src="https://pbs.twimg.com/media/Cp2_ftRWYAASiDz.jpg">
</body>
</html>
... View more
08-14-2016
09:32 PM
1 Kudo
If I put HTML in the message area and it doesn't send HTML it just sends tags.
... View more
Labels:
- Labels:
-
Apache NiFi
08-13-2016
01:25 PM
3 Kudos
Use the jdbc interpreter https://zeppelin.apache.org/docs/0.6.0/interpreter/jdbc.html#hive It can be used for Hive and Phoenix.
... View more
08-11-2016
02:59 PM
Sqoop is just regular sqoop. You call it with executeprocess. https://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_literal_sqoop_create_hive_table_literal NiFi + Spark (can be site-to-site, command trigger, kafka) https://community.hortonworks.com/articles/30213/us-presidential-election-tweet-analysis-using-hdfn.html https://community.hortonworks.com/articles/12708/nifi-feeding-data-to-spark-streaming.html
... View more
08-11-2016
01:42 PM
if you have that many rows you need to go parallel and run on multiple nodes. You should probably trigger a Sqoop job or Spark SQL job from NiFi. have a few nodes running at once.
... View more
08-11-2016
01:02 PM
2 Kudos
With Apache NiFi 1.0 you can now act as a simple SMTP server (though it is recommended to sit behind a real SMTP MTA and just get mail forwards). It makes for an easy way to ingest mail, headers and attachments. The first thing you will notice is the awesome new UI, which is much cleaner and a joy to use.' First add a processor, ListenSMTP, this will be your mail gateway/SMTP server. As you can see there's also processors for extracting attachments and headers from Email. You need to make sure you set Listening Port, SMTP hostname and Max. # of Connections. The entire flow for mail processing is pretty simple, but easy to follow. We listen for SMTP over TCP Port (I chose 2025, but with Root access you could run on 25). I send the original flow file right to HDFS. I extract the attachments and put them in a separate HDFS directory and finally pull out the email headers and also send them to an HDFS file. I have a little test flow in the bottom to read a file and send email to our ListenSMTP for testing. If you are running this on an HDP 2.4 sandbox, you will need to install Java 8 and set it as an alternative JDK. http://tecadmin.net/install-java-8-on-centos-rhel-and-fedora/ alternatives --config java Pick Java 8 I added
Java 8 as an alternative and specified Java_HOME in top of bin/nifi.sh so I
could run with Java 8 which is required now. To send a test SMTP message from the command line: telnet localhost 2025
Trying 127.0.0.1...
Connected to localhost.
Escape character is '^]'.
220 sandbox.hortonworks.com ESMTP Apache NiFi
ehlo sandbox
250-sandbox.hortonworks.com
250-8BITMIME
250-SIZE 67108864
250 Ok
MAIL
FROM: <tim@sparkdeveloper.com>
250 Ok
RCPT TO: <tspann@hortonworks.com>
250 Ok
DATA
354 End data with
<CR><LF>.<CR><LF>
hello
.
250 Ok
A better way to test SMTP is with SWAKS See: https://debian-administration.org/article/633/Testing_SMTP_servers_with_SWAKS From Mac: brew install swaks From Centos/RHEL: sudo yum -y install swaks Test Send Email: swaks --to tspann@hortonworks.com --server localhost:2025
Received: from hw13125.home (localhost [127.0.0.1])
by sandbox.hortonworks.com with SMTP (Apache NiFi) id IRPEF4WI
for tspann@hortonworks.com; Wed, 10 Aug 2016 17:19:12 -0400 (EDT)
Date: Wed, 10 Aug 2016 17:19:12 -0400To:
tspann@hortonworks.com
From: tspann@hw13125.home
Subject: test Wed,
10 Aug 2016 17:19:12 -0400
X-Mailer: swaks v20130209.0 jetmore.org/john/code/swaks/
MIME-Version: 1.0
Content-Type: multipart/mixed;
boundary="----=_MIME_BOUNDARY_000_98059"------=_MIME_BOUNDARY_000_98059
Content-Type: text/plain
This is a test mailing
------=_MIME_BOUNDARY_000_98059
Content-Type: application/octet-stream
Content-Disposition: attachment
Content-Transfer-Encoding:
BASE64
------=_MIME_BOUNDARY_000_98059--
It is very easy to configure send an email message to our server you need to put in a hostname and port. Once your down building your flow, make sure you create a template and save the XML off to version control. Creating a template has now moved to the Operate control. If you get lost on what you are working for you can use the search feature from the top right. Remember this is a beta product, not yet ready for production. Wait for HDF 2.0 for supported production usage.
... View more
Labels: