Community Articles

Find and share helpful community-sourced technical articles.
avatar
Master Guru

The all new Cloudera Data Engineering ExperienceThe all new Cloudera Data Engineering Experience

 

I recently had the opportunity to work with Cloudera Data Engineering to stream data from Kafka.  It's quite interesting how I was able to deploy code without much worry about how to configure the back end components.  

 

Demonstration

This demo will pull from the Twitter API using NiFi, write to payload to a Kafka topic named "twitter".  Spark Streaming on Cloudera Data Engineering Experience CDE will pull from the twitter topic, extract the text field from the payload (which is the tweet itself) and write back to another Kafka topic named "tweet"

 

The following is an example of a twitter payload.  The objective is to extract only the text field:

2020-08-28_11-27-32.jpg

What is Cloudera Data Engineering?

Cloudera Data Engineering (CDE) is a serverless service for Cloudera Data Platform that allows you to submit Spark jobs to an auto-scaling cluster. CDE enables you to spend more time on your applications, and less time on infrastructure.

How do I begin with Cloudera Data Engineering (CDE)?

Complete setup instructions here.

 

Prerequisites

  • Access to a CDE
  • Some understanding of Apache Spark
  • Access to a Kafka cluster
    • In this demo, I use Cloudera DataHub, Streamings Messaging for rapid deployment of a Kafka Cluster on AWS
  • An IDE
    • I use Intellij
    • I do provide the jar later on in this article
  • Twitter API developer Access: https://developer.twitter.com/en/portal/dashboard
  • Setting up a twitter stream
    • I use Apache NiFi deployed via Cloudera DataHub on AWS

Source Code

I posted all my source code here.

If you're not interested in building the jar, that's fine.  I’ve made the job Jar available here.

Oc t26, 2020 update - I added source code for how to connect CDE to Kafka DH available here.  Users should be able to run the code as is without need for jaas or keytab. 

Kafka Setup

This article is focused on Spark Structured Streaming with CDE.  I'll be super brief here

  • Create two Kafka topics
    • twitter
      • This topic is used to ingest the firehose data from twitter API
    • tweet
      • This topic is used post tweet extraction performed via Spark Structured streaming

NiFi Setup

This article is focused on Spark Structured Streaming with CDE.  I'll be super brief here.

Use the GetTwitter processor (which requires twitter api developer account, free) and write to the Kafka twitter topic

2020-08-28_09-52-14.jpg

 

 

 

Spark Code (Scala)

 

What does the code do?

It will pull from the source Kafka topic (twitter), extract the text value from the payload (which is the tweet itself) and write to the target topic (tweet)

CDE

  1. Assuming CDE access is available, navigate to virtual clusters->View Jobs
    2020-08-28_09-58-17.jpg
  2. Click on Create Job:
    2020-08-28_10-01-50.jpg

    Job Details

    • Name
      • Job Name
    • Spark Application File
    • Main Class
      •  com.cloudera.examples.KafkaStreamExample
    • Arguments
      • arg1
        • Source Kafka topic: twitter
      • arg2
        • Target Kafka topic: tweet
      • arg3
        • Kafka brokers: kafka1:9092,kafka2:9092,kafka3:9092

    2020-08-28_10-05-26.jpg

  3. From here jobs can be created and run or simply created.  Click on Create and Run to view the job run:2020-08-28_11-19-39.jpg
  4. To view the metrics about the streaming:2020-08-28_11-20-06.jpg

At this point, only the text (tweet) from the twitter payload is being written to the tweet Kafka topic.

 

That's it! You now have a spark structure stream running on CDE fully autoscaled.  Enjoy

10,367 Views