Member since
11-17-2021
1128
Posts
256
Kudos Received
29
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2899 | 11-05-2025 10:13 AM | |
| 482 | 10-16-2025 02:45 PM | |
| 1040 | 10-06-2025 01:01 PM | |
| 821 | 09-24-2025 01:51 PM | |
| 629 | 08-04-2025 04:17 PM |
01-25-2024
08:27 AM
1 Kudo
@Rafe Welcome to the Cloudera Community! To help you get the best possible solution, I have tagged our CDP experts @venkatsambath @aakulov who may be able to assist you further. Please keep us updated on your post, and we hope you find a satisfactory solution to your query.
... View more
01-18-2024
11:05 AM
@Adhitya This is a rather old post. Can your provide details on your specific setup (processors and configurations including scheduling) used, info around your data and what you expect versus what you are seeing? Is your NiFi a cluster setup or standalone? How is source data ingested into your NiFi for this dataflow? Typically issues like this are related to dataflow design, but there is not enough info here to reproduce or make suggestions yet. Thanks, Matt
... View more
01-18-2024
09:17 AM
1 Kudo
@MPHSpeed Working with the actual data instead of sample data I build, i would recommend making these two changes: 1. In extractText processor change "Enable Unix Lines Mode" to true. 2. In each dynamic property in RouteOnAttribute, change "equals" function to "contains" function. If you found any of the suggestions/solutions provided helped you with your issue, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
01-14-2024
07:34 AM
so this is also a scam. I was contacted for work after I looked for wirk online. Work is to complete tesks daily to generate reviews.
... View more
01-13-2024
05:51 AM
Oh, I successfully managed to integrate and run NiFi 2.0 with Python on Windows using the method you suggested. Thank you so much!
... View more
01-12-2024
01:40 AM
Hello, sorry but due to the urgency I abandoned the migration of the components and finally reinstalled everything. I now have a hbase table export problem and submitted a new topic. If anyone here would take the time to help me find a solution, that would be very kind.
... View more
01-11-2024
11:50 AM
@NAMC Welcome to the Cloudera Community! To help you get the best possible solution, I have tagged our CDP experts @venkatsambath @aakulov who may be able to assist you further. Please keep us updated on your post, and we hope you find a satisfactory solution to your query.
... View more
01-10-2024
12:57 PM
@pratschavan FetchFile is typically used in conjunction with ListFile so that it only fetches the content for the FlowFile it is passed. ListFile would only list the file once. If you are using only the FetchFile processor, I am guessing you configured the "File to Fetch" property with the absolute path to you file. Using this processor in this way means that it will fetch that same file every time it is scheduled to execute via the processor's "Scheduling" tab configuration. Can you share screenshots of how you have these two processors configured? If you found any of the suggestions/solutions provided helped you with your issue, please take a moment to login and click "Accept as Solution" on one or more of them that helped. Thank you, Matt
... View more
01-09-2024
08:14 PM
@SylvainL It seems your API request has some issues, Can you share the full request you are sending to /flow-files?
... View more
01-09-2024
03:40 AM
As I was already using the Hadoop Credential Provider, I found a solution that does not require decrypting the password as follows: PySpark code: # Spark session
spark = SparkSession.builder \
.config("spark.yarn.keytab=/etc/security/keytabs/<APPLICATION_USER>.keytab") \
.appName('SPARK_TEST') \
.master("yarn") \
.getOrCreate()
credential_provider_path = 'jceks://hdfs/<PATH>/<CREDENTIAL_FILE>.jceks'
credential_name = 'PASSWORD.ALIAS'
# Hadoop credential
conf = spark.sparkContext._jsc.hadoopConfiguration()
conf.set('hadoop.security.credential.provider.path',credential_provider_path)
credential_raw = conf.getPassword(credential_name)
for i in range(credential_raw.__len__()):
password = password + str(credential_raw.__getitem__(i)) The important point above is the .config() line in SparkSession. You must enter the keytab to access the password. Otherwise you will get the encrypted value. I can't say that I'm very happy with being able to directly manipulate the password value in the code. I would like to delegate this to some component in a way that the programmer does not have direct access to the password value. Maybe what I'm looking for is some kind of authentication provider, but for now the solution above works for me.
... View more