Member since
02-16-2016
89
Posts
24
Kudos Received
10
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
8401 | 05-14-2018 01:54 PM | |
938 | 05-08-2018 05:07 PM | |
481 | 05-08-2018 04:46 PM | |
1902 | 02-13-2018 08:53 PM | |
1964 | 11-09-2017 04:24 PM |
05-08-2018
05:07 PM
A "multi node HDP cluster in Azure VM" can be created using normal Ambari managed HDP installation guide, this is not much different from setting a cluster on on-premise hardware or VMs on your desktop: https://docs.hortonworks.com/HDPDocuments/Ambari-2.6.1.5/bk_ambari-installation/content/ch_Getting_Ready.html Prerequisite will be to set up Azure VMs with storage and networking. You only need to pay Hortonworks for support, if needed.
... View more
05-08-2018
05:00 PM
Try: SELECT ARRAY_LENGTH( REGEXP_SPLIT(line, ' ')) FROM twitter_feed; This is a common SQL challenge. Count will be incorrect if you have more than one spaces between words.
... View more
05-08-2018
04:46 PM
1 Kudo
Enable LDAP for Ambari: https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.4/bk_security/content/_synchronizing_ldap_users_and_groups.html Use Ambari View > Files to access HDFS. If you need access from command line, you will need to enable LDAP on operating system through ldap/sssd etc.
... View more
05-08-2018
03:23 PM
If you deployed Knox after enabling Kerberos, you will need to recreate keytabs from Ambari > Security Ranger will only sync from one provider at a time UNIX or LDAP/AD. When you switch to LDAP/AD, new users in UNIX will not be synced. So if you added Knox after sync, knox user will not be synced. The second part of your question, enabling kerberos does not, just, enable users to access cluster resources. This process actually creates principals and keytabs for all HDP service accounts. This is totally separate from user access. If you need just user access with no security you can do with just LDAP, no kerberos (not recommended on production systems).
... View more
05-08-2018
02:17 PM
As mentioned before, make sure your Hive Service is connected to the new location of MySQL. Removing MySQL without will result in losing all you metadata. You can use Ambari REST API to remove services manually: curl -u admin:admin -H "X-Requested-By: ambari" -X DELETE http: //AMBARI_SERVER_HOST:8080/api/v1/clusters/c1/services/SERVICENAME https://cwiki.apache.org/confluence/display/AMBARI/Using+APIs+to+delete+a+service+or+all+host+components+on+a+host
... View more
04-05-2018
12:27 AM
1 Kudo
Pierre solution is correct. If you installed Atlas after Ranger UserSync has been configured to use LDAP, new local users will not get synced in ranger like atlas. This user is needed to setup hbase tables. To fix, revert UserSync to UNIX, restart only Ranger UserSync, Switch back to UserSync LDAP config. In Ranger add user atlas to HBase all policy. Restart Atlas.
... View more
02-13-2018
09:11 PM
Usually by default the ticket expires 24 hours and cache expires 7 days. Depends on your directory services policies. Within 7 days you can do kinit -R for users. klist will show ticket and cache expiry time. Or you can use keytabs to automate ticket renewal. You don't have to kinit for hadoop services (ever), ticket renewal is managed automatically.
... View more
02-13-2018
09:04 PM
Check your file-provider property in Nifi. Should be set to kerberos if you are using kerberos to authenticate user: https://nifi.apache.org/docs/nifi-docs/html/administration-guide.html#kerberos_login_identity_provider
... View more
02-13-2018
08:53 PM
1 Kudo
We usually use Nifi Content and Provenance repository to troubleshoot failed flow files. Both are set to 7 days of retention. Plus you can replay content to debug. From zero API reporting perspective you can use row counts as attributes collecting total rows, success rows, failed rows etc. Depending on format of your source file this can be as simple as executing wc -l. Later converting these attributes to JSON and use MergeContent, Schema registry and query record to create an email report.
... View more
02-13-2018
08:08 PM
Best place to start will be to check YARN tuning: https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.4/bk_command-line-installation/content/determine-hdp-memory-config.html There can be number of reasons why your jobs might be running slow like what is the current load on cluster, how is capacity scheduling set up like priority, queues, resources etc. You can check Hive execution plan to see bottlenecks.
... View more
02-13-2018
07:40 PM
Similar issue: https://community.hortonworks.com/questions/86429/cannot-connect-to-zookeeper-server-from-zookeeper.html Problem is your zookeeper address which is pointing to localhost:2181, the worker/slave node does not have a local zookeeper running hence the "Connection refused". Check link above for detailed solution.
... View more
11-09-2017
04:24 PM
1 Kudo
Template attached. I am using Nifi 1.3.0. In this template if a file is not found in GetFile processor every 10 seconds an event will be generated. You can put an email processor in front of Monitor Activity to send mail. You can also set custom subject headers in Monitoring processor (also in the template). nifimonitoring.xml
... View more
11-09-2017
04:05 PM
Please provide more information like job you are trying to run, services you using for import and specific log messages.
... View more
11-07-2017
07:47 PM
I don't think it is your ConvertAvroToJSON processor failing it is Nifi itself. Check bootstrap logs and nifi-app.log for any suspicious messages. Maybe memory or disk issues.
... View more
11-07-2017
07:39 PM
2 Kudos
Put a MonitorActivity processor after your GetFile processor. Set the threshold to your interval. If a file is not found in that interval MonitorActivity will be triggered. https://nifi.apache.org/docs/nifi-docs/components/org.apache.nifi/nifi-standard-nar/1.4.0/org.apache.nifi.processors.standard.MonitorActivity/index.html
... View more
11-07-2017
07:30 PM
Please verify your source files for any new lines at the end. Best will be to open the queue and copy the content to a notepad, sometime using Nifi web reader is not ideal. Alternatively you can use a ReplaceText processor to strip out any new lines.
... View more
11-07-2017
07:25 PM
1 Kudo
Try stopping Nifi and purging everything within your provenance repository then start Nifi. Check nifi-app.log file for any provenance related events. Check if the user running Nifi process has access to read/write in set directory. I had a similar issue today instead my provenance implementation was set to Volatile, which I changed to WriteAhead. Also note, by default implementation is PersistentProvenanceRepository and if you have been changing implementations back and forth you will need to delete provenance data. (WriteAhead can read PersistentProvenanceRepository but not other way around).
... View more
08-16-2017
08:18 PM
This is a bug in Ambari 2.5.1 [https://issues.apache.org/jira/browse/AMBARI-21473] Resolution: Remove /etc/zeppelin/conf/interpreter.json file and restart Zeppelin service
... View more
05-05-2017
08:57 PM
This is not a Hive issue rather a file system or file encoding issue. SELECT * in Hive actually does nothing except read the file from file system. So if you run a hadoop fs cat on your underlying file, you should see the same behavior. You can check file encoding on bash as $ file -i filename You can change the encoding using iconv. And convert to utf-8 which is printable encoding. iconv -f current_encoding -t new_encoding input.file -o out.file
... View more
05-05-2017
07:11 PM
1 Kudo
Trying to give closure to this topic: This is a very misleading situation where it seems Nifi is not running but it actually is. And you get ERR_CONNECTION_CLOSED or 'can't display page' etc. error when you try to hit the secure url. Reason: Nifi is configured to authenticate client for connection and since there is not authentication provided the UI errors out. Resolution: Import client certificate in your browser. This certificate needs to be trusted by the same authority as Nifi. Once imported, close and re-open the browser (refreshing will not work). Given you have imported the certificates, you should be prompted to select a certificate to login.
... View more
05-04-2017
05:45 PM
Please provide information on how you are generating and defining your keytabs. try klist -k nifi-1-service-keytab If you principals have HOST (machine name or IP) as part of the definition like xxxx/HOST_NAME@domain you will not be able to use the keytab on any other machine. Renaming the keytab will not work as content of the file still point to a specific host. It is best practice to have separate keytab for separate machines. Reusing the same keytab is not the most secure option. Alternatively, if you define a principal in AD as headless that is without HOST attribute. And then create a keytab, that keytab can be used on any host typically this is your hdfs principal. But, not too secure.
... View more
03-27-2017
09:20 PM
Try something like: hadoop fs -cat /path_to_hdfs_file/test.csv | head -c 40000000
... View more
03-15-2017
03:55 PM
1 Kudo
Please verify what have you configured for "Maximum-Value Column" in the QueryDatabaseTable processor. Plus how is your table structured. Nifi needs an increment key to determine the maximum id previously fetched. Following article will help" https://community.hortonworks.com/articles/51902/incremental-fetch-in-nifi-with-querydatabasetable.html
... View more
03-15-2017
03:48 PM
Start by look into the container sizing and queues configuration. Seems the container serving JDBC is either not getting enough resource or is hanging. Check your yarn logs for the application id running JDBC (thrift) process.
... View more
03-14-2017
03:32 PM
I would strongly recommend you reconsider your usecase to import a table from SQL Server to HBase. HBase is not a relation database, and most of the practice applied on relational database will not work or degrade HBase performance. Consider Hive or similar technologies for SQL server or relation DB offload. Example, if you don't know your row key prefix in HBase you will end up doing a full table scan, which is an expensive operation in HBase. So designing rowkey is the most important step in HBase unlike relational DBs. More information here and why: https://hbase.apache.org/book.html#rowkey.design
... View more
03-14-2017
03:21 PM
You can use Multi-tenant tables to create user specific connections. More detail here: https://phoenix.apache.org/multi-tenancy.html
... View more
03-14-2017
03:17 PM
see fit Deem appropriate, as in He's entitled to divide up his property as he sees fit, or If we see fit to attend, we'll be there. This expression uses see in the sense of "view as," a usage dating from about 1325. http://idioms.thefreedictionary.com/see+fit
... View more
03-14-2017
03:13 PM
Before going into compatibility, from your logs it seems your /hadoop directory have different permissions than what you expect/ java.security.AccessControlException: Permission denied: user=hadoop, path="file:/hadoop":root:root:drwxr-xr-x /hadoop is owned by root:root with write permission only to root. No other user will be able to write to this directory. Change ownership to root:hadoop with permission 775. Hope it helps.
... View more
03-14-2017
03:05 PM
Please provide more information like: 1. Ranger version 2. Have you enabled audit logs in Ranger configuration 3 Where are you storing audit logs; Db, Solr, HDFS or log4j Depending on the version of Ranger please follow the documentation below to troubleshoot your configs: https://cwiki.apache.org/confluence/display/RANGER/Ranger+0.5+Audit+Configuration
... View more
03-02-2017
05:27 PM
2 Kudos
The patch to create Hive external tables from Sqoop is still unresolved: https://issues.apache.org/jira/browse/SQOOP-816 Unfortunately you will not be able to pull updates from source tables using only id column. You will need a timestamped (last modified) column for Sqoop to know which rows were updated. So the best practice is rather at your database side where it is always best to keep columns like 'modified', 'modified by' in your tables. https://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_incremental_imports
... View more
- « Previous
-
- 1
- 2
- Next »