Member since
01-19-2017
3679
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1001 | 06-04-2025 11:36 PM | |
| 1568 | 03-23-2025 05:23 AM | |
| 784 | 03-17-2025 10:18 AM | |
| 2819 | 03-05-2025 01:34 PM | |
| 1861 | 03-03-2025 01:09 PM |
12-26-2019
02:10 AM
@hud When you run nifi as a microservice then you can configure PVC [Persistent Volume Claims] using helm in AKS or Kubernetes which will ensure that evenif the nifi pods restarts it will always have the same volume mounted. Under the persitence configuration the parameter persistence.enabled should be set to true see Helm Chart for Apache Nifi HTH
... View more
12-25-2019
12:38 PM
@saivenkatg55 For the screenshots, the 2 notebooks you create are Untitled Note 1 and Untitled Note 2 which should appear in the drop-down list under Notebook on the top menu. Below I will create a Spark interpreter notebook named saivenkatg55 from step 2 above This should appear on the under the notebook Launched a test I can see the job was accepted and running in RM UI So where exactly are you encountering issues? Happy hadooping !
... View more
12-25-2019
09:44 AM
@kiranpune DistCp (distributed copy) is a tool used for large inter/intra-cluster copying. It uses MapReduce to effect its distribution, error handling and recovery and reporting. It expands a list of files and directories into the input to map tasks, each of which will copy a partition of the files specified in the source list that basic description. But one can use different command-line options when running DISTCP see the official dictcp documentation below are a few options for your different use cases. OPTIONS -append: Incremental copy of the file with the same name but different length -update: Overwrite if source and destination differ in size, block size, or checksum -overwrite: Overwrite destination -delete: Delete the files existing in the destination but not in the source I think you can schedule or script a daily copy
... View more
12-24-2019
01:06 AM
@saivenkatg55 I have tried to analyze your situation but with access to the Linux box it rather difficult,but I think there is a workaround. The chattr linux command makes important files IMMUTABLE (Unchangeable). The immutable bit [ +i ] can only be set by superuser (i.e root) user or a user with sudo privileges can be able to set. This will prevent the file from being deleted forcefully, renamed or change the permissions, but it won’t be allowed says 'Operation not permitted“' # ls -al /var/run/hadoop-yarn/yarn/ total 8 . .. -rw-r--r-- 1 yarn hadoop 0 Dec 24 09:34 hadoop-yarn-nodemanager.pid Set immutable bit # chattr +i hadoop-yarn-nodemanager.pid Verify the attribute with command the below command # lsattr ----i--------e-- ./hadoop-yarn-nodemanager.pid The normal ls command shows no difference # ls -al /var/run/hadoop-yarn/yarn/ total 8 drwxr-xr-x 2 root root 4096 Dec 24 09:34 . drwxr-xr-x 3 root root 4096 Dec 24 09:34 .. -rw-r--r-- 1 yarn hadoop 0 Dec 24 09:34 hadoop-yarn-nodemanager.pid Deletion protection # rm -rf /var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid rm: cannot remove ‘/var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid’: Operation not permitted Permission change protected # chmod 755 /var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid chmod: changing permissions of ‘/var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid’: Operation not permitted How to unset attribute on Files # chattr -i /var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid After resetting permissions, verify the immutable status of files using lsattr command # lsattr ---------------- ./var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid Please do that and revert
... View more
12-22-2019
10:13 AM
2 Kudos
@Prakashcit That by design: A NOVALIDATE constraint is basically a constraint that can be enabled but for which hive will not check the existing data to determine whether there might be data that currently violate the constraint. This is useful if we know there’s data that violates the constraint but we want to quickly put on a constraint to prevent further violations, with the intention to clean up any possible violations at some future point in time. It’s also potentially useful if we know the data is clean and so want to prevent the potentially significant overheads of hive having to check all the data to ensure there are indeed no violations.
... View more
12-21-2019
06:47 AM
@Uppal Best way to duplicate a partitioned table in Hive Create the new target table with the schema from the old table the describe formatted could help with the SQL Use hadoop fs -cp to copy all the partitions from source to the target table Run MSCK REPAIR TABLE table_name; on the target table HTH
... View more
12-20-2019
01:02 PM
@GrahamB No you don't need to wait for 24 hours to destroy a kerberos ticket you will need to run on the Kerberos server as the user Check valid ticketTo list all of the entries in the default credentials cache $ klist You should have some out here To delete the default credentials cache for the user $ kdestroy Then to obtain a ticket-granting ticket with a lifetime of 10 hours, which is renewable for five days, type: $ kinit -l 10h -r 5d your_principal HTH
... View more
12-20-2019
03:13 AM
@saivenkatg55 The file permission should be 644 not 444 # chmod 644 /var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid Do that and revert please
... View more
12-19-2019
02:12 PM
@saivenkatg55 This "Exiting with status 1: java.io.IOException: Problem starting http server" error should be linked to your other question I just have responded to https://community.cloudera.com/t5/Support-Questions/Unable-to-start-the-node-manager/td-p/286013 If this is resolved then the java.io.IOException shouldn't occur HTH
... View more
12-19-2019
01:55 PM
@saivenkatg55 I think there is a permission issue with the pid file Can you check the permissions, if for any reason the are not as shown in the screenshot please run the chown as root to rectify that # chown yarn:hadoop /var/run/hadoop-yarn/yarn/hadoop-yarn-nodemanager.pid Do that for all files in the directory whose permissions are not correct. HTH
... View more