Member since
01-19-2017
3679
Posts
632
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 928 | 06-04-2025 11:36 PM | |
| 1535 | 03-23-2025 05:23 AM | |
| 761 | 03-17-2025 10:18 AM | |
| 2738 | 03-05-2025 01:34 PM | |
| 1810 | 03-03-2025 01:09 PM |
02-02-2019
01:21 AM
@Harry Li How many data nodes do you have in your cluster? Can you try to isolate the culprit $ hdfs dfs -du -h / If you enabled snapshots then that could be one reason can you check its existence? $ hdfs lsSnapshottableDir HTH
... View more
02-02-2019
12:12 AM
@Harry Li When a file is deleted by a user or an application, it is not immediately removed from HDFS. Instead, HDFS first renames it to a file in the /trash directory. The file can be restored quickly as long as it remains in /trash. The retention time in the /trash is configurable. After the expiry of its life in /trash, the NameNode deletes the file from the HDFS namespace. The deletion of a file causes the blocks associated with the file to be freed. Note that there could be an appreciable time delay between the time a file is deleted by a user and the time of the corresponding increase in free space in HDFS. If you want to change the default setting then it needs to be updated in the core-site properties, which you can find in the Ambari menu. Simply follow this path; from the Ambari Dashboard, click HDFS -> Configs -> Advanced -> Advanced core-site. Then set the 'fs.trash.interval' to 0 to disable. This will require a restart of the related components to pick up the changes. Check the HDFS structure to see where the most data is held. This will give you the space on each data node $ hdfs dfsadmin -report Breakdown of the HDFS across the cluster and each of the data nodes run the below command, you should give it some time to complete. $ hdfs dfs -expunge By default, HDFS uses trash. You can bypass this with rm -skipTrash or just delete the trash with The other option when cleaning up your data use the -skipTrash flag: $ hdfs dfs -rm -R -skipTrash /folder-path HTH
... View more
02-01-2019
10:24 PM
@Dukool SHarma Yes ,you can when executing from the command line by adding the -Dmapreduce parameter see below bin/hadoop jar -Dmapreduce.job.maps=5 yourapp.jar ... HTH
... View more
02-01-2019
10:10 PM
1 Kudo
@Siva A Sqoop will by default import NULL values as string null. Hive is, however, using string \N to denote NULL values and therefore predicates dealing with NULL (like IS NULL) will not work correctly. You should append parameters --null-string and --null-non-string in case of import job or --input-null-string and --input-null-non-string in case of an export job if you wish to properly preserve NULL values. Because sqoop is using those parameters in generated code, you need to properly escape value \N to \N: HTH
... View more
02-01-2019
05:07 PM
@Jean-François Vandemoortele Thanks for reverting to @kanna k question, I would just like to know your initial problem org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken): token (HDFS_DELEGATION_TOKEN token 7140339for xxxxxx) can't be found in cache was it resolved with the jaas.conf file because that was specific to the Kerberos error above.
... View more
02-01-2019
03:30 PM
@Sandeep Nemuri I think we responded at almost the same time, when some is clicking submit, there is no logic that checks whether a similar answer has already been give 🙂 Maybe you should have added that he needs to run the script as Atlas admin user as illustrated which he wasn't aware of 🙂
... View more
02-01-2019
03:24 PM
@Yee Zee What version of HDP? Please give a much info as position so that a third party can understand your exact situation.
... View more
02-01-2019
03:21 PM
@kanna k Unfortunately, I never got feedback from Jean-François but I guess it worked, That's why it's important to close a thread and share the solution. Maybe you open a new thread and tag me, what is your exact problem is it a kerberized cluster? Please let me know
... View more
01-30-2019
10:55 AM
@Ali Erdem Any updates on this thread?
... View more
01-29-2019
12:02 PM
@Ali Erdem YES it's possible to connect and run a sqoop job against an SQL server without a password. Hadoop credential provider API the CredentialProvider API in Hadoop allows for the separation of applications and how they store their required passwords/secrets. With Sqoop 1.4.5 or higher, the credential API keystore is supported by Sqoop. The AD user ONLY needs to include the -Dhadoop.security.crendential.provider.path in the sqoop command. Here are the steps, The API expects the password .jceks file to be in HDFS and accessible to that user preferably in his/her home directory Assumption password for Production sqlserver it's good to standardize eg sql_prod,sql_dev or ora_prod,ora_dev etc $ hadoop credential create sql_prod.password -provider jceks://hdfs/user/erdem/sql_prod.password.jceks The above command will prompt for the target database password see output below Enter password: {the_target_database_password}
Enter password again: {the_target_database_password}ora_prod.password
has been successfully created.org.apache.hadoop.security.alias.JavaKeyStoreProvider
has been updated. Now the password should be in your home directory,the file should be readable $ hdfs dfs -ls /user/erdem
Found 1 items
-rwx------ 3 erdem erdem 502 2019-01-29 11:08 /user/erdem/sql_prod.password.jceks Now the user erdem can run a sqoop job sqoop import
-Dhadoop.security.crendential.provider.path jceks//hdfs/user/erdem/sql_prod.password.jceks
-Doraoop.timestamp.string=false -Dmapreduce.job.user.classpath.first=true \
--verbose --connect jdbc:sqlserver://sqlserver-name \
--username erdem \
--password alias ora_prod.password \
--driver com.microsoft.sqlserver.jdbc.SQLServerDriver \
--table test \
--target-dir "{some_dir}" \
--split-by NOOBJETRISQUECONTRAT --direct --as-parquetfile In the above, I modified the output from my oracle sqoop output especially for the driver part. But it should work without issue you will realise the user erdem didn't key in a password on the CLI a security loophole. There you go revert if you need more help.
... View more