Member since
03-16-2016
707
Posts
1753
Kudos Received
203
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
5129 | 09-21-2018 09:54 PM | |
6495 | 03-31-2018 03:59 AM | |
1969 | 03-31-2018 03:55 AM | |
2179 | 03-31-2018 03:31 AM | |
4833 | 03-27-2018 03:46 PM |
08-26-2016
07:07 PM
3 Kudos
@AravinthRaj Soundararajan As part of the import, you need to determine data type mappings. There are a number of data types that will not get implicitly converted to Hive. As @Satish Bomma suggested, you need to explicitly help sqoop to use a target data type.
... View more
08-26-2016
06:50 PM
@Arun A K Before bringing new user records from LDAP to Ambari, did you have an Ambari user Arun?
... View more
08-26-2016
06:04 PM
@Sooraj Antony @bpreachuk response is appropriate. Please accept or vote to appreciate the effort.
... View more
08-26-2016
05:59 PM
3 Kudos
@Sooraj Antony m is used for minute M is used for month In your case should be YYYY-MM-DD hh:mm:ss or yyyy-MM-dd hh:mm:ss select UNIX_TIMESTAMP('2000-01-01 10:20:30','yyyy-MM-dd hh:mm:ss');
... View more
08-26-2016
04:36 PM
@Vasilis Vagias Let me know if you still have issues.
... View more
08-26-2016
04:20 PM
3 Kudos
@John Swartz, @Mark Herring , @Sagar Shimpi http://hortonworks.com/blog/how-to-size-your-hadoop-cluster page mentions a link to size-o-tron: http://hortonworks.com/resources/cluster-sizing-guide/ That's what John is looking for, and not only him. With the new redo of the portal, another page that supposed one to download a PDF is broken: http://info.hortonworks.com/SizingGuide.html Until then, the best bet is http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.2/bk_cluster-planning-guide/content/index.html. That is for HDP 2.4.2, but similar can be found for other versions.
... View more
08-26-2016
03:23 PM
@Sunile Manjee Just the cluster planning guide link works. That is anyway, the official guide maintained up-to-date.
... View more
08-26-2016
03:13 PM
5 Kudos
@Vasilis Vagias That, but VirtualBox is smart enough to import even a VMWare image. My opinion is that you downloaded an incomplete file due to your Chrome Download Manager, not capable to resume download in case of temporary connectivity issues. What is the size of your file? It should be around 9.5 GB. Try to use a download manager that is capable to resume download and do a size check at the end. Try folx, that if you are on mac.
... View more
08-26-2016
03:08 PM
5 Kudos
@Sami Ahmad The output you posted shows that 3.6 GB is used and available space is 150.8 GB. You are concerned that available should show 240.5 GB - 3.6 GB = 236.9 GB and not 150.8 GB. Here is how it goes. HDFS has blocks and each block has a size, let's assume 128 MB/block.If you have multiple small files they will underuse the block size. For example, if you have 10 files of 64 MB each stored in 10 blocks of 128 MB they will underuse the blocks at 50%. The remaining space in those blocks CANNOT be used by other files. It is just wasted and it is not reported as AVAILABLE. The way hdfs dfs -df -h command works for AVAILABLE is this: it determines the number of blocks available for storing new data (empty) and multiplies that with the block size. Just looking at your numbers above, it shows that the wasted space is 236.9-150.8=86.1 GB. That shows that your block size is set to a value higher than your average file size, about 50%. This is not uncommon, but be aware. I hope this explanation is good enough. Please don't forget to vote/accept answer that answered your question.
... View more