Member since
01-18-2016
169
Posts
32
Kudos Received
21
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1627 | 06-27-2025 06:00 AM | |
| 1340 | 01-14-2025 06:30 PM | |
| 1861 | 04-06-2018 09:24 PM | |
| 2009 | 05-02-2017 10:43 PM | |
| 5202 | 01-24-2017 08:21 PM |
11-17-2016
03:09 PM
Awesome. If it continues to happen you'll need to figure out why you're getting OOM (assuming that was what was happening). Intermittent exceptions are often a symptom of OOM exceptions. Solr loves memory but there are a lot of factors that can contribute to it. Sometimes giving the JVM more memory is the solution, but not always. Good luck.
... View more
11-16-2016
10:48 PM
@Wing Lo - The solution may depend on what is actually wrong. It may be that the node is just out of memory. If that's the case, a restart may resolve the issue (but it could occur again). If the index data is actually corrupt, you can take the bad node offline and the errors will stop. However, if the index is actually corrupt, you will need to fix/replace the bad data. I have not used this technique, but you might look at this https://support.lucidworks.com/hc/en-us/articles/202091128-How-to-deal-with-Index-Corruption
... View more
11-16-2016
10:28 PM
@Daniel Scheiner To add to the point, I believe that in the future HDP and HDF will be able to use the same Ambari host, but it is not possible currently as Constantin said. This blog post says, "Currently, nodes can not be shared between HDP and HDF. Completely separate clusters (each with its own Ambari and Ranger) are required at this point."
... View more
11-16-2016
02:12 PM
@Prem Kripalani I'm glad yo got it worked out. Sorry to hear that it was such a pain.
... View more
11-16-2016
03:25 AM
Awesome. Glad you got it!
... View more
11-16-2016
03:23 AM
@Prem Kripalani That is weird. I'm looking at code for nifi 1.0.0 and line 82 does not have JAVAHOME anywhere near it. In fact, I don't see JAVAHOME anywhere. Typically that variable has an underscore "_" like JAVA_HOME. The error message "syntax error near unexpected token `newline'" is associated with an invalid redirect ">" with a newline after it. Can you find the file nifi.sh and look around line 82 to see if you see anything about JAVAHOME or if you see a redirect with no filename after it? This could be an empty variable after the ">" as well. In any case, it may be easier to reinstall Nifi than to track it down.
... View more
11-16-2016
02:22 AM
Can yo check if Postgres is running? [root@sandbox ~]# service postgresql status
postmaster (pid 524) is running...
... View more
11-16-2016
02:09 AM
Apparently there is a newer option "-Dorg.apache.sqoop.splitter.allow_text_splitter=true" for splitting on string there is no guarantee it will split evenly.
... View more
11-16-2016
01:56 AM
It looks like you're right about --direct-split-size being limited to postgresql. So, I think you'll need to depend on the number of mappers and the proper split-by column. What size files are you getting when you export with 4 mappers? I noticed that in your original post, you have --split-by STR, which makes me think you're trying to split by a string column which is not possible and I think it will produce an error with more than one mapper. But, let's assume you use a split-by column is actually numeric. Do you know for sure that your data is evenly distributed by the split-by column? If the numeric field is not evenly distributed, you will end up with some larger and some smaller files. Sqoop first does select min(<split-coumn>), max(split-column) from <table>. It then divides it by the number of mappers. For example, suppose you have 25 records in the database. The primary key field "id" has 24 records with id's 1 through 24, plus one other record with id=100. If, we run 4 mappers, we get min 1, max 100, divided into 4 groups by id ranges for the 4 mappers. Each mapper writes the records in its ID range. So we will end up with ONE file containing 24 records, three empty files and the one with id = 100.
... View more