Member since
08-14-2014
13
Posts
4
Kudos Received
0
Solutions
03-13-2019
08:59 PM
1 Kudo
Hello again, I commented to you that I found something additional, apparently the configuration "hive.strict.managed.tables" is in two configuration blocks and the value of "True" must be changed to "False" in both. I attached an image regards
... View more
03-01-2019
05:15 PM
Hello, I told you that I had the same problem and make the following changes to my hive configuration. I'm still checking the performance of hive by changing these settings, but at least it lets me create tables. regards
... View more
10-03-2017
04:21 PM
Greetings to all, Please maybe someone knows some way to audit the users and scripts running on spark. What I would like to know in the audit would be: - User running the script - Script or statement that execute - Date and Time - IP - Etc. Thank you
... View more
Labels:
06-09-2016
04:46 PM
Hello, Maybe must assign the data type to count, for example: 1) COUNT (uniqueids) AS id_count: double 2) COUNT (uniqueids) AS id_count: chararray I usually go for option 2 because pig does not resolve well the double data type when you want to apply a filter. Greetings and good luck
... View more
06-08-2016
09:28 PM
Hello, Settings flume agents must do it in Ambari UI.
I made the capture of twits with flume and has worked for me in the following link: http://blog.hubacek.uk/streaming-tweets-into-hadoop-part-ii/
Do not forget that the jar that downloads cloudera should be in the path of your lib flume, for example in centos (usr/HDP/current/flume/ lib).
A step followed you set up agents in ambari UI, restart flume from the ambari and should begin to capture you twits. If you have any error you can check the log flume, for example in centos (var/log/flume). Greetings and good luck
... View more
06-06-2016
05:37 PM
Dear,
I have the following configuration flume already have used it before and it worked: (also Add the flume-sources-1.0-SNAPSHOT jar in the directory of jars flume) config-flume.png
The only thing that changed was update to version 1.5.2.2.4 flume and is giving me the following error.
ERROR [lifecycleSupervisor-1-2] (org.apache.flume.lifecycle.LifecycleSupervisor$MonitorRunnable.run:253) - Unable to start EventDrivenSourceRunner: { source:com.cloudera.flume.source.TwitterSource{name:Twitter,state:IDLE} } - Exception follows.
java.lang.NoSuchMethodError: twitter4j.FilterQuery.setIncludeEntities(Z)Ltwitter4j/FilterQuery;
at com.cloudera.flume.source.TwitterSource.start(TwitterSource.java:139)
at org.apache.flume.source.EventDrivenSourceRunner.start(EventDrivenSourceRunner.java:44)
at org.apache.flume.lifecycle.LifecycleSupervisor$MonitorRunnable.run(LifecycleSupervisor.java:251)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
Please your help if someone has the same error.
Thank you
... View more
Labels:
02-03-2016
05:50 PM
Excellent, thank you very much, I was looking for this configuration
... View more
02-03-2016
05:38 PM
Good afternoon, I wanted to know how they could have a cluster with nodes of different characteristics in RAM, for example:
slave1: 12 GB RAM slave2: 12 GB RAM Slave3: 32 GB RAM Slave4: 32 GB RAM The nodes would slave1 and slave2 a group and slave3 and slave4 nodes are another group.
The problem is the time to set the parameter of Yarn "yarn.nodemanager.resource.memory-mb" and I do not know if I set it to the first group or the second group .. ??
... View more
Labels:
02-03-2016
03:01 PM
Thank you very much, looks very interesting, I will perform a test on the cluster we have implemented, I will comment soon as worked
... View more
02-03-2016
02:39 PM
1 Kudo
Thanks for your help, the issue of incremental loads Sqoop and I work very well when you have an update on the Oracle table.
On the issue of oracle golden gate, we are verifying costs to try, if we can implement them as we would remark was, but according to articles I reviewed is very efficient for incremental loads in real time
... View more
12-11-2015
03:52 PM
2 Kudos
Hi Estimated, Please, as you might consume an incremental loading of data from Oracle to Hadoop. I will comment briefly as work currently I use Oracle materialized views to get only the data that changed last time and new records, so I avoid each time bringing the entire database. But now that we are entering in the world of BigData, I want to know if we can replace Oracle materialized views and incremental loads with a tool Hadoop as flume, storm, Kafka nifi, what would be the most appropriate
If you need additional information do not hesitate to ask Thanks in advance Angel
... View more
Labels:
08-14-2014
01:40 PM
Excelente, despues de buscar un dia entero la solución esta es la correcta
... View more
- Tags:
- me
08-14-2014
07:09 AM
when adding a new host to the cluster CDH stays in the message "Adquiriendo bloqueo de instalación"
... View more