Member since
04-11-2016
535
Posts
148
Kudos Received
77
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
7497 | 09-17-2018 06:33 AM | |
1827 | 08-29-2018 07:48 AM | |
2744 | 08-28-2018 12:38 PM | |
2125 | 08-03-2018 05:42 AM | |
1991 | 07-27-2018 04:00 PM |
12-29-2016
10:17 AM
3 Kudos
The table definition "LINES TERMINATED BY" only supports newline '\n' right now. This is a known issue and Jira Hive 11996 has already been raised for the issue:
To handle the newline characters within the data, you can use the Omniture Data format which uses a EscapedLineReader which gets around Omniture's pesky escaped tabs and newlines.
Please note that the data files need to include '\' characters before the newline character within the data and run the below command in sequence and required jars are attached along with data file: add jar /tmp/omnituredata-1.0.2-SNAPSHOT-jar-with-dependencies.jar;
add jar /tmp/omnituredata-1.0.2-SNAPSHOT-javadoc.jar;
add jar /tmp/omnituredata-1.0.2-SNAPSHOT-sources.jar;
add jar /tmp/omnituredata-1.0.2-SNAPSHOT.jar;
(Note: jars are available on the HDFS /tmp folder). CREATE TABLE test8(id string,desc string)
ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
STORED AS INPUTFORMAT 'org.rassee.omniture.hadoop.mapred.OmnitureDataFileInputFormat'
OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat'
LOCATION '/apps/hive/warehouse/test8';
Sample file under HDFS location '/apps/hive/warehouse/test8' is as: [hive@sindhu root]$ hdfs dfs -cat /apps/hive/warehouse/test8/file.txt
id desc
1 Hi\
I am a member and would like to open savings accts for both my kids aged 12 and 16.\
Is that possible and what documents do I need to bring?\
Also do I need to make an appt first?\
Thx!
Also, the inputformat as TEXT does not understand the escaped newline characters. 4 rows selected (0.165 seconds)
0: jdbc:hive2://sindhu:2181/> CREATE TABLE test9(id string,desc string)
0: jdbc:hive2://sindhu:2181/> ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'
0: jdbc:hive2://sindhu:2181/> STORED AS textfile LOCATION '/apps/hive/warehouse/test9';
No rows affected (0.209 seconds)
0: jdbc:hive2://sindhu:2181/> select * from test9;
+---------------------------------------------------------------------------------------+-------------------+--+
| test9.id | test9.desc |
+---------------------------------------------------------------------------------------+-------------------+--+
| id | desc |
| 1 | Hi\ |
| I am a member and would like to open savings accts for both my kids aged 12 and 16.\ | NULL |
| Is that possible and what documents do I need to bring?\ | NULL |
| Also do I need to make an appt first?\ | NULL |
| Thx! | NULL |
| 2 | hi jihidp\ |
| uiunoo! | NULL |
| 3 | hi who are you\ |
| talking with | NULL |
+--------------------------+
... View more
Labels:
11-22-2017
05:34 AM
Hi, This issue was resolved by the settings as follows: hadoop.proxyuser.root.hosts=* You can also see the answer on the below comment. https://community.hortonworks.com/comments/144449/view.html
... View more
11-09-2016
05:01 AM
@swathi thukkaraju Try using the option --password-file to remove the possibility of entering / exposing the password. Below is the link for creating the password file (link)
... View more
10-31-2018
11:53 AM
Hi all, I have a similar issue, all 3 node managers in my cluster are marked as bad with local dirs bad alerts and I can assure I have plenty of space on the YARN disk. (just 35% of the disk is used). Does anyone know of this alert other than YARN reaching its disk threshold value?? Thanks, Priyanshu
... View more
08-30-2016
07:17 AM
This solution is works for me. Thanks for the immediate reply.
... View more
07-27-2016
07:13 AM
@Sindhu I tried using sqoop import command and it gives me the below error.
... View more
07-28-2016
09:17 AM
Is the issue resolved?
... View more
07-26-2016
06:41 PM
@Prakash Punj Good lesson learned for all of us. It is a good practice to save a file with a name like hue.ini.old for anything unix. Hue is not the only that reads all .ini files from a directory.
... View more
07-07-2016
05:37 AM
@Sindhu The code snippet didn't work see [root@sandbox ~]# su - hdfs [hdfs@sandbox ~]$ hdfs dfs -chown -R sqoop:hdfs /user/root
chown: `/user/root': No such file or directory So what I did I just run the sqoop command as hdfs and it run successfully [hdfs@sandbox ~]$ sudo sqoop import --connect jdbc:oracle:thin:@192.168.0.15:1521/PROD --username sqoop -P --table DEPT_INFO
We trust you have received the usual lecture from the local System
Administrator. It usually boils down to these three things:
#1) Respect the privacy of others.
#2) Think before you type.
#3) With great power comes great responsibility.
[sudo] password for hdfs:
[hdfs@sandbox ~]$ sqoop import --connect jdbc:oracle:thin:@192.168.0.15:1521/PROD --username sqoop -P --table DEPT_INFO
Warning: /usr/hdp/2.3.2.0-2950/accumulo does not exist! Accumulo imports will fail.
Please set $ACCUMULO_HOME to the root of your Accumulo installation.
16/07/07 05:25:32 INFO sqoop.Sqoop: Running Sqoop version: 1.4.6.2.3.2.0-2950
Enter password:
16/07/07 05:25:39 INFO oracle.OraOopManagerFactory: Data Connector for Oracle and Hadoop is disabled.
16/07/07 05:25:39 INFO manager.SqlManager: Using default fetchSize of 1000
16/07/07 05:25:39 INFO tool.CodeGenTool: Beginning code generation
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/2.3.2.0-2950/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/2.3.2.0-2950/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
16/07/07 05:25:52 INFO manager.OracleManager: Time zone has been set to GMT
16/07/07 05:25:53 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM DEPT_INFO t WHERE 1=0
16/07/07 05:25:53 INFO orm.CompilationManager: HADOOP_MAPRED_HOME is /usr/hdp/2.3.2.0-2950/hadoop-mapreduce
Note: /tmp/sqoop-hdfs/compile/075dc9427b098234773ffaadf17b1b5f/DEPT_INFO.java uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
16/07/07 05:25:57 INFO orm.CompilationManager: Writing jar file: /tmp/sqoop-hdfs/compile/075dc9427b098234773ffaadf17b1b5f/DEPT_INFO.jar
16/07/07 05:25:57 INFO manager.OracleManager: Time zone has been set to GMT
16/07/07 05:25:57 INFO manager.OracleManager: Time zone has been set to GMT
16/07/07 05:25:57 INFO mapreduce.ImportJobBase: Beginning import of DEPT_INFO
16/07/07 05:25:58 INFO manager.OracleManager: Time zone has been set to GMT
16/07/07 05:26:00 INFO impl.TimelineClientImpl: Timeline service address: http://sandbox.hortonworks.com:8188/ws/v1/timeline/
16/07/07 05:26:01 INFO client.RMProxy: Connecting to ResourceManager at sandbox.hortonworks.com/192.168.0.104:8050
16/07/07 05:26:06 INFO db.DBInputFormat: Using read commited transaction isolation
16/07/07 05:26:06 INFO db.DataDrivenDBInputFormat: BoundingValsQuery: SELECT MIN(DEPT_ID), MAX(DEPT_ID) FROM DEPT_INFO
16/07/07 05:26:06 INFO mapreduce.JobSubmitter: number of splits:4
16/07/07 05:26:06 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1467711992449_0004
16/07/07 05:26:07 INFO impl.YarnClientImpl: Submitted application application_1467711992449_0004
16/07/07 05:26:07 INFO mapreduce.Job: The url to track the job: http://sandbox.hortonworks.com:8088/proxy/application_1467711992449_0004/
16/07/07 05:26:07 INFO mapreduce.Job: Running job: job_1467711992449_0004
16/07/07 05:26:32 INFO mapreduce.Job: Job job_1467711992449_0004 running in uber mode : false
16/07/07 05:26:32 INFO mapreduce.Job: map 0% reduce 0%
16/07/07 05:26:55 INFO mapreduce.Job: map 25% reduce 0%
16/07/07 05:27:22 INFO mapreduce.Job: map 50% reduce 0%
16/07/07 05:27:25 INFO mapreduce.Job: map 100% reduce 0%
16/07/07 05:27:26 INFO mapreduce.Job: Job job_1467711992449_0004 completed successfully
16/07/07 05:27:27 INFO mapreduce.Job: Counters: 30
File System Counters
FILE: Number of bytes read=0
FILE: Number of bytes written=584380
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=417
HDFS: Number of bytes written=95
HDFS: Number of read operations=16
HDFS: Number of large read operations=0
HDFS: Number of write operations=8
Job Counters
Launched map tasks=4
Other local map tasks=4
Total time spent by all maps in occupied slots (ms)=164533
Total time spent by all reduces in occupied slots (ms)=0
Total time spent by all map tasks (ms)=164533
Total vcore-seconds taken by all map tasks=164533
Total megabyte-seconds taken by all map tasks=41133250
Map-Reduce Framework
Map input records=9
Map output records=9
Input split bytes=417
Spilled Records=0
Failed Shuffles=0
Merged Map outputs=0
GC time elapsed (ms)=871
CPU time spent (ms)=10660
Physical memory (bytes) snapshot=659173376
Virtual memory (bytes) snapshot=3322916864
Total committed heap usage (bytes)=533200896
File Input Format Counters
Bytes Read=0
File Output Format Counters
Bytes Written=95
16/07/07 05:27:27 INFO mapreduce.ImportJobBase: Transferred 95 bytes in 87.9458 seconds (1.0802 bytes/sec)
16/07/07 05:27:27 INFO mapreduce.ImportJobBase: Retrieved 9 records.
... View more
- « Previous
- Next »