Member since
11-18-2017
4
Posts
0
Kudos Received
0
Solutions
09-16-2023
08:51 AM
Hi, You need to set in hive-site.xml, these three tags to get this working with hive: <property>
<name>google.cloud.auth.service.account.json.keyfile</name>
<value>/home/hadoop/keyfile.json</value>
</property>
<property>
<name>fs.gs.reported.permissions</name>
<value>777</value>
</property>
<property>
<name>fs.gs.path.encoding</name>
<value>/home/hadoop/</value>
</property> Same xml tags we can have it on hadoop in core-site.xml to have it working with hdfs, On beeline, just execute this and it shall work: INSERT OVERWRITE DIRECTORY 'gs://bucket/table' ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' SELECT * FROM table; Please upvote if you found helpful!
... View more
05-02-2018
08:55 PM
I am getting below error,if I am using "key" as column. mysql> create table session(key char(16) not null,data blob,expiry int(11) unsigned not null,primary key (key)) ENGINE=MyISAM default charset=utf8;
ERROR 1064 (42000): You have an error in your SQL syntax; check the manual that corresponds to your MySQL server version for the right syntax to use near 'char(16) not null,data blob,expiry int(11) unsigned not null,primary key (key)) ' at line 1 By replacing "key" to "key1" it works fine. mysql> create table session(key1 char(16) not null,data blob,expiry int(11) unsigned not null,constraint pk_example primary key (key1)) ENGINE=MyISAM default charset=utf8;
Query OK, 0 rows affected (0.00 sec)
... View more