Member since
12-09-2015
115
Posts
43
Kudos Received
12
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
8353 | 07-10-2017 09:38 PM | |
5879 | 04-10-2017 03:24 PM | |
1345 | 03-04-2017 04:08 PM | |
4683 | 02-17-2017 10:42 PM | |
6139 | 02-17-2017 10:41 PM |
12-02-2016
02:17 PM
1 Kudo
I was able to create hive table on top of json files. below is the syntax i used to create external table..so i donot have to move data, all i need is to add partition CREATE EXTERNAL
TABLE hdfs_audit( access string, agenthost string, cliip string, enforcer string, event_count bigint, event_dur_ms bigint, evttime timestamp, id string, logtype string, policy bigint, reason string, repo string, repotype bigint, requser string, restype string, resource string, result bigint, seq_num bigint) PARTITIONED BY ( evt_time string) ROW FORMAT SERDE 'org.apache.hive.hcatalog.data.JsonSerDe' STORED AS
INPUTFORMAT 'org.apache.hadoop.mapred.TextInputFormat' OUTPUTFORMAT
'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat' LOCATION 'hdfs://CLUSTERNAME/ranger/database/hdfs'; Add partition: ALTER
TABLE ranger_audit.hdfs_audit ADD PARTITION (evt_time='20160601') LOCATION
'/ranger/audit/hdfs/20160601/hdfs/20160601';
... View more
11-05-2018
06:42 PM
@Raja Sekhar Chintalapati: Can you please tell me if you find solution to this problem
... View more
01-04-2017
08:39 PM
@Sergey Soldatov or @Raja Sekhar Chintalapati Do you know if this still is the case? Is there any plans for this in the future? Thanks! It looks like the jira mentioned here is resolved
... View more
07-20-2016
06:09 PM
@Kuldeep Kulkarni great stuff. I find myself getting this confused as well.
... View more
09-06-2017
10:58 AM
1 Kudo
Symptom Not able to use reflect function using beeline, but the query works OK with Hvi CLI. Error Message: Error while compiling statement: FAILED: SemanticException UDF reflect is not allowed (state=42000,code=40000) Cause When running set hive.server2.builtin.udf.blacklist from beeline, it will return the following as blacklisted: jdbc:hive2://localhost:10000/default> set hive.server2.builtin.udf.blacklist;
+------------------------------------------------------------------+--+
| set |
+------------------------------------------------------------------+--+
| hive.server2.builtin.udf.blacklist=reflect,reflect2,java_method |
+------------------------------------------------------------------+--+ Reflect UDF is blacklisted by default when running queries through HiveServer2 (beeline, ODBC, JDBC connections), as it was found to be a security risk. The code was modified so if the parameter hive.server2.builtin.udf.blacklist has not been configured or it is blank, its default value will be "reflect,reflect2,java_method". Resolution 1. Open the Ambari UI 2. Add the custom property in Ambari hive.server2.builtin.udf.blacklist under Hive / Configs / Advanced / Custom hive-site and give it any value, for example "empty_blacklist". 3. Restart services as requested by Ambari. 4. Connect again with beeline and verify that blacklist only includes the dummy value now. 0: jdbc:hive2://localhost:10000/default> set hive.server2.builtin.udf.blacklist; +-------------------------------------------+--+
| hive.server2.builtin.udf.blacklist=empty_blacklist |
+-------------------------------------------+--+
5. Reflect should work now without issues.
... View more
07-27-2016
02:10 PM
@Kuldeep Kulkarni I got the same error message. only difference is my env is kerborized & both my rm's are not in standby mode. [yarn@m1 root]$ yarn rmadmin -getServiceState rm1 standby [yarn@m1 root]$ yarn rmadmin -getServiceState rm2 active Ambari doesn't show the state of RM, but getting same exception as above. i tried to switch the roles and that didnot help. Any help is appreciated.
... View more
03-18-2016
09:10 PM
3 Kudos
it is an issue with meta data..we just dropped it and recreated the table and all is well...
... View more
02-05-2016
12:35 AM
@Neeraj Sabharwal yea...we did follow with support and they said it is a know issue...i posted the comments below..alll we need to do is change the DB engine for all the tables which are MyISAM to InnoDB. thanks for your response though..
... View more
01-28-2016
12:29 AM
@Raja Sekhar Chintalapati Did you mean feature or future? 🙂 Not sure why do you want to do this? You can create a config group, assign HS2 and make changes only to that host hive-site using ambari.
... View more
01-27-2016
01:16 AM
3 Kudos
@Raja Sekhar Chintalapati There is no Spark authentication against LDAP in a non kerberized environment. If a Spark job reads from HDFS and the user running the job does not have sufficient HDFS permission, Spark will fail to read data. Spark HiveContext does not connect to HiveServer2. It connects to Hive metastore once you provide the Hive configuration (hive-site.xml) to Spark, else it creates its own metastore in it's working directory I don't know a way to suppress the info in sparl-sql The Spark Master UI is typically on the node with Driver running on port 4040. You can define ports for the Driver, File Server, Executor, UI etc. See doc here
See also setting Spark Configuratin here: https://spark.apache.org/docs/1.1.0/configuration.html See also for YARN Mode: http://spark.apache.org/docs/latest/security.html
Example SPARK_MASTER_OPTS="-Dspark.driver.port=7001 -Dspark.fileserver.port=7002
-Dspark.broadcast.port=7003 -Dspark.replClassServer.port=7004
-Dspark.blockManager.port=7005 -Dspark.executor.port=7006
-Dspark.ui.port=4040 -Dspark.broadcast.factory=org.apache.spark.broadcast.HttpBroadcastFactory"
SPARK_WORKER_OPTS="-Dspark.driver.port=7001 -Dspark.fileserver.port=7002
-Dspark.broadcast.port=7003 -Dspark.replClassServer.port=7004
-Dspark.blockManager.port=7005 -Dspark.executor.port=7006
-Dspark.ui.port=4040 -Dspark.broadcast.factory=org.apache.spark.broadcast.HttpBroadcastFactory" Programmatic Example import org.apache.spark.SparkConf
import org,apache.spark.SparkContext
val conf = new SparkConf()
.setMaster(master)
.setAppName("namexxx")
.set("spark.driver.port", "7001")
.set("spark.fileserver.port", "7002")
.set("spark.broadcast.port", "7003")
.set("spark.replClassServer.port", "7004")
.set("spark.blockManager.port", "7005")
.set("spark.executor.port", "7006")
val sc= new SparkContext(conf)
... View more