Member since 
    
	
		
		
		09-24-2015
	
	
	
	
	
	
	
	
	
	
	
	
	
	
			
      
                527
            
            
                Posts
            
        
                136
            
            
                Kudos Received
            
        
                19
            
            
                Solutions
            
        My Accepted Solutions
| Title | Views | Posted | 
|---|---|---|
| 2882 | 06-30-2017 03:15 PM | |
| 4357 | 10-14-2016 10:08 AM | |
| 9584 | 09-07-2016 06:04 AM | |
| 11622 | 08-26-2016 11:27 AM | |
| 1910 | 08-23-2016 02:09 PM | 
			
    
	
		
		
		10-25-2016
	
		
		03:17 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 After create a table with 3 primary key on ROW KEY with phoenix:  CREATE TABLE IF NOT EXISTS tabla1(
c1 VARCHAR not null,
c2 VARCHAR not null,
c3 VARCHAR not null,
c4 DOUBLE,
c5 VARCHAR,
CONSTRAINT pk PRIMARY KEY (c1,c2,c3) );  and try to insert from official site:    df.write \
.format("org.apache.phoenix.spark") \
.mode("overwrite") \
.option("table", "XXXXXX") \
.option("zkUrl", "XXXXXXX:2181") \
.save()  y receive this error  Caused by: org.apache.phoenix.schema.ConstraintViolationException: org.apache.phoenix.schema.TypeMismatchException: ERROR 203 (22005): Type mismatch. VARCHAR cannot be coerced to DOUBLE
at org.apache.phoenix.schema.types.PDataType.throwConstraintViolationException(PDataType.java:282)
at org.apache.phoenix.schema.types.PDouble.toObject(PDouble.java:129)
at org.apache.phoenix.jdbc.PhoenixPreparedStatement.setObject(PhoenixPreparedStatement.java:442)
at org.apache.phoenix.spark.PhoenixRecordWritable$$anonfun$write$1.apply(PhoenixRecordWritable.scala:53)
at org.apache.phoenix.spark.PhoenixRecordWritable$$anonfun$write$1.apply(PhoenixRecordWritable.scala:44)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at org.apache.phoenix.spark.PhoenixRecordWritable.write(PhoenixRecordWritable.scala:44)
at org.apache.phoenix.mapreduce.PhoenixRecordWriter.write(PhoenixRecordWriter.java:78)
at org.apache.phoenix.mapreduce.PhoenixRecordWriter.write(PhoenixRecordWriter.java:39)
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12$$anonfun$apply$4.apply$mcV$sp(PairRDDFunctions.scala:1113)
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12$$anonfun$apply$4.apply(PairRDDFunctions.scala:1111)
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12$$anonfun$apply$4.apply(PairRDDFunctions.scala:1111)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1250)
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12.apply(PairRDDFunctions.scala:1119)
at org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1$$anonfun$12.apply(PairRDDFunctions.scala:1091)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
at org.apache.spark.scheduler.Task.run(Task.scala:89)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
... 1 more
Caused by: org.apache.phoenix.schema.TypeMismatchException: ERROR 203 (22005): Type mismatch. VARCHAR cannot be coerced to DOUBLE
at org.apache.phoenix.exception.SQLExceptionCode$1.newException(SQLExceptionCode.java:71)
at org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
... 22 more  There is any problem wiht the row keys???  Regards 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
 - 
						
							
		
			Apache HBase
 - 
						
							
		
			Apache Phoenix
 - 
						
							
		
			Apache Spark
 
			
    
	
		
		
		10-25-2016
	
		
		09:22 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi:  finally the jar that I have needed was /usr/hdp/current/hbase-client/lib/hbase-server.jar.  Regards 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-25-2016
	
		
		08:15 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi:  I have HDP 2.4.0.0 and Iam trying to use Phoenix 2.4.0 on Spark 1.6.0:  I am using the next classpath but iam receiving this error, I need anything else?     Traceback(most recent call last):  File"<stdin>", line 5,in<module>  File"/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/readwriter.py", line 385,in save  self._jwrite.save()  File"/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813,in __call__  File"/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45,in deco  return f(*a,**kw)  File"/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 308,in get_return_value  py4j.protocol.Py4JJavaError:An error occurred while calling o67.save.  : java.lang.NoClassDefFoundError: org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy  Causedby: java.lang.ClassNotFoundException: org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy  
    Driverand executor path Iamusing:  /usr/hdp/2.4.0.0-169/phoenix/lib/phoenix-spark-4.4.0.2.4.0.0-169.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/hbase-client.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/hbase-common.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/phoenix-core-4.4.0.2.4.0.0-169.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/hbase-protocol.jar   Regards 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
 - 
						
							
		
			Apache HBase
 - 
						
							
		
			Apache Phoenix
 - 
						
							
		
			Apache Spark
 
			
    
	
		
		
		10-25-2016
	
		
		07:55 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi:  I have HDP 2.4.0.0 and Iam trying to use Phoenix 2.4.0 on Spark 1.6.0:  I am using the next classpath but iam receiving this error, I need anything else?  Traceback (most recent call last):
  File "<stdin>", line 5, in <module>
  File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/readwriter.py", line 385, in save
    self._jwrite.save()
  File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__
  File "/usr/hdp/2.4.0.0-169/spark/python/pyspark/sql/utils.py", line 45, in deco
    return f(*a, **kw)
  File "/usr/hdp/2.4.0.0-169/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 308, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling o67.save.
: java.lang.NoClassDefFoundError: org/apache/hadoop/hbase/regionserver/ConstantSizeRegionSplitPolicy
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy
  Driver and executor path Iam using:
/usr/hdp/2.4.0.0-169/phoenix/lib/phoenix-spark-4.4.0.2.4.0.0-169.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/hbase-client.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/hbase-common.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/phoenix-core-4.4.0.2.4.0.0-169.jar:/usr/hdp/2.4.0.0-169/phoenix/lib/hbase-protocol.jar  Regards 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-19-2016
	
		
		11:43 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi: the HDP Version is 2.5.0.0 .      The error is:  HTTP ERROR 500
Problem accessing /views/ZEPPELIN/1.0.0/ZEPPELIN/. Reason:
    Server Error
Caused by:
org.apache.jasper.JasperException: PWC6033: Unable to compile class for JSP
PWC6199: Generated servlet error:
The type java.io.ObjectInputStream cannot be resolved. It is indirectly referenced from required .class files
 at org.apache.jasper.compiler.DefaultErrorHandler.javacError(DefaultErrorHandler.java:123)
 at org.apache.jasper.compiler.ErrorDispatcher.javacError(ErrorDispatcher.java:296)
 at org.apache.jasper.compiler.Compiler.generateClass(Compiler.java:376)
 at org.apache.jasper.compiler.Compiler.compile(Compiler.java:437)
 at org.apache.jasper.JspCompilationContext.compile(JspCompilationContext.java:608)
 at org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:360)
 at org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:486)
 at org.apache.jasper.servlet.JspServlet.service(JspServlet.java:380)
 at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
 at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:684)
 at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:501)
 at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
 at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:575)
 at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
 at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1086)
 at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:427)
 at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
 at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1020)
 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-19-2016
	
		
		08:34 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
	
		1 Kudo
		
	
				
		
	
		
					
							 hi:  Any Reason why i can see zeppelin view from ambari?, here the error:  HTTP ERROR 500
Problem accessing /views/ZEPPELIN/1.0.0/ZEPPELIN/. Reason:
    Server Error
Caused by:
org.apache.jasper.JasperException: PWC6033: Unable to compile class for JSP
PWC6199: Generated servlet error:
The type java.io.ObjectInputStream cannot be resolved. It is indirectly referenced from required .class files
	at org.apache.jasper.compiler.DefaultErrorHandler.javacError(DefaultErrorHandler.java:123)
	at org.apache.jasper.compiler.ErrorDispatcher.javacError(ErrorDispatcher.java:296)
	at org.apache.jasper.compiler.Compiler.generateClass(Compiler.java:376)
	at org.apache.jasper.compiler.Compiler.compile(Compiler.java:437)
	at org.apache.jasper.JspCompilationContext.compile(JspCompilationContext.java:608)
	at org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:360)
	at org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:486)
	at org.apache.jasper.servlet.JspServlet.service(JspServlet.java:380)
	at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
	at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:684)
	at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:501)
	at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:137)
	at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:575)
	at org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:231)
	at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1086)
	at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:427)
	at org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:193)
	at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1020)
	at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:135)
	at org.eclipse.jetty.server.Dispatcher.forward(Dispatcher.java:276)
	at org.eclipse.jetty.server.Dispatcher.forward(Dispatcher.java:103)
	at org.apache.ambari.view.zeppelin.ZeppelinServlet.doGet(ZeppelinServlet.java:68)
	at javax.servlet.http.HttpServlet.service(HttpServlet.java:687)
	at javax.servlet.http.HttpServlet.service(HttpServlet.java:790)
	at org.eclipse.jetty.servlet.ServletHolder.handle(ServletHolder.java:684) 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
 - 
						
							
		
			Apache Ambari
 - 
						
							
		
			Apache Zeppelin
 
			
    
	
		
		
		10-16-2016
	
		
		02:34 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi:  I have edit vi /usr/hdp/current/zeppelin-server/lib/conf/shiro.ini:  [urls]
/api/version = anon
#/** = anon
/** = authcBasic
[users]
admin = admin
hdfs = hdfs    
 and restart zeppelin, but the login doesnt appear, just the anonimous user. I need anything else ??? 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-16-2016
	
		
		08:18 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi:  i tried the new code, and its working now like that:  REPLACE($0,'[^a-zA-Z\\0361\\n\\.\\-]+','')
  Many thanks. 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-14-2016
	
		
		01:57 PM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 hi:  i tried this but still doesnt work, any other suggestion? many many thanks  REPLACE($0,'[^\\361a-zA-Z\\n\\.\\-]+)',''))
or
REPLACE($0,'([^a-zA-Z\\n\\.\\-\\361]+)','')) 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
			
    
	
		
		
		10-14-2016
	
		
		10:17 AM
	
	
	
	
	
	
	
	
	
	
	
	
	
	
		
	
				
		
			
					
				
		
	
		
					
							 Hi:  I need that the 'ñ' caracter DOESNT be replaced in this code,   REPLACE($0,'([^a-zA-Z\\n\\.\\-]+)',''))  anny sugestion about how can do that?  thanks 
						
					
					... View more
				
			
			
			
			
			
			
			
			
			
		
		
			
				
						
							Labels:
						
						
		
			
	
					
			
		
	
	
	
	
				
		
	
	
- Labels:
 - 
						
							
		
			Apache Pig