Member since
04-14-2016
54
Posts
9
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
21793 | 06-27-2016 07:20 AM | |
2001 | 05-09-2016 10:10 AM |
06-27-2016
12:11 PM
Thank you. But I've already done this step and I needed to handle multiple files.
Currently this is solved thank you
... View more
06-27-2016
07:20 AM
Hello ! Thank you very much for your suggestions.
These methods have worked and I found another very suitable method is that of using the dataframe . Cordially
... View more
06-27-2016
07:16 AM
Hello,
I want to know is how can I run a python script that contains commands spark ? Here is my python script that I would run into a python environment : #!/usr/bin/python2.7 from pyspark.sql import HiveContext from pyspark import SparkContext from pandas.DataFrame.ix import DataFrame as df hive_context = HiveContext(sc) qvol1 = hive_context.table("table") qvol2 = hive_context.table("table") qvol1.registerTempTable("qvol1_temp") qvol2.registerTempTable("qvol2_temp") df=hive_context.sql("request") df.show()
... View more
Labels:
- Labels:
-
Apache Spark
06-24-2016
07:50 AM
2 Kudos
Hello,
Please I want to read a hive table from a python script.
Can you help me please?
My cordial thanks
... View more
Labels:
- Labels:
-
Apache Hive
06-22-2016
01:54 PM
Thank you. but I would go directly from the csv file to the hive orc table format without creating the textfile data. Thank
... View more
06-22-2016
01:39 PM
Hello, Is it possible to import data from a CSV file into a hive table the orc format? Thank
... View more
Labels:
- Labels:
-
Apache Hive
06-21-2016
08:30 AM
Hello,
Thank you for the directive. But I 'm new to the dataframe and what I try to do is be able to make it to retrieve the values of the indices i and i + 1 for example.
Best regards
... View more
06-14-2016
08:43 AM
Hello,
Please I will like to iterate and perform calculations accumulated in a column of my dataframe but I can not. Can you help me?
Thank you Here the creation of my dataframe. I would like to calculate an accumulated blglast the column and stored in a new column from pyspark.sql import HiveContext
from pyspark import SparkContext
from pandas import DataFrame as df
sc =SparkContext()
hive_context = HiveContext(sc)
tab = hive_context.table("table")
tab.registerTempTable("tab_temp")
df=hive_context.sql("SELECT blglast FROM tab_temp AS b limit 50") df.show()
... View more
Labels:
- Labels:
-
Apache Spark
06-13-2016
11:28 AM
Thank you. By adding the attribute --map-column-hive Date=Timestamp to Sqoop everything works.
... View more
06-10-2016
02:38 PM
Thanks. But it creates another error : "Hive does not support the SQL type for column date" Thanks
... View more