Member since
05-09-2016
280
Posts
58
Kudos Received
31
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3749 | 03-28-2018 02:12 PM | |
3023 | 01-09-2018 09:05 PM | |
1649 | 12-13-2016 05:07 AM | |
5050 | 12-12-2016 02:57 AM | |
4317 | 12-08-2016 07:08 PM |
10-31-2016
11:03 PM
HI @azeltov, I am trying to install R-studio on Hortonworks sandbox 2.5, running through the exception in verify installation step: initctl: Unable to connect to Upstart: Failed to connect to socket /com/ubuntu/upstart: Connection refused I have tried starting, stopping rstudio server, it shows the same message. PS: Since it is a docker container, 8787 port is not opened so I have configured /etc/rstudio/rserver.conf to use port 9000.
... View more
10-31-2016
07:19 PM
Thank you so much @Constantin Stanca, I followed the other link and installed R, rhdfs, rmr2 and rhive. It was a great help.
... View more
10-27-2016
03:02 AM
1 Kudo
Hi guys, I am trying to install R on Hortonworks Sandbox, referring this right now: https://github.com/randerzander/r-service Install step got successful but getting the following error in Check R step. resource_management.core.exceptions.Fail: Execution of 'R --file=/var/lib/ambari-agent/cache/stacks/HDP/2.5/services/r-service/package/scripts/r/smoke_test.r' returned 1. Error in objects(db.pos, all.names = TRUE) :
2 arguments passed to .Internal(ls) which requires 3
Execution halted This repo was meant for HDP2.2 but I am using with 2.5, is there a updated version for this repo?
... View more
Labels:
10-19-2016
11:35 PM
No problem @Sabine Maennel , please accept the best answer to close the thread.
... View more
10-19-2016
10:52 PM
It worked, I changed regexp_replace to replace function. Used the following command: %pyspark from pyspark.sql.functions import * extension_df4 = extension_df1.replace('','None','Extension').alias('Extension') extension_df4.show(100,truncate=False) It gives me the following output: |Extension| |gif | |gif | |gif | |gif | |html | |gif | |html | |html | |None | |None | |gif |
|
... View more
10-19-2016
10:30 PM
More information, when I am doing: %pyspark from pyspark.sql.functions import * extension_df3 = extension_df1.select(regexp_replace('Extension','','None').alias('Extension')) extension_df3.show(100,truncate=False) It changes the data frame in the state which I do not want: Extension | |NonegNoneiNonefNone | |NonegNoneiNonefNone | |NonegNoneiNonefNone | |NonegNoneiNonefNone | |NonehNonetNonemNonelNone| |NonegNoneiNonefNone | |NonehNonetNonemNonelNone|
| NonehNonetNonemNonelNone| |None | |None | |NonegNoneiNonefNone |
... View more
10-19-2016
09:22 PM
I am using Spark 1.6.2 and I have a data frame like this: |Extension| |gif | |gif | |gif | |gif | |html | |gif | |html | |html | | | | | |gif | As you can see, there are some blank rows. They are not null because when I ran isNull() on the data frame, it showed false for all records. Then I thought of replacing those blank values to something like 'None' using regexp_replace. It does not affect the data frame column values. Running the following command right now: %pyspark from pyspark.sql.functions import * extension_df3 = extension_df1.select(regexp_replace('Extension','\\s','None').alias('Extension')) extension_df3.show(100,truncate=False) I am replacing on the basis of white space which I guess is wrong. Can somebody please guide me how to do it?
... View more
Labels:
- Labels:
-
Apache Spark
10-19-2016
06:23 PM
2 Kudos
@Sabine Maennel , you have got it wrong. You are running scp command from the sandbox terminal itself which doesn't have ~/Downloads/trace_50.zip. Run this command from the local machine terminal and the file will be copied. Let us know if you get any issue. PS - Do Not change the port to 2122. 2122 is for VM and 2222 is for docker container. Use 2222 because HDP components are installed in the docker container.
... View more
10-17-2016
06:21 PM
@Yukti Agrawal can you try shutting down all Zookeepers and start one at a time?
... View more
10-14-2016
08:39 PM
@prabhavathi Muthusenapathy, it is because that Ambari is not managing NiFi, you have installed it on your own. But Nifi is still there, you can start/stop it, view its logs and libraries under /root/HDF-2.0.0.0/ from the terminal.
... View more