Member since
03-19-2019
67
Posts
38
Kudos Received
19
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3773 | 08-03-2018 03:48 PM | |
945 | 05-04-2018 03:26 PM | |
1104 | 01-12-2018 07:16 PM | |
1474 | 09-13-2017 02:29 PM | |
3609 | 05-10-2017 07:59 PM |
03-27-2017
09:05 AM
Hi, Unfortunately in this case you can update the DB the following way: docker exec -it cbreak_cbdb_1 psql -U postgres
select id,attributes from credential where cloudplatform='AZURE_RM';
update credential set attributes='XX' where id=X;
1st step to exec into the DB docker container 2nd to select the credential you want to update (please note the ID, and copy the whole attributes filed which you want to update) 3rd update the field with the desired changes (please make sure the formatting is correct and the same as before)
... View more
03-23-2017
10:08 PM
I think this is an Ambari bug. You can report the issue here after you've registered: http://issues.apache.org/jira/browse/AMBARI . Also could you please accept the answer here so others can find the solution while it's beeing fixed. Thank you. Br, Krisz
... View more
03-23-2017
03:02 PM
Hi, 22 and 9443 must be open otherwise the provision will fail. 9443 is used with 2-way-ssl for CB to orchestrate the install.
... View more
03-23-2017
02:55 PM
Hi, I can reproduce your issue on HDP 2.5 as well. It seems like the spark assembly contains invalid azure storage jars. To fix this quickly I did the following: mkdir -p /tmp/jarupdate && cd /tmp/jarupdate
find /usr/hdp/ -name "azure-storage*.jar"
cp /usr/hdp/2.5.0.1-210/hadoop/lib/azure-storage-2.2.0.jar .
cp /usr/hdp/current/spark-historyserver/lib/spark-assembly-1.6.3.2.5.0.1-210-hadoop2.7.3.2.5.0.1-210.jar .
unzip azure-storage-2.2.0.jar
jar uf spark-assembly-1.6.3.2.5.0.1-210-hadoop2.7.3.2.5.0.1-210.jar com/
mv -f spark-assembly-1.6.3.2.5.0.1-210-hadoop2.7.3.2.5.0.1-210.jar /usr/hdp/current/spark-historyserver/lib/spark-assembly-1.6.3.2.5.0.1-210-hadoop2.7.3.2.5.0.1-210.jar
cd .. && rm -rf /tmp/jarupdate
Basically I put the desired class files into the assembly jar and updated the original jar file. Once it's done just start the history server and it should be ok. I changed in the spark defaults 2 configuration: spark.eventLog.dir = wasb://cloudbreak492@kriszwasbnorth.blob.core.windows.net/spark-history
spark.history.fs.logDirectory = wasb://cloudbreak492@kriszwasbnorth.blob.core.windows.net/spark-history Once the history server started I was able to start the spark teragen job: spark-submit --class com.nexr.spark.terasort.TeraGen --deploy-mode cluster --master yarn-cluster --num-executors 1 spark-terasort-0.1.jar 1G wasb://cloudbreak492@kriszwasbnorth.blob.core.windows.net/teradata
... View more
03-22-2017
06:14 AM
1 Kudo
Hi, You can download the CLI like this: curl -kL https://ec2-34-251-140-175.eu-west-1.compute.amazonaws.com/hdc-cli_$(uname)_x86_64.tgz | tar -xz Here is an article about this: HDCloud using the HDC cli
... View more
03-21-2017
09:53 AM
Hi, Are you using wasb as the default filesystem? If so then you can check the hdfs-site for the fs.defaultFS. I assume it should look like something like this: wasb://<storage_account>/spark-history
... View more
03-09-2017
04:55 PM
2 Kudos
Hi, It looks like it's a bug in Cloudbreak when you don't attach volumes to the instances. It can be configured when you create a template on the UI. We'll get this fixed.
... View more
03-02-2017
11:54 AM
6 Kudos
Download the HDC cli from the cloud controller Although it's possible to download the cli from the cloud controller UI we're going to use CURL for this, since we're about to automate the process.
All we need is the address of the cloud controller (we make sure with $(uname) that we download the appropriate binary for our OS, note: all of the instance references are examples, there is no real running aws instance): curl -kL https://ec2-34-251-140-175.eu-west-1.compute.amazonaws.com/hdc-cli_$(uname)_x86_64.tgz | tar -xz Once it's downloaded we can verify it's version: ./hdc --version
hdc version 1.13.0-2017-02-09T08:56:06
Configure the cli to use our cloud controller The cli can connect to multiple cloud controllers by providing it's address and credentials. Each command has parameters to provide these informations (even can be configured with environment variables):
--server value server address [$CB_SERVER_ADDRESS] --username value user name (e-mail address) [$CB_USER_NAME] --password value password [$CB_PASSWORD] For convenience we could save them into a file so we wouldn't have to provide these parameters for each command: ./hdc configure --server https://ec2-34-251-140-175.eu-west-1.compute.amazonaws.com --username admin@hortonworks.com --password 'adminPassword123!' but it's not really a good idea from automation perspective, because in the logs we want to see each command as it is. Create a cluster using a JSON skeleton Let's generate our base json skeleton: ./hdc create-cluster generate-cli-skeleton
{
"ClusterName": "",
"HDPVersion": "2.5",
"ClusterType": "EDW-ETL: Apache Hive 1.2.1, Apache Spark 1.6",
"Master": {
"InstanceType": "m4.4xlarge",
"VolumeType": "gp2",
"VolumeSize": 32,
"VolumeCount": 1,
"InstanceCount": 1,
"Recipes": []
},
"Worker": {
"InstanceType": "m3.xlarge",
"VolumeType": "ephemeral",
"VolumeSize": 40,
"VolumeCount": 2,
"InstanceCount": 3,
"Recipes": [],
"RecoveryMode": "AUTO"
},
"Compute": {
"InstanceType": "m3.xlarge",
"VolumeType": "ephemeral",
"VolumeSize": 40,
"VolumeCount": 1,
"InstanceCount": 0,
"Recipes": [],
"RecoveryMode": "AUTO",
"SpotPrice": "0"
},
"SSHKeyName": "",
"RemoteAccess": "",
"WebAccess": true,
"HiveJDBCAccess": true,
"ClusterComponentAccess": false,
"ClusterAndAmbariUser": "",
"ClusterAndAmbariPassword": "",
"InstanceRole": "CREATE",
"Network": {
"VpcId": "",
"SubnetId": ""
},
"Tags": {},
"HiveMetastore": {
"Name": "",
"Username": "",
"Password": "",
"URL": "",
"DatabaseType": ""
},
"Configurations": []
}
As you can see there are default values for certain properties, but it is also missing a few which we need to provide. To manipulate the json we're going to use a really handy tool called JQ. In this tutorial we're not going to change instance types, volumes etc.. as we could, but for now for demonstration purposes let's set the missing properties and write it to a file called cluster.json: ./hdc create-cluster generate-cli-skeleton | jq '.ClusterName = "tutorial-cluster" | .Worker.InstanceCount = 1 | .Compute.SpotPrice = "0.5" | .Compute.InstanceCount = 1 | .Compute.RecoveryMode = "MANUAL" | .SSHKeyName = "my-aws-key" | .RemoteAccess = "0.0.0.0/0" | .ClusterComponentAccess = true | .ClusterAndAmbariUser = "admin" | .ClusterAndAmbariPassword = "admin"' > cluster.json
and the result should look something like this: cat cluster.json
{
"ClusterName": "tutorial-cluster",
"HDPVersion": "2.5",
"ClusterType": "Data Science: Apache Spark 1.6, Apache Zeppelin 0.6.0",
"Master": {
"InstanceType": "m4.4xlarge",
"VolumeType": "gp2",
"VolumeSize": 32,
"VolumeCount": 1,
"InstanceCount": 1,
"Recipes": []
},
"Worker": {
"InstanceType": "m3.xlarge",
"VolumeType": "ephemeral",
"VolumeSize": 40,
"VolumeCount": 2,
"InstanceCount": 1,
"Recipes": [],
"RecoveryMode": "AUTO"
},
"Compute": {
"InstanceType": "m3.xlarge",
"VolumeType": "ephemeral",
"VolumeSize": 40,
"VolumeCount": 1,
"InstanceCount": 1,
"Recipes": [],
"RecoveryMode": "MANUAL",
"SpotPrice": "0.5"
},
"SSHKeyName": "my-aws-key",
"RemoteAccess": "0.0.0.0/0",
"WebAccess": true,
"HiveJDBCAccess": true,
"ClusterComponentAccess": true,
"ClusterAndAmbariUser": "admin",
"ClusterAndAmbariPassword": "admin",
"InstanceRole": "CREATE",
"Network": {
"VpcId": "",
"SubnetId": ""
},
"Tags": {},
"HiveMetastore": {
"Name": "",
"Username": "",
"Password": "",
"URL": "",
"DatabaseType": ""
},
"Configurations": []
}
We used 0.0.0.0/0 for the RemoteAccess, but in production clusters it is highly discouraged. Let's create this cluster and wait until it finishes. We're going to use the --wait flag so we don't have to write some custom functions to poll the cluster state: ./hdc create-cluster --cli-input-json cluster.json --server https://ec2-34-251-140-175.eu-west-1.compute.amazonaws.com --username admin@hortonworks.com --password 'adminPassword123!' --wait true Once the command returned we can check the instances: ./hdc describe-cluster instances --cluster-name tutorial-cluster --server https://ec2-34-251-140-175.eu-west-1.compute.amazonaws.com --username admin@hortonworks.com --password 'adminPassword123!'
[
{
"InstanceId": "i-xxxx",
"Hostname": "ip-10-0-1-159.eu-west-1.compute.internal",
"PublicIP": "x.x.x.x",
"PrivateIP": "10.0.1.159",
"InstanceStatus": "REGISTERED",
"HostStatus": "HEALTHY",
"Type": "master - ambari server"
},
{
"InstanceId": "i-xxxx",
"Hostname": "ip-10-0-1-185.eu-west-1.compute.internal",
"PublicIP": "x.x.x.x",
"PrivateIP": "10.0.1.185",
"InstanceStatus": "REGISTERED",
"HostStatus": "HEALTHY",
"Type": "worker"
},
{
"InstanceId": "i-xxxx",
"Hostname": "ip-10-0-1-197.eu-west-1.compute.internal",
"PublicIP": "x.x.x.x",
"PrivateIP": "10.0.1.197",
"InstanceStatus": "REGISTERED",
"HostStatus": "HEALTHY",
"Type": "compute"
}
]
Terminate the cluster Now that we have a cluster we can execute jobs, queries etc.. which in this tutorial we're not going to cover and if we no longer need the cluster we can simply terminate it: ./hdc terminate-cluster --cluster-name tutorial-cluster --server https://ec2-34-251-140-175.eu-west-1.compute.amazonaws.com --username admin@hortonworks.com --password 'adminPassword123!' --wait true
... View more
Labels:
02-15-2017
12:03 PM
I think the root has been changed can you try with: cb/api/v1/stacks/3 instead of /api/v1/stacks/3.
... View more
12-08-2016
08:24 AM
You can see the supported versions here: http://sequenceiq.com/cloudbreak-docs/latest/openstack/ Unfortunately we didn't get a chance to test it with newton, but most likely it works with that version too.
... View more