Member since
08-08-2017
1652
Posts
30
Kudos Received
11
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1964 | 06-15-2020 05:23 AM | |
| 16025 | 01-30-2020 08:04 PM | |
| 2108 | 07-07-2019 09:06 PM | |
| 8244 | 01-27-2018 10:17 PM | |
| 4664 | 12-31-2017 10:12 PM |
12-19-2017
01:07 PM
on ambari GUI by clicking on HDFS --> config --> (we can see the data node folders that actually represented the worker disks) how to use API in order to capture the number of disks from the blueprint ?
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
12-18-2017
03:05 PM
Another quastion , is it possible to find the real problem from the API and not from the log , ?
... View more
12-18-2017
02:53 PM
Dear Aditya , sorry for the late response here the ERROR from the ambari-server.log regarding the ambari GUI , all services/hosts/components are up ( green ) , so I not see any issue from this side grep ERROR ambari-server.log
17 Dec 2017 19:37:58,236 ERROR [pool-3-thread-1] ViewRegistry:1056 - Failed to update 'use.hive.interactive.mode' parameter for viewName: HIVE, version: 1.5.0. Exception: org.apache.ambari.server.ServiceComponentNotFoundException: ServiceComponent not found, clusterName=hdp, serviceName=HIVE, serviceComponentName=HIVE_SERVER_INTERACTIVE
17 Dec 2017 19:37:58,695 ERROR [ambari-client-thread-89] ClusterImpl:2716 - No service found for config types '[cluster-env]', service config version not created
17 Dec 2017 19:38:04,119 ERROR [pool-3-thread-1] ViewRegistry:1056 - Failed to update 'use.hive.interactive.mode' parameter for viewName: HIVE, version: 2.0.0. Exception: org.apache.ambari.server.ServiceComponentNotFoundException: ServiceComponent not found, clusterName=hdp, serviceName=HIVE, serviceComponentName=HIVE_SERVER_INTERACTIVE
17 Dec 2017 19:47:16,278 ERROR [ambari-heartbeat-processor-0] HeartbeatProcessor:554 - Operation failed - may be retried. Service component host: SPARK2_JOBHISTORYSERVER, host: master03.sys6363.com Action id 9-3 and taskId 90
17 Dec 2017 19:47:17,025 ERROR [ambari-action-scheduler] ActionScheduler:440 - Operation completely failed, aborting request id: 9
17 Dec 2017 19:59:19,161 ERROR [Stack Version Loading Thread] URLStreamProvider:245 - Received HTTP 404 response from URL: http://master02.sys6363.com/HDP/hdp_urlinfo.json
17 Dec 2017 19:59:22,550 ERROR [main] ViewRegistry:1056 - Failed to update 'use.hive.interactive.mode' parameter for viewName: HIVE, version: 1.5.0. Exception: org.apache.ambari.server.ServiceComponentNotFoundException: ServiceComponent not found, clusterName=hdp, serviceName=HIVE, serviceComponentName=HIVE_SERVER_INTERACTIVE
17 Dec 2017 19:59:25,500 ERROR [main] ViewRegistry:1056 - Failed to update 'use.hive.interactive.mode' parameter for viewName: HIVE, version: 2.0.0. Exception: org.apache.ambari.server.ServiceComponentNotFoundException: ServiceComponent not found, clusterName=hdp, serviceName=HIVE, serviceComponentName=HIVE_SERVER_INTERACTIVE
17 Dec 2017 17:45:59,151 ERROR [qtp-ambari-agent-67] HeartBeatHandler:200 - CurrentResponseId unknown for worker02.sys6363.com - send register command
... View more
12-18-2017
11:08 AM
this is very huge file , whant you want to search in this file?
... View more
12-18-2017
10:34 AM
after we installed the Ambari cluster and blueprint ( hadoop version 2.6 ) , I run the following API in order to see if blueprint complete successfully the installation curl -s --user "admin"":""admin" -H "X-Requested-By: ambari" -X GET http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1 and it is very strange because when I access the ambari GUI , I see all services are up include the compliments , and all host are up also as master01 , master02 , master03 , kafka , worker01-03 but from the API we get failed on request_status "request_status" : "FAILED", please advice how to know exactly why requisite status is fail ? , what was wrong or how to know fail content ? partital of the output from the API curl -s --user "admin"":""admin" -H "X-Requested-By: ambari" -X GET http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1 | more
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1",
"Requests" : {
"aborted_task_count" : 1,
"cluster_name" : "hdp",
"completed_task_count" : 122,
"create_time" : 1513540772178,
"end_time" : -1,
"exclusive" : false,
"failed_task_count" : 1,
"id" : 1,
"inputs" : null,
"operation_level" : null,
"progress_percent" : 100.0,
"queued_task_count" : 0,
"request_context" : "Logical Request: Provision Cluster 'hdp'",
"request_schedule" : null,
"request_status" : "FAILED",
"resource_filters" : null,
"start_time" : -1,
"task_count" : 122,
"timed_out_task_count" : 0,
"type" : null
},
"stages" : [
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/stages/1",
"Stage" : {
"cluster_name" : "hdp",
"request_id" : 1,
"stage_id" : 1
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/stages/2",
"Stage" : {
"cluster_name" : "hdp",
"request_id" : 1,
"stage_id" : 2
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/stages/3",
"Stage" : {
"cluster_name" : "hdp",
"request_id" : 1,
"stage_id" : 3
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/stages/4",
"Stage" : {
"cluster_name" : "hdp",
"request_id" : 1,
"stage_id" : 4
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/stages/5",
"Stage" : {
"cluster_name" : "hdp",
"request_id" : 1,
"stage_id" : 5
}
}
],
"tasks" : [
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/tasks/10000",
"Tasks" : {
"cluster_name" : "hdp",
"id" : 10000,
"request_id" : 1,
"stage_id" : 1
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/tasks/10001",
"Tasks" : {
"cluster_name" : "hdp",
"id" : 10001,
"request_id" : 1,
"stage_id" : 1
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/tasks/10002",
"Tasks" : {
"cluster_name" : "hdp",
"id" : 10002,
"request_id" : 1,
"stage_id" : 1
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/tasks/10003",
"Tasks" : {
"cluster_name" : "hdp",
"id" : 10003,
"request_id" : 1,
"stage_id" : 1
}
},
{
"href" : "http://master02.sys6363.com:8080/api/v1/clusters/hdp/requests/1/tasks/10004",
"Tasks" : {
"cluster_name" : "hdp",
"id" : 10004,
"request_id" : 1,
"stage_id" : 1
}
},
{
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
12-09-2017
08:12 PM
we want to define the RAID configuration on our masters machines
we have 3 masters machines ( while ambari-server installed on master02 machine )
each master machine have 3 partitions
one partition for hadoop
one partition for zookeeper
one partition for hdfs
as the following: /dev/sdb 393G 252M 392G 1% /hadoop
/dev/sdc 393G 298M 392G 1% /BD/hadoop/zookeeper
/dev/sdd 908G 257M 9008G 1% /BD/hadoop/hdfs for now we configured the RAID as the following: raid 0 for /dev/sdb and /dev/sdc raid 1+0 (raid10) for /dev/sdd
dose the raid configuration that we configured for the partitions on the masters machine is OK , or maybe need to configured raid10 also for the other partitions?
... View more
Labels:
- Labels:
-
Apache Hadoop
12-07-2017
07:09 AM
worker08 machine was deleted by mistake from ambari cluster ( by host action --> delete host ) now we want to join this machine again to the ambari cluster we cant use the option from the ambari ( add new host ) , because this option is for new machine and not for machine that already installed as in our case * worker08 is already installed with (datanode , nodemanager , matrix monitor and all client components ) is there any API command to do the join process on worker08?
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
12-06-2017
12:52 PM
@Jay what about just set the port 2182 in the "Port for running ZK Server" instead 2181 , this isn't elegant solution but just to start the Zk on node - master02 , it will works?
... View more
12-06-2017
12:38 PM
one of them is - ambari-metrics-collector service for example
... View more
12-06-2017
12:28 PM
@Jay one note - but if I will kill the process it will affected on other services , so killing all them will showdown 3-4 services from ambari cluster , so my quastion is how to avoid the appliaction not to use the port that alocated already
... View more