Member since
06-21-2017
30
Posts
1
Kudos Received
2
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
537 | 08-24-2018 02:07 PM | |
2191 | 06-20-2018 12:16 PM |
12-10-2018
02:08 PM
Hi Team, I have been monitoring resource utilization on node managers through Ambari metrics. Cache memory is keep increasing on all node manager nodes. Does it impact performance? our cluster has been installed on Centos 7.3. Can you please clarify my doubt? Thanks in Advance.
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache NiFi
-
Apache YARN
11-20-2018
03:21 PM
hi Aviram, how did you resolve this?
... View more
08-24-2018
02:07 PM
Got resolved by configuring adl url with "fs.defaultFS" property in core-site.xml file.
... View more
08-20-2018
03:22 PM
Hi Team, I want to use Azure Data Lake Store in place of HDFS. Want to avoid HDFS completely, is it possible? Can some one help me to where can i configure this? Thanks in advance.
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache YARN
07-17-2018
03:51 PM
Hi Team, I have enabled kerberos on my cluster. followed below steps. 1) Installed kerberos in one node which can be accessible from all cluster nodes. 2) Enabled kerberos from Ambari by giving details of previously created kerberos server. Every thing went fine. But when i try to run spark-submit (word count spark job) from any of data node i am getting below error. "Invalid credentials, no valid tgt found." I tried with all users, hdfs, spark and Yarn (sudo -u hdfs) etc. but no luck. But when i run the same from master node the job is getting executed. What should i do to be able to run spark/hive/any other jobs from any of the data node/edge node? Please advice. Thanks in advance.
... View more
Labels:
- Labels:
-
Apache Spark
-
Apache YARN
06-20-2018
12:21 PM
HI All, I have got lot of dr.who yarn jobs. does any one have any idea? How these jobs are getting created?
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache YARN
06-20-2018
12:16 PM
It got resolved, once move the RM to another node in the cluster.
... View more
06-14-2018
06:00 AM
Hi Vinicius, yes it is already enabled. I am still having this issue.
... View more
05-30-2018
01:56 PM
Hi Team,
ResourceManager servcie is stopping automatically with in few sec.
I have not found any error/exceptions in resourcemanager logs. I am suspecting that there is some issue with Zookeeper. I have three zookeeper services. below are the logs of resource manager and two zookeeper services. Please Help with it.
Resource Manager Logs: 2018-05-30 09:08:38,058 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:java.io.tmpdir=/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir
2018-05-30 09:08:38,058 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:java.compiler=<NA>
2018-05-30 09:08:38,058 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:os.name=Linux
2018-05-30 09:08:38,058 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:os.arch=amd64
2018-05-30 09:08:38,059 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:os.version=3.10.0-693.el7.x86_64
2018-05-30 09:08:38,059 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:user.name=yarn
2018-05-30 09:08:38,059 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:user.home=/home/yarn
2018-05-30 09:08:38,059 INFO zookeeper.ZooKeeper (Environment.java:logEnv(100)) - Client environment:user.dir=/usr/hdp/2.6.3.0-235/hadoop-yarn
2018-05-30 09:08:38,060 INFO zookeeper.ZooKeeper (ZooKeeper.java:<init>(438)) - Initiating client connection, connectString=hdp01.mydomain.com:2181,hdp03.mydomain.com:2181,hdp02.mydomain.com:2181 sessionTimeout=10000 watcher=null
2018-05-30 09:08:38,179 INFO recovery.ZKRMStateStore (ZKRMStateStore.java:createConnection(1276)) - Created new ZK connection
2018-05-30 09:08:38,200 INFO zookeeper.ClientCnxn (ClientCnxn.java:logStartConnect(1019)) - Opening socket connection to server hdp02.mydomain.com/192.168.3.19:2181. Will not attempt to authenticate using SASL (unknown error)
2018-05-30 09:08:38,220 INFO zookeeper.ClientCnxn (ClientCnxn.java:primeConnection(864)) - Socket connection established, initiating session, client: /192.168.3.18:56340, server: hdp02.mydomain.com/192.168.3.19:2181
2018-05-30 09:08:38,279 INFO zookeeper.ClientCnxn (ClientCnxn.java:onConnected(1279)) - Session establishment complete on server hdp02.mydomain.com/192.168.3.19:2181, sessionid = 0x263b11a44120001, negotiated timeout = 10000
2018-05-30 09:08:38,495 INFO recovery.ZKRMStateStore (ZKRMStateStore.java:run(359)) - Fencing node /rmstore/ZKRMStateRoot/RM_ZK_FENCING_LOCK doesn't exist to delete
2018-05-30 09:08:38,793 INFO resourcemanager.ResourceManager (ResourceManager.java:serviceStart(597)) - Recovery started
2018-05-30 09:08:38,851 INFO recovery.RMStateStore (RMStateStore.java:checkVersion(639)) - Loaded RM state version info 1.2 Zookeeper 1 logs: 2018-05-30 09:08:38,208 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@197] - Accepted socket connection from /192.168.3.18:56340
2018-05-30 09:08:38,246 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:ZooKeeperServer@868] - Client attempting to establish new session at /192.168.3.18:56340
2018-05-30 09:08:38,272 - INFO [CommitProcessor:2:ZooKeeperServer@617] - Established session 0x263b11a44120001 with negotiated timeout 10000 for client /192.168.3.18:56340
2018-05-30 09:08:38,285 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0x1 zxid:0x800000042 txntype:-1 reqpath:n/a Error Path:/rmstore Error:KeeperErrorCode = NodeExists for /rmstore
2018-05-30 09:08:38,344 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0x2 zxid:0x800000043 txntype:-1 reqpath:n/a Error Path:/rmstore/ZKRMStateRoot Error:KeeperErrorCode = NodeExists for /rmstore/ZKRMStateRoot
2018-05-30 09:08:38,447 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@590] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:multi cxid:0x4 zxid:0x800000045 txntype:-1 reqpath:n/a aborting remaining multi ops. Error Path:/rmstore/ZKRMStateRoot/RM_ZK_FENCING_LOCK Error:KeeperErrorCode = NoNode for /rmstore/ZKRMStateRoot/RM_ZK_FENCING_LOCK
2018-05-30 09:08:38,510 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0x5 zxid:0x800000046 txntype:-1 reqpath:n/a Error Path:/rmstore/ZKRMStateRoot/RMAppRoot Error:KeeperErrorCode = NodeExists for /rmstore/ZKRMStateRoot/RMAppRoot
2018-05-30 09:08:38,535 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0x6 zxid:0x800000047 txntype:-1 reqpath:n/a Error Path:/rmstore/ZKRMStateRoot/RMDTSecretManagerRoot Error:KeeperErrorCode = NodeExists for /rmstore/ZKRMStateRoot/RMDTSecretManagerRoot
2018-05-30 09:08:38,602 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0x7 zxid:0x800000048 txntype:-1 reqpath:n/a Error Path:/rmstore/ZKRMStateRoot/RMDTSecretManagerRoot/RMDTMasterKeysRoot Error:KeeperErrorCode = NodeExists for /rmstore/ZKRMStateRoot/RMDTSecretManagerRoot/RMDTMasterKeysRoot
2018-05-30 09:08:38,666 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0x8 zxid:0x800000049 txntype:-1 reqpath:n/a Error Path:/rmstore/ZKRMStateRoot/RMDTSecretManagerRoot/RMDelegationTokensRoot Error:KeeperErrorCode = NodeExists for /rmstore/ZKRMStateRoot/RMDTSecretManagerRoot/RMDelegationTokensRoot
2018-05-30 09:08:38,724 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0x9 zxid:0x80000004a txntype:-1 reqpath:n/a Error Path:/rmstore/ZKRMStateRoot/RMDTSecretManagerRoot/RMDTSequentialNumber Error:KeeperErrorCode = NodeExists for /rmstore/ZKRMStateRoot/RMDTSecretManagerRoot/RMDTSequentialNumber
2018-05-30 09:08:38,765 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@643] - Got user-level KeeperException when processing sessionid:0x263b11a44120001 type:create cxid:0xa zxid:0x80000004b txntype:-1 reqpath:n/a Error Path:/rmstore/ZKRMStateRoot/AMRMTokenSecretManagerRoot Error:KeeperErrorCode = NodeExists for /rmstore/ZKRMStateRoot/AMRMTokenSecretManagerRoot
2018-05-30 09:08:45,736 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@197] - Accepted socket connection from /192.168.3.19:38248
2018-05-30 09:08:45,736 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@827] - Processing ruok command from /192.168.3.19:38248
2018-05-30 09:08:45,767 - INFO [Thread-35:NIOServerCnxn@1008] - Closed socket connection for client /192.168.3.19:38248 (no session established for client)
2018-05-30 09:09:02,037 - WARN [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@362] - Exception causing close of session 0x263b11a44120001 due to java.io.IOException: Connection reset by peer
2018-05-30 09:09:02,038 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@1008] - Closed socket connection for client /192.168.3.18:56340 which had sessionid 0x263b11a44120001
2018-05-30 09:09:12,009 - INFO [SessionTracker:ZooKeeperServer@347] - Expiring session 0x263b11a44120001, timeout of 10000ms exceeded
2018-05-30 09:09:12,017 - INFO [ProcessThread(sid:2 cport:-1)::PrepRequestProcessor@492] - Processed session termination for sessionid: 0x263b11a44120001 Zookeeper 2 logs: 2018-05-30 09:08:46,033 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@197] - Accepted socket connection from /192.168.3.18:33482
2018-05-30 09:08:46,033 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@827] - Processing ruok command from /192.168.3.18:33482
2018-05-30 09:08:46,083 - INFO [Thread-20:NIOServerCnxn@1008] - Closed socket connection for client /192.168.3.18:33482 (no session established for client)
2018-05-30 09:09:46,017 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxnFactory@197] - Accepted socket connection from /192.168.3.18:33584
2018-05-30 09:09:46,018 - INFO [NIOServerCxn.Factory:0.0.0.0/0.0.0.0:2181:NIOServerCnxn@827] - Processing ruok command from /192.168.3.18:33584
2018-05-30 09:09:46,052 - INFO [Thread-21:NIOServerCnxn@1008] - Closed socket connection for client /192.168.3.18:33584 (no session established for client) Please help with it. Thanks in Advance. -Paramesh.
... View more
Labels:
02-27-2018
12:49 PM
Hi team, how to downgrade Cloudbreak to 1.16.5 from 2.4.0? I recently upgrade clodubreak to 2.4.0 and image catalog is not working, so i want to degrade it? can you please help me on how to downgrade it to 1.16.5? Thanks.
... View more
Labels:
- Labels:
-
Hortonworks Cloudbreak
02-27-2018
12:29 PM
custom-images-1.txt hi Team, I recently updated my Cloudbreak to 2.4.0 from 1.16.5. I am trying to create custom-image catalog but not able to do that. i am getting below error. 2018-02-27 12:04:06,980 [http-nio-8080-exec-12] propagateImagesIfRequested:332 ERROR c.s.c.s.i.ImageCatalogService - [instance:caa446d9-80af-4efe-9496-a6bd190696d5] [type:ImageCatalogV1] [id:] [name:intucustomimage] [tracking:8ce02a59-efeb-4152-a339-6e438a2f2720] No images was found: com.sequenceiq.cloudbreak.core.CloudbreakImageCatalogException: Failed to process image catalog from 'https://s3-us-west-2.amazonaws.com/xyzabc/pmalla/custom-images_1.json' due to: 'Error reading entity from input stream.'
I have taken the default json and replaced my .vhd image path in the json and changed UUID for this. my json file was uploaded to s3 and it is accessible from browser. I am also able to download this file from wget command. Any help is much appriciated.
... View more
Labels:
- Labels:
-
Hortonworks Cloudbreak
02-21-2018
02:22 PM
@pdarvasi Hi, thanks for your quick reply. What if i want to update some configuration/package in image? is there any way? like can i run some command (through cloud break) after creating new VM from image? Thanks Again.
... View more
02-21-2018
02:13 PM
@pdarvasi Hi, Can i update and use new image, Once i launched the cluster with my custom image? if so, can you tell me how? Thanks in Advance.
... View more
02-06-2018
08:31 AM
Hi, I am using cloud break for cluster auto scaling. When it adds new node to cluster, it is taking almost 15 min to create and install all client stack in it. As i am using custom image, i want to install all client stack in the image. So is there a way to skip client stack installation? while adding node from pre installed custom image? Thanks in Advance.
... View more
Labels:
01-18-2018
10:36 AM
Hi @mmolnar, Thanks, that worked. But i am getting an error, as that image is not having public access. Any tip to resolve it other than giving publicly available image?
... View more
11-23-2017
11:43 AM
Hi, I have installed Nifi (HDF alongside HDP) over HDP through Ambari. It is showing like "No data Available" for Nifi graphs on Ambari dashboards. I checked Nifi logs and observed that Nifi sending metrics Ambari through AmbariReportingTask. INFO [Timer-Driven Process Thread-5] o.a.n.r.ambari.AmbariReportingTask AmbariReportingTask[id=3b90bc0f-a6c0-49db-b721-4dbc04cef28e] Successfully sent metrics to Ambari in 0 ms I have followed the below link and succeeded in getting graphs on Grafana. https://community.hortonworks.com/articles/138614/no-default-grafana-dashboard-for-nifi-in-hdf-on-to-1.html But still Ambri showing no data available for Nifi. Can any one help me out on it. Thanks.
... View more
- Tags:
- ambari-metrics
- ambari-metrics-collector
- dashboard
- Data Ingestion & Streaming
- nifi-controller-service
- nifi-reporting
Labels:
- Labels:
-
Apache Ambari
-
Apache NiFi
11-23-2017
11:26 AM
@Shu, Thanks Shu. That worked.
... View more
11-20-2017
03:22 PM
Hi Team, I am new to Nifi. I just wanted to fetch all rows (3645000) from a SQL table. For this i wanted to provide initial value to the processor. I have given this as below. but it is not taking. Can you please give me some example on it. Initial Max Value = initial.3600000.{max_value_column} Thanks in Advance.
... View more
Labels:
- Labels:
-
Apache NiFi
11-01-2017
10:43 AM
Hi Team, We are exploring Cloudbreak in Azure. We want to use our custom image for HDP installation as some libraries needs to be present in the cluster nodes. As suggested in the article http://sequenceiq.com/cloudbreak-docs/master/images/, i have created "arm-images.yml" in etc directory in cloudbreak-deployment directory. The whole path is "/root/cloudbreak-deployment/etc". below is the content in arm-images.yml file. azure_rm: East Asia: https://sequenceiqeastasia2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd East US: https://mystoredisks1558.blob.core.windows.net/vhds/Hdptestimage120171031154408.vhd Central US: https://sequenceiqcentralus2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd North Europe: https://sequenceiqnortheurope2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd South Central US: https://sequenceiqouthcentralus2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd North Central US: https://sequenceiqorthcentralus2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd East US 2: https://mystoredisks1558.blob.core.windows.net/vhds/Hdptestimage120171031154408.vhd Japan East: https://sequenceiqjapaneast2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd Japan West: https://sequenceiqjapanwest2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd Southeast Asia: https://sequenceiqsoutheastasia2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd West US: https://sequenceiqwestus2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd West Europe: https://sequenceiqwesteurope2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd Brazil South: https://sequenceiqbrazilsouth2.blob.core.windows.net/images/cb-2016-06-14-03-27.vhd I have put my custom image path for East US and East US 2. When i start deploying cluster it is taking the default image "https://sequenceiqeastus12.blob.core.windows.net/images/hdc-hdp--1706211640.vhd" came up with Cloudbreak. i have checked the log, but did not get any related info. Does any one have any idea. Thanks in Advance....
... View more
Labels:
- Labels:
-
Hortonworks Cloudbreak
09-27-2017
03:22 PM
Hi team, I have 3 node HDFS cluster with replication factor as 1. I have
copied 10GB file in to HDFS with “dfs hdfs -put” command then it was divided
into 86 blocks of 128MB each. But all these 86 blocks were stored in one data
node. Is it common behaviour? I expected it to be distribute all 86 blocks across all 3
nodes? Is there any configuration to do this distribution?
... View more
Labels:
- Labels:
-
Apache Hadoop
07-10-2017
10:47 AM
1 Kudo
I am trying to create and overwrite a file in file system. The below code working with local filesystem and NFS as well. But not working with HDFS NFS gateway. open_file = open("/hdfs_nfs/hdfs_Data/sampledata/testWrite.txt", 'wb+')
open_file.write("This is just a sample data")
open_file.close() The file "testWrite.txt" is getting created on HDFS mount point when i run above code for first time. But it is not working when I run the code from 2nd time onwards, with or with out changing the content. I am getting below error. IOError: [Errno 22] invalid mode ('wb+') or filename: '/hdfs_nfs/hdfs_Data/sampledata/testWrite.txt'
... View more
Labels:
- Labels:
-
Apache Hadoop
07-07-2017
11:40 AM
Hi Kuldeep, Thanks for the reply. I tried the method that was mentioned on above given link. still it is not working. I am getting below error. Traceback (most recent call last):
File "write_test.py", line 7, in <module>
open_file.truncate()
IOError: [Errno 22] Invalid argument.
... View more
06-27-2017
05:10 AM
Hi Selavan, First of all thanks for your reply. I used below code with the python write mode of w,w+ and wb+. open_file = open("/hdfs_nfs/hdfs_Data/sampledata/testWrite.txt", 'wb+')
open_file.write("This is just a sample data")
open_file.close() The file "testWrite.txt" is getting created in given path (HDFS Mount point) when I run this for first time. I am getting below error, when I execute above code from second time onwards. IOError: [Errno 22] invalid mode ('wb+') or filename: '/hdfs_nfs/hdfs_Data/sampledata/testWrite.txt' Thanks.
... View more
06-23-2017
12:51 PM
Hi, cp is working with the option "-f". But when i try with python it is failing with below error. IOError: [Errno 22] invalid mode ('wb+') or filename: '/file_path/testWrite.txt' Can some one help with it. Thanks.
... View more
06-21-2017
02:19 PM
I am trying to over write the file with cp command and it failed with below error. cp: cannot create regular file "filename" : Invalid argument i checked the error log and got the below error. ERROR nfs3.RpcProgramNfs3 (RpcProgramNfs3.java:setattr(436)) - Setting file size is not supported when setattr, fileId: 18463 Can some one Help me with it. Thanks in Advance.
... View more
Labels:
- Labels:
-
Apache Hadoop
06-21-2017
02:09 PM
Hi Raghav, I am also facing same issue.. is it resolved for you?
... View more