Member since
01-18-2016
169
Posts
32
Kudos Received
21
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1627 | 06-27-2025 06:00 AM | |
| 1339 | 01-14-2025 06:30 PM | |
| 1860 | 04-06-2018 09:24 PM | |
| 2008 | 05-02-2017 10:43 PM | |
| 5200 | 01-24-2017 08:21 PM |
07-13-2016
12:11 PM
@Saurabh Kumar - Since this example is using the Sandbox zookeeper rather than embedded zk, try adding /solr to the end of your zookeeper entry in the create command like this sandbox.hortonworks.com:2181/solr. I'm not sure if that will solve the issue but when Solr runs in cloud mode with an external zookeeper it makes /solr the zookeeper root to keep it's data separate from other services.
... View more
07-12-2016
06:39 PM
It's telling you that it cannot find the file in your confdir (the directory after -d in your command). Do an ls on the directory and either the directory is missing or the solrconfig.xml is missing from /opt/lucidworks-hdpsearch/solr/server/solr/configsets/data_driven_schema_configs_hdfs/conf. You may have overlooked a step in the setup (Step 2). It includes creating the conf directory and modifying the solrconfig.xml file... cp -R /opt/lucidworks-hdpsearch/solr/server/solr/configsets/data_driven_schema_configs /opt/lucidworks-hdpsearch/solr/server/solr/configsets/data_driven_schema_configs_hdfs
... View more
06-30-2016
04:14 PM
Thanks for the info @vpoornalingam. We resolved the issue. It turns out that when we changed the path, Ambari suggested making other changes, which we mindlessly accepted. Obviously we should have paid more attention. One of the changes was to drop the embedded hbase master heap size to a much lower value. I realized this after looking at the log hbase-ams-master-<FQDN>.out rather than ambari-metrics-collector.log that I was looking at.
... View more
06-29-2016
10:53 PM
Changing it broke Metrics Collector... I need to move the AMS hbase.rootdir to another partition, so I created a directory, did a chown -R ams:hadoop MYDIR, changed the configuration value and restarted AMS. The Metrics Collector will not start. It's throwing a connection refused exception when trying to connect to zookeeper on localhost:61181. Unfortunately, I don't have access to the exact exception at the moment). This is on HDP 2.4 Nothing is listening on port 61181, which I believe should be the embedded ZK port. hbase.zookeeper.property.clientPort={{zookeeper_clientPort}}. I killed all ams processes to be sure something was not in a bad state. I also tried copying the old hbase.rootdir to my new directory with the same permissions but it still fails. When I switch back to the old location it works fine. This seems very similar to changing to distributed mode, so I don't understand what's going wrong.
... View more
Labels:
- Labels:
-
Apache Ambari
06-29-2016
05:12 PM
Awesome. Thanks. I wasn't sure if under the covers Ranger was just doing sql grants.
... View more
06-29-2016
04:54 PM
1 Kudo
We are enabling the Ranger authorization in Hive, but previously we created roles & grants in beeline. Should we remove manually created Hive grants and roles in beeline before switching from StdSqlAuth to Ranger authorization or will magic happen? The roles we created match our AD group names in the policies. (HDP 2.4)
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Ranger
06-27-2016
02:52 PM
Awesome. Thanks.
... View more
06-23-2016
03:20 AM
1 Kudo
@rbiswas, You may have read this but there's some good info here in what they describe as a "real world" production configuration using the new cross-data-center replication: https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=62687462 Since this feature only came out in 6.0 which was released less than 2 months ago, there's probably been limited production use. ALSO....Not a best practice, but since way before Solr Cloud existed, we used a brute force method of cross-data-center replication for stand-by Solrs with the magic of rsync. You can reliably use rsync to copy indexes as they are being updated, but there's a bit of scripting required. I have only done this in non-cloud environments, but I'm pretty sure it can be done in cloud as well. It is crude, but it worked for years and uses some of the great features of linux. Example script, run in crontab from the DR site nodes: #step 1 - create a backup first, assuming your current copy is good.
cp -rl ${data_dir} ${data_dir}.BAK
#step 2 - Now copy from the primary site
status=1
while [ $status != 0 ]; do
rsync -a --delete ${primary_site_node}:${data_dir} ${data_dir}
status=$?
done
echo "COPY COMPLETE!"
That script will create local backup (instantly via hard-links, not soft links) and then copies [only] new files and deletes files from DR that are have been deleted from Primary/remote. If files disappear during the rsync copy, it will copy again until nothing changes during the rsync. This can be run from crontab, but it does need a bit of bullet-proofing. Simple. Crude. It works.
... View more
06-21-2016
05:25 PM
1 Kudo
Note that Solr Cloud's replication is not intended to go across data centers due to volume of traffic and dependency on zookeeper ensembles. However, the recently released 6.x added a special replication to go across data centers. https://issues.apache.org/jira/browse/SOLR-6273, which is based on this description: http://yonik.com/solr-cross-data-center-replication/ Basically, this is a cross-cluster replication, which is different from the standard Solr Cloud's replication mechanism.
... View more
06-08-2016
12:41 PM
1 Kudo
@David Lam did this work for you?
... View more