Member since
10-25-2019
16
Posts
8
Kudos Received
4
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1808 | 05-21-2020 06:39 AM | |
| 1873 | 05-21-2020 05:05 AM | |
| 9219 | 05-17-2020 11:58 AM | |
| 8477 | 05-05-2020 01:22 AM |
03-05-2025
05:24 AM
1 Kudo
Hello I had a similar problem in the past: i had snapshots policies that did hourly snapshots and retain the last hours + 3 days + 4 weeks . the disk space was released after i deleted snapshots as the deletion of files doesn't delete the snapshoted hdfs data's blocs i also emptied my HDFS Trash and checked that all users HDFS Trash DIRs where empty or small. maybe tha could help Regards
... View more
05-21-2020
06:39 AM
2 Kudos
apt-get installation doesn't seem to install any bitcoin package same thing via python package manager (pip ...) it's probably a mistake in the dockerfile anyway the docker image is old, and the github repo doesn't seem to exist any more
... View more
05-21-2020
05:05 AM
1 Kudo
cloudera CDP is based on a Cloudera Runtime Version + a Cloudera Manager version that is compatible (with that cloudera runtime) https://docs.cloudera.com/cdpdc/7.0/release-guide/topics/cdpdc-release-notes-links.html at time of writing: CDP DC 1.0 uses Cloudera runtime 7.0.3 and cloudera manager 7.0.3 the Cloudera Runtime Component Version is aimed to keep a set of consistent hadoop components versions that can work together. it will also make more easy to migrate from CDH/HDP if services/components versions are the same or close to the runtime components versions of CDP if i'm not wrong, there is actually only one CDP DC (1.0) version with minor updates of CM and cloudera components runtime versions https://docs.cloudera.com/cloudera-manager/7.0.3/release-notes/topics/cm-release-notes.html https://docs.cloudera.com/runtime/7.0.3/release-notes/topics/rt-runtime-component-versions.html
... View more
05-17-2020
11:58 AM
1 Kudo
in your nifi app log if you have "out of memory" errors it may indicate that you need to add more memory (or a possible memory leak) increasing heap too much will cause long GC (especially with Old Generation Objects) you may need to use a more efficient GC if you use large heap size in bootstrap.conf uncomment line : #java.arg.13=-XX:+UseG1GC as you know, some processors are memory intensive queues between processors use memory, ... until nifi.queue.swap.threshold is reached (default 20000 -> if reached nifi will push to disk remaining DFs ) -> too much swapping in queues will realy affect performances. (memory + disk) there are many points that can help you to determine where the PB may be for ex. (processor 1) --> queue --> (processor 2) - check queue occupation : if always full for some processors parallelize more in following processors (especially if processor a re not memory intensive, and for ex if the processor is cpu intenssive and look likes it's not ingestings quickly from queue adding more threads will help to relieve the queue ). also ading back-pressus/control-rate before. in tha case just checking "in" and "out" of processor 2 will confirm that (high IN and low OUT), increase nb of threads -> (threadsScheduling>Concurrent Tasks), sometimes tuning "Run Schedule" and "Run Duration" may also help depending of ingres DFs nature/size and processor type and how parallelization is impacted take a look at "Total queued data" in the status bar (under components bar on top of the UI) keep in mind that processors may be I/O or CPU or memory intensive or all, and parallelizing more to reduce backlogged data may solve memory PBs but it will cause more threads overuse. adding more NIFI nodes will remain the ultimate solution to increase resources to increase performances, you may also want to take a look at nifi.provenance.repository.index.threads nifi.provenance.repository.query.threads nifi.provenance.repository.index.shard.size using different filesystems (with good IOPS) for directories for provenance repository, content repository, flowfile repository and more
... View more
05-13-2020
04:07 AM
1 Kudo
the point is to correctly set the fqdn no matter how it's done as long as it is correctly configured (it needs to be configured on all hosts in the cluster) ( /etc/sysconfig/network , network manager comands, /etc/hosts (avoiding 127.0.0.1, ::1 <-> fqdn etc... or via another admin tool ) some key services need a correctly set fqdn (kerberos REALM trust for ex will be based on domain thus localy get fqdn . geting host instead of host.domain.etc may cause issues in many hadoop services)
... View more
05-09-2020
12:21 PM
could you confirm you can do queries on any database first and check ambari DB is restored sudo -u postgres psql \l+ ... if ok could you check ambari jdbc connector is updatet ? (using adequate connector path, and if needed download it and install it) ambari-server setup --jdbc-db=postgres --jdbc-driver=/usr/share/java/postgresql96-jdbc.jar in case new jdbc version in installed in same location make sure old one is not already in use ( lsof kill or restart host/vm )
... View more
05-09-2020
02:43 AM
after you renamed old installation pg DIR and created new one using initDb (in old installation DIR) a pg_upgrade step needs to be done (https://www.postgresql.org/docs/9.6/pgupgrade.html) for ex ( with /usr/pgsql-9.6 as your new pgsql9.6 bin DIR installation path and /usr/bin/ as olg binDIR) /usr/pgsql-9.6/bin/pg_upgrade --old-datadir /var/lib/pgsql/data/ --new-datadir /var/lib/pgsql/9.6/data/ --old-bindir /usr/bin/ --new-bindir /usr/pgsql-9.6/bin/ restart postgres after pg_upgrade is done if (psql --version) gives old version update PATH also systemd unit files etc... and pgsql_profile (/var/lib/pgsql/.pgsql_profile) / bash_profile /etc/ files etc...
... View more
05-08-2020
05:54 PM
your problem is probably related to /var/lib/pgsql/<PGSQL_VERSION>/data/pg_hba.conf please check you did not restore pg_hba config file to old postgresql version DIR in /var/lib/psql if it's not the case please try to allow any source connections (to eliminate this possible cause) for ex by adding the following in the end of pg_hba vi /var/lib/pgsql/<NEW_POSTGRES_VERSION>/data/pg_hba.conf host all all 0.0.0.0/0 md5 also check bellow file (for ex if you use many network interfaces in your host/VM): vi /var/lib/pgsql/<NEW_POSTGRES_VERSION>/data/postgresql.conf listen_addresses = '*' and restart postgres
... View more
05-05-2020
01:22 AM
1 Kudo
@Mondi just update dfs.cluster.administrators with the admin username you want in hdfs config (and restart hdfs,yarn,MR2,... services) ex: dfs.cluster.administrators = hdfs,ops you can also use a HDFS administrators group (only one administrator group) using dfs.permissions.superusergroup ex : dfs.permissions.superusergroup = operations to verify config has been updated, once services restarted hdfs getconf -confKey dfs.cluster.administrators or hdfs getconf -confKey dfs.permissions.superusergroup
... View more