Member since
10-14-2015
93
Posts
52
Kudos Received
5
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3265 | 04-20-2016 02:37 PM |
12-08-2016
09:52 AM
We are pleased to announce the release of the Cloudera JDBC Driver for Apache Hive v2.5.18. This release includes the following fixes and enhancements: When using Kerberos authentication, you can now configure the KrbAuthType connection property to specify how the driver obtains the Kerberos Subject. For more information, see the Installation and Configuration Guide. In some cases, when the heap size is restricted, the driver retrieves fewer rows than expected but does not return any exceptions or errors. This issue has been resolved. The driver now retrieves the correct number of rows when the heap size is restricted. See Release Notes for details on all of the fixes and enhancements. Getting started with the Cloudera Drivers: Read the Cloudera JDBC v2.5.18 Driver for Hive Release Notes and Installation Guide. Download the connector from the Cloudera Connectors page. As always, we welcome your feedback. Please send your comments and suggestions to the user group or through our community forums. You can also file bugs through our external JIRA projects on issues.cloudera.org.
... View more
12-01-2016
08:52 AM
We are pleased to announce the release of Hive ODBC v2.5.21, Impala ODBC v2.5.36, and Impala JDBC v2.5.35. Resolved issues in Impala JDBC 2.5.35 The driver does not properly detect a TOP statement and inserts a TOP 0 before it to generates the limit 0 query. This happens if there is an additional space in the query before the original TOP. This issue has been resolved. Queries with YEAR and ISO_WEEK functions failing due to a data type mismatch. This issue has been resolved. Please see below links for details on all of the fixes and enhancements Release Notes Installation Guide Enhancements and new features in Impala ODBC 2.5.36 Support added for Kerberos delegation You now have the option for the driver to forward the user credentials to the server when using Kerberos authentication. Support for Windows Trust Store You now have the option to use the CA certificates in the Windows Trust Store for server verification when using SSL. Auto Reconnect You can configure the driver to automatically attempt reconnection to the Impala server if communications are lost. Support for query, login, and connection timeout The driver now supports the following ODBC API calls: • SQL_ATTR_CONNECTION_TIMEOUT • SQL_ATTR_LOGIN_TIMEOUT • SQL_ATTR_QUERY_TIMEOUT [14244] Optimization of metadata retrieval Resolved issues in Impala ODBC v2.5.36 If a user attempted to connect without proper permissions to access the schema set in the connection string, the driver would fail to connect. The driver will now display a warning under these circumstances, and use the default schema. Specifying an alias for a column name in the SELECT list when SELECT DISTINCT is present and using that column name in the ORDER BY clause causes a system message to be displayed. Driver returning incorrect order for character data types in the SQLGetTypeInfo result set under some circumstances. Please see below links for details on all of the fixes and enhancements Release Notes installation Guide Enhancements and new features in Hive ODBC 2.5.21 Support for the Windows trust store You now have the option to use the CA certificates in the Windows trust store for server verification when using SSL. Auto Reconnect You can now configure the driver to automatically attempt reconnection to the Hive server if communications are lost. Resolved issues in Hive ODBC 2.5.21 Driver fails to connect to the server when using TLS 1.2. When executing a query that contains a CASE statement, the driver replaces the greater than sign (>) with an equal sign (=). Driver does not correctly translate queries that contain the IS NOT NULL operator. Please see below links for details on all of the fixes and enhancements Release Notes Installation Guide Getting started with the Cloudera Drivers: Download the connector from the Cloudera Connectors page. As always, we welcome your feedback. Please send your comments and suggestions to the user group or through our community forums. You can also file bugs through our external JIRA projects on issues.cloudera.org.
... View more
11-29-2016
10:40 AM
3 Kudos
Cloudera is excited to announce the general availability of Cloudera Enterprise 5.9! Cloudera Enterprise 5.9 contains a long list of new features, quality enhancements, bug fixes, and other improvements across the stack. Here is a partial list of those improvements; see the Release Notes for a full list: What's New In CDH 5.9.0 Apache Hadoop You can use temporary credentials to log in to Amazon S3 and obtain temporary credentials from Amazon's Security Token Service (STS). Apache HBase A tool has been added to dump existing replication peers, configurations, and queues when using HBase replication. For more information, see Class DumpReplicationQueues. Metrics have been added that expose the amount of replayed work occurring in the HBase replication system. For more information, see Replication Metrics in the Apache HBase Reference Guide. Apache Hive HIVE-14270 : Added parameters to optimize write performance for Hive tables and partitions that are stored on Amazon S3. See Optimizing Hive Write Performance on Amazon S3. Hue HUE-2915: Integrates Hue with Amazon S3. You can now access both S3 and HDFS in the File Browser, create tables from files in S3, and save query results in S3. See how to Enable S3 Cloud Storage. HUE-4039: Improves SQL Autocompleter. The new Autocompleter understands Hive and Impala SQL dialects and provides smart suggestions based on statement structure and cursor position. See how to manually Enable and Disable Autocompleter. HUE-3877: Adds support for Amazon RDS. You can now deploy Hue against an Amazon RDS database instance with MySQL, PostgreSQL, and Oracle engines. Rebase of Hue on upstream Hue 3.11. Apache Impala (incubating) Performance improvements: [IMPALA-3206] Speedup for queries against DECIMAL columns in Avro tables. [IMPALA-3674] Improved efficiency in LLVM code generation can reduce codegen time, especially for short queries. [IMPALA-2979] Improvements to scheduling on worker nodes, enabled by the REPLICA_PREFERENCEquery option. See REPLICA_PREFERENCE Query Option (CDH 5.9 or higher only) for details. [IMPALA-1683] The REFRESH statement can be applied to a single partition, rather than the entire table. See REFRESH Statement and Refreshing a Single Partition for details. Improvements to the Impala web user interface: [IMPALA-2767] You can now force a session to expire by clicking a link in the web UI, on the /sessions tab. [IMPALA-3715] The /memz tab includes more information about Impala memory usage. [IMPALA-3716] The Details page for a query now includes a Memory tab. [IMPALA-3499] Scalability improvements to the catalog server. [IMPALA-3677] You can send a SIGUSR1 signal to any Impala-related daemon to write a Breakpad minidump. See Breakpad Minidumps for Impala (CDH 5.8 or higher only) for details about the Breakpad minidump feature. [IMPALA-3687] The schema reconciliation rules for Avro tables have changed slightly for CHAR and VARCHAR columns. See Creating Avro Tables for details about column definitions in Avro tables. [IMPALA-3575] Some network operations now have additional timeout and retry settings. Apache Sentry Sentry adds support for securing data on Amazon RDS. As a result, Sentry can secure URIs with an RDS schema. SENTRY-1233 - Logging improvements for SentryConfigToolSolr. SENTRY-1119 - Allow data engines to obtain the ActionFactory directly from the configuration, instead of having hardcoded component-specific classes. SENTRY-1229 - Added a basic configurable cache to SentryGenericProviderBackend. Apache Spark You can now set up AWS credentials for Spark with the Hadoop credential provider, to avoid exposing the AWS secret key in configuration files. Apache Sqoop The mainframe import module extension has been added to support data sets on tape. Cloudera Search The Solr watchdog is now configured to use the fully qualified domain name (FQDN) of the host on which the Solr process is running (instead of 127.0.0.1). You can override this configuration by setting the SOLR_HOSTNAME environment variable to appropriate value. Cloudera Search adds support for index snapshots. For more information, see Backing Up and Restoring Cloudera Search. What's New in Cloudera Manager 5.9.0 You can create virtual images of Cloudera Manager and cluster hosts. See Creating Virtual Images of Cluster Hosts. Security External/Cloud account configuration in Cloudera Manager. Account configuration for access to Amazon Web Services is now available through the centralized UI menu External Accounts. Key Trustee Server rolling restart. Key Trustee Server now supports rolling restart. Backup and Disaster Recovery You can now replicate HDFS files and Hive data to and from an Amazon S3 instance. See HDFS Replication to Amazon S3 and Hive Replication To and From Amazon S3. You can now download performance data about HDFS replication jobs from the Replication Schedules and Replication History pages. See Monitoring the Performance of HDFS Replications. Hive replication now stores Hive UDFs in the Hive metastore. Replication of Impala and Hive User Defined Functions (UDFs). YARN jobs now include the BDR schedule ID that launched the job so you can connect logs with existing schedules, if multiple schedules exist. Resource Management You can create custom Cluster Utilization reports that you can export data from. See Creating a Custom Cluster Utilization Report. When Cloudera Manager manages multiple clusters, Historical Applications and User and Historical Queries by User show applications per user and per pool. Directory usage reports can be exported as a CSV file. Cloudera Manager API Added the update_user() method to the Python API client api_client.py. New API endpoints have been added that allow users to add, list and remove Watched Directories in HDFS service. Logging Kafka log4j log files now include the hostname in the format kafka-broker-${host}.log. Similarly, MirrorMaker logs now include the hostname in the format kafka-mirrormaker-${host}.log. Cloudera Manager displays the History and Rollback support for the Cloudera Manager Settings. This helps Cloudera Support provide better service when certain Cloudera Manager administrative settings are modified. Diagnostic Bundles You can specify information to be redacted in the diagnostic bundle in the UI using Administration > Settings > Redaction Parameters for Diagnostic Bundles. Upgrade Informs you when a simple restart is performed instead of rolling restart on a service because rolling restart is not available. Oozie The Actions menu in the Oozie service has two new commands, Dump Database and Load Database, which make it easier to migrate an Oozie database to another database supported by Oozie. Install Oozie ShareLib command assigns correct permissions to the uploaded libraries. This prevents breaking Oozie workflows with a custom umask setting. Configuration Changes Added the zkClientTimeout parameter for ZooKeeper. Added a new option for setting the file format used by an ApplicationMaster when generating the .jhist file. Adds graceful decommission on YARN NodeManager roles. The NodeManager is not assigned new containers, and it waits for any currently running applications to finish before being decommissioned, unless a timeout occurs. Configure the timeout using the Node Manager Graceful Decommission Timeout configuration property in the YARN Service. stdout and stderr log links are now shown in the UI when a failure occurs while deploying client configurations. Added the configuration parameter, Extra Space Ratio for Indexing, to Reports Manager. Use the parameter to increase indexing speed by allocating additional memory. The default amount of time that HBase Indexer roles attempt to connect to ZooKeeper has been increased from 30 to 60 seconds. Cloudera Manager can identify whether or not a customer is using the embedded PostgreSQL database. If Cloudera Manager is configured to use the embedded PostgreSQL database, a yellow banner appears in the UI, recommending that you upgrade to a supported external database. When Impala uses SSL, TLS Connection to Catalog Server is now supported. You can enable replication for any Impala UDFs/Metadata (in Hive Replication). When running wizards from the Cloudera Manager Admin Console that add a cluster, add a service, perform an upgrade, and other tasks, steps do not display when they are not reachable or do not apply to the current configuration. Improve Cloudera Manager provisioning performance on AWS. Add support for resetting Cloudera Manager GUID/UUID by checking the UUID file. Over the next few weeks, we will publish blog posts that cover some of these features in detail. In the meantime: Download Cloudera Enterprise 5.9 Explore documentation As always, we value your feedback; please provide any comments and suggestions through our community forums. You can also file bugs through issues.cloudera.org.
... View more
11-09-2016
09:00 PM
We are pleased to announce the general availability of the Cloudera Connector for Netezza 1.4c5. New Features in Cloudera Connector for Netezza 1.4c5 Added the escapeColName method to handle, for example, column names with lowercase characters. For more details on new features and usage of Cloudera Connector for Netezza, see: Release Notes Cloudera Connector for Netezza Version 1.4c5 Cloudera Connector for Netezza User Guide, version 1.4c5 As always, we welcome your feedback. Please send your comments and suggestions through our new community forums. You can also file bugs in the CDH project at issues.cloudera.org.
... View more
10-13-2016
02:12 PM
We are pleased to announce the general availability of the Cloudera Connector Powered by Teradata 1.6c5. New Features in Cloudera Connector Powered by Teradata Version 1.6c5 Upgrades the JDBC driver to version 15.10.00.22 and the TDCH library to version 1.5.0. These libraries contain several bug fixes and improvements. Adds the --schema argument, used to override the <td-instance> value in the connection string of the Sqoop command. For example, if the connection string in the Sqoop command is jdbc:teradata:// <td-host>/DATABASE=database1, but you specify --schema database2, your data is imported from database2 and not database1. If the connection string does not contain the DATABASE parameter — for example jdbc:teradata://<td-host>/CHARSET=UTF8) — you can also use the --schema databasename argument to have Sqoop behave as if you specified the jdbc:teradata://<td-host>/DATABASE=databasename,CHARSET=UTF8 connection string. For more details on new features and usage of Cloudera Connector Powered by Teradata, see Download Cloudera Connector Powered by Teradata version 1.6c5 Release Notes Cloudera Connector Powered by Teradata version 1.6c5 Cloudera Connector Powered by Teradata User Guide, version 1.6c5 As always, we welcome your feedback. Please send your comments and suggestions through our community forums. You can also file bugs in issues.cloudera.org.
... View more
10-11-2016
07:14 AM
5 Kudos
Cloudera is pleased to announce a new minor release of our supported packaging of Apache Accumulo for use on CDH 5. This release adds support for: Kerberos authentication for Accumulo Clients RHEL 7 / CentOS 7 support Running Accumulo on top of Apache HDFS Transparent Encryption AccumuloStorgeHandler for Apache Hive access to data stored in Accumulo Apache Spark processing of data stored in Accumulo through AccumuloInputFormat and AccumuloOutputFormat Users are cautioned to read the installation guide for known issues and unsupported upstream features. Existing users should see a new parcel available for Accumulo 1.7.2-cdh5.5.0. This release is supported with Cloudera Manager and CDH versions 5.5.0 or later. Cloudera recommends using the latest version of Cloudera Manager available. Docs: http://tiny.cloudera.com/accumulo-docs Cloudera Manager 5.8.2 Download: http://tiny.cloudera.com/cm-5.8.2 Accumulo 1.7.2 on CDH 5 (CDH 5.5.0 or later): http://archive.cloudera.com/accumulo-c5/ List of JIRAs included in addition to Apache Accumulo 1.7.2 release: http://tiny.cloudera.com/accumulo-1.7.2-cdh5.5.0-release-notes
... View more
10-04-2016
04:26 PM
We are pleased to announce the release of Cloudera Enterprise 5.7.4 (CDH 5.7.4, Cloudera Manager 5.7.4, Cloudera Navigator 2.6.4, and Key Trustee KMS 5.7.4). This release fixes key bugs and includes the following: CDH fixes for the following issues: HADOOP-8436 - NPE In getLocalPathForWrite ( path, conf ) when the required context item is not configured HDFS-8269 - getBlockLocations() does not resolve the .reserved path and generates incorrect edit logs when updating the atime HDFS-9781 - FsDatasetImpl#getBlockReports can occasionally throw NullPointerException MAPREDUCE-6628 - Potential memory leak in CryptoOutputStream HBASE-16284 - Unauthorized client can shutdown the cluster HIVE-14436 - Hive 1.2.1/Hitting "ql.Driver: FAILED: IllegalArgumentException Error" HIVE-14743 - ArrayIndexOutOfBoundsException - HBASE-backed views query with JOINs IMPALA-3682 - Do not retry unrecoverable socket creation errors IMPALA-4020 - Handle external conflicting changes to HMS gracefully SPARK-8428 -Fix integer overflows in TimSort For a full list of upstream JIRAs fixed in CDH 5.7.4, see the issues fixed section of the Release Notes. Cloudera Manager fixes for the following issues: Agent orphan cleanup removes process dir from in flight process Host inspector incorrectly warns about kernel version "2.6.32-504.16.2" If total_space_bytes is really big, heartbeats fail Oozie points to older sharelib even after running sharelib install command OOMKiller script does not work for Impala Catalog For a full list of issues fixed in Cloudera Manager 5.7.4, see the issues fixed section of the Release Notes. Issues Fixed in Key Trustee KMS KMS does not failover to Passive Key Trustee Server in some network failure scenarios: In some situations, if the Active Key Trustee Server is unreachable on the network, Key Trustee KMS does not fail over to the Passive Key Trustee Server. For a full list of issues fixed in Key Trustee KMS 5.7.4, see the issues fixed section of the Release Notes. We look forward to you trying it, using the information below: Download Cloudera Enterprise View the documentation As always, we are happy to hear your feedback. Please send your comments and suggestions to the user group or through our community forums. You can also file bugs through our external JIRA projects on issues.cloudera.org.
... View more
09-30-2016
04:33 PM
We are pleased to announce the release of Impala JDBC v2.5.34 driver. This release has the following fixes and enhancements: Enhancements and new features: Authentication method and transport protocol. You can now specify if SASL should be used in conjunction with the 'Username and Password' Authentication Mechanism (AuthMech=3) by using the Use Sasl connection property. Previously this was done through the transport Mode connection property, which has now been removed. See Release Notes for JDBC for details on all of the fixes. Getting Started with the Cloudera Drivers Read the Cloudera JDBC 2.5.34 Driver for Impala Release Notes and Installation Guide. Download the connector from the Cloudera Connectors page. As always, we welcome your feedback. Please send your comments and suggestions to the user group or through our community forums. You can also file bugs through our external JIRA projects on issues.cloudera.org.
... View more
Labels:
- Labels:
-
Impala
09-20-2016
11:51 AM
Cloudera is happy to announce the availability of parcels and packages for Kudu 1.0.0. After approximately a year of beta releases, Apache Kudu has reached version 1.0.0. This version number signifies that the development team feels that Kudu is stable enough for usage in production environments.This is the first non-beta release of the Apache Kudu project. (Although because Kudu is not currently integrated into CDH, it is not yet an officially supported CDH component.) Kudu 1.0.0 delivers a number of new features, bug fixes, and optimizations. We are also releasing a refresh of the Impala Kudu parcel. To upgrade Kudu to 1.0.0, see Upgrade Parcels or Upgrade Packages. In addition, you can read more about Kudu 1.0.0 in the vision blog post we just published. New Features in Kudu 1.0.0 Removal of multiversion concurrency control (MVCC) history is now supported. This is known as tablet history GC. This allows Kudu to reclaim disk space, where previously Kudu would keep a full history of all changes made to a given table since the beginning of time. Previously, the only way to reclaim disk space was to drop a table. Kudu will still keep historical data, and the amount of history retained is controlled by setting the configuration flag --tablet_history_max_age_sec, which defaults to 15 minutes (expressed in seconds). The timestamp represented by the current time minus tablet_history_max_age_sec is known as the ancient history mark (AHM). When a compaction or flush occurs, Kudu will remove the history of changes made prior to the ancient history mark. This only affects historical data; currently-visible data will not be removed. A specialized maintenance manager background task to remove existing "cold"historical data that is not in a row affected by the normal compaction process will be added in a future release. Most of Kudu’s command line tools have been consolidated under a new top-level kudu tool. This reduces the number of large binaries distributed with Kudu and also includes much-improved help output. The Kudu Flume Sink now supports processing events containing Avro-encoded records, using the newAvroKuduOperationsProducer. Administrative tools including kudu cluster ksck now support running against multi-master Kudu clusters. The output of the ksck tool is now colorized and much easier to read. The C++ client API now supports writing data in AUTO_FLUSH_BACKGROUND mode. This can provide higher throughput for ingest workloads. Performance The performance of comparison predicates on dictionary-encoded columns has been substantially optimized. Users are encouraged to use dictionary encoding on any string or binary columns with low cardinality, especially if these columns will be filtered with predicates. The Java client is now able to prune partitions from scanners based on the provided predicates. For example, an equality predicate on a hash-partitioned column will now only access those tablets that could possibly contain matching data. This is expected to improve performance for the Spark integration as well as applications using the Java client API. The performance of compaction selection in the tablet server has been substantially improved. This can increase the efficiency of the background maintenance threads and improve overall throughput of heavy write workloads. The policy by which the tablet server retains write-ahead log (WAL) files has been improved so that it takes into account other replicas of the tablet. This should help mitigate the spurious eviction of tablet replicas on machines that temporarily lag behind the other replicas. Wire protocol compatibility Kudu 1.0.0 maintains client-server wire-compatibility with previous releases. Applications using the Kudu client libraries may be upgraded either before, at the same time, or after the Kudu servers. Kudu 1.0.0 does not maintain server-server wire compatibility with previous releases. Therefore, rolling upgrades between earlier versions of Kudu and Kudu 1.0.0 are not supported. Incompatible Changes in Kudu 1.0.0 Command line tools The kudu-pbc-dump tool has been removed. The same functionality is now implemented as kudu pbc dump. The kudu-ksck tool has been removed. The same functionality is now implemented as kudu cluster ksck. The cfile-dump tool has been removed. The same functionality is now implemented as kudu fs cfile dump. The log-dump tool has been removed. The same functionality is now implemented as kudu wal dumpand kudu local_replica dump wals. The kudu-admin tool has been removed. The same functionality is now implemented within kudu tableand kudu tablet. The kudu-fs_dump tool has been removed. The same functionality is now implemented as kudu fs dump. The kudu-ts-cli tool has been removed. The same functionality is now implemented within kudu master, kudu remote_replica, and kudu tserver. The kudu-fs_list tool has been removed and some similar useful functionality has been moved underkudu local_replica. Configuration flags Some configuration flags are now marked as "unsafe" and "experimental". Such flags are disallowed by default. Users may access these flags by enabling the additional flags --unlock_unsafe_flags and --unlock_experimental_flags. Usage of such flags is not recommended, as the flags may be removed or modified with no deprecation period and without notice in future Kudu releases. Client APIs (C++/Java/Python) The TIMESTAMP column type has been renamed to UNIXTIME_MICROS in order to reduce confusion between Kudu’s timestamp support and the timestamps supported by other systems such as Apache Hive and Apache Impala (incubating). Existing tables will automatically be updated to use the new name for the type. Clients upgrading to the new client libraries must move to the new name for the type. Clients using old client libraries will continue to operate using the old type name, even when connected to clusters that have been upgraded. Similarly, if clients are upgraded before servers, existing timestamp columns will be available using the new type name. KuduSession methods in the C++ library are no longer advertised as thread-safe to have one set of semantics for both C++ and Java Kudu client libraries. The KuduScanToken::TabletServers method in the C++ library has been removed. The same information can now be found in the KuduScanToken::tablet method. Apache Flume Integration The KuduEventProducer interface used to process Flume events into Kudu operations for the Kudu Flume Sink has changed, and has been renamed KuduOperationsProducer. The existing KuduEventProducers have been updated for the new interface, and have been renamed similarly. Known Issues and Limitations of Kudu 1.0.0 Schema and Usage Limitations Kudu is primarily designed for analytic use cases. You are likely to encounter issues if a single row contains multiple kilobytes of data. The columns which make up the primary key must be listed first in the schema. Key columns cannot be altered. You must drop and recreate a table to change its keys. Key columns must not be null. Columns with DOUBLE, FLOAT, or BOOL types are not allowed as part of a primary key definition. Type and nullability of existing columns cannot be changed by altering the table. A table's primary key cannot be changed. Dropping a column does not immediately reclaim space. Compaction must run first. There is no way to run compaction manually, but dropping the table will reclaim the space immediately. Partitioning Limitations Tables must be manually pre-split into tablets using simple or compound primary keys. Automatic splitting is not yet possible. Range partitions may be added or dropped after a table has been created. See Schema Design for more information. Data in existing tables cannot currently be automatically repartitioned. As a workaround, create a new table with the new partitioning and insert the contents of the old table. Replication and Backup Limitations Kudu does not currently include any built-in features for backup and restore. Users are encouraged to use tools such as Spark or Impala to export or import tables as necessary. Impala Limitations To use Kudu with Impala, you must install a special release of Impala called Impala_Kudu. Obtaining and installing a compatible Impala release is detailed in Installing and Using Apache Impala (incubating) With Apache Kudu. To use Impala_Kudu alongside an existing Impala instance, you must install using parcels. Updates, inserts, and deletes via Impala are non-transactional. If a query fails part of the way through, its partial effects will not be rolled back. All queries will be distributed across all Impala hosts which host a replica of the target table(s), even if a predicate on a primary key could correctly restrict the query to a single tablet. This limits the maximum concurrency of short queries made via Impala. No TIMESTAMP and DECIMAL type support. (The underlying Kudu type formerly known as TIMESTAMP has been renamed to UNIXTIME_MICROS; currently, there is no Impala-compatible TIMESTAMP type.) The maximum parallelism of a single query is limited to the number of tablets in a table. For good analytic performance, aim for 10 or more tablets per host or use large tables. Impala is only able to push down predicates involving =, <=, >=, or BETWEEN comparisons between any column and a literal value, and < and > for integer columns only. For example, for a table with an integer key ts, and a string key name, the predicate WHERE ts >= 12345 will convert into an efficient range scan, whereas where name > 'lipcon' will currently fetch all data from the table and evaluate the predicate within Impala. Security Limitations Authentication and authorization features are not implemented. Data encryption is not built in. Kudu has been reported to run correctly on systems using local block device encryption (e.g. dmcrypt). Client and API Limitations ALTER TABLE is not yet fully supported via the client APIs. More ALTER TABLE operations will become available in future releases. Other Known Issues The following are known bugs and issues with the current release of Kudu. They will be addressed in later releases. Note that this list is not exhaustive, and is meant to communicate only the most important known issues. If the Kudu master is configured with the -log_fsync_all option, tablet servers and clients will experience frequent timeouts, and the cluster may become unusable. If a tablet server has a very large number of tablets, it may take several minutes to start up. It is recommended to limit the number of tablets per server to 100 or fewer. Consider this limitation when pre-splitting your tables. If you notice slow start-up times, you can monitor the number of tablets per server in the web UI. Due to a known bug in Linux kernels prior to 3.8, running Kudu on ext4 mount points may cause a subsequent fsck to fail with errors such as Logical start <N> does not match logical start <M> at next level . These errors are repairable using fsck -y, but may impact server restart time. This affects RHEL/CentOS 6.8 and below. A fix is planned for RHEL/CentOS 6.9. RHEL 7.0 and higher are not affected. Ubuntu 14.04 and later are not affected. SLES 12 and later are not affected. Issues Fixed in Kudu 1.0.0 See Issues resolved for Kudu 1.0.0 and Git changes between 0.10.0 and 1.0.0. For a complete list of new features, changes, bug fixes, and known issues, see the Kudu 1.0.0 Release Notes . As always, your feedback is appreciated. For general Kudu questions, visit the community page. If you have any questions related to the Kudu packages provided by Cloudera, including installation or configuration using Cloudera Manager, visit the Cloudera Community Forum.
... View more
09-07-2016
04:18 PM
1 Kudo
We are pleased to announce the release of CDH 5.7.3 This release fixes key bugs and includes the following: CDH fixes for the following issues: FLUME-2821 - KafkaSourceUtil Can Log Passwords at Info remove logging of security related data in older releases. HADOOP-11361 - Fix a race condition in MetricsSourceAdapter.updateJmxCache. HDFS-8581 - ContentSummary on / skips further counts on yielding lock. MAPREDUCE-6675 - TestJobImpl.testUnusableNode failed. YARN-5048 - DelegationTokenRenewer#skipTokenRenewal may throw NPE. HBASE-14818 - user_permission does not list namespace permissions. HIVE-11980 - Follow up on HIVE-11696, exception is thrown from CTAS from the table with table-level serde is Parquet while partition-level serde is JSON. HIVE-13749 - Memory leak in Hive Metastore. HUE-4477 - [security] Select All is not filtering out the non visible roles from the selection . IMPALA-3441 - , IMPALA-3659: check for malformed Avro data. SENTRY-1252 - grantServerPrivilege and revokeServerPrivilege should treat "*" and "ALL" as synonyms when action is not explicitly specified. For a full list of upstream JIRAs fixed in CDH 5.7.3, see the issues fixed section of the Release Notes. We look forward to you trying it, using the information below: Download CDH View the documentation As always, we are happy to hear your feedback. Please send your comments and suggestions to the user group or through our community forums. You can also file bugs through our external JIRA projects on issues.cloudera.org.
... View more