Back to Hadoop

Licensed to the Apache Software Foundation (ASF) under one

hadoop-common-project/hadoop-common/src/site/markdown/release/2.9.0/RELEASENOTES.2.9.0.md

2.0.5-alpha-rc231.0 KB
Original Source
<!--- # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -->

Apache Hadoop 2.9.0 Release Notes

These release notes cover new developer and user-facing incompatibilities, important issues, features, and major improvements.


  • HDFS-9624 | Major | DataNode start slowly due to the initial DU command operations

Make it configurable how long the cached du file is valid. Useful for rolling upgrade.


  • HDFS-9525 | Blocker | hadoop utilities need to support provided delegation tokens

If hadoop.token.files property is defined and configured to one or more comma-delimited delegation token files, Hadoop will use those token files to connect to the services as named in the token.


  • YARN-4762 | Blocker | NMs failing on DelegatingLinuxContainerRuntime init with LCE on

Fixed CgroupHandler's creation and usage to avoid NodeManagers crashing when LinuxContainerExecutor is enabled.


  • HDFS-1477 | Major | Support reconfiguring dfs.heartbeat.interval and dfs.namenode.heartbeat.recheck-interval without NN restart

Steps to reconfigure:

  1. change value of the parameter in corresponding xml configuration file
  2. to reconfigure, run hdfs dfsadmin -reconfig namenode <nn_addr>:<ipc_port> start
  3. to check status of the most recent reconfigure operation, run hdfs dfsadmin -reconfig namenode <nn_addr>:<ipc_port> status
  4. to query a list reconfigurable properties on NN, run hdfs dfsadmin -reconfig namenode <nn_addr>:<ipc_port> properties

  • YARN-4732 | Trivial | *ProcessTree classes have too many whitespace issues

  • HDFS-9349 | Major | Support reconfiguring fs.protected.directories without NN restart

Steps to reconfigure:

  1. change value of the parameter in corresponding xml configuration file
  2. to reconfigure, run hdfs dfsadmin -reconfig namenode <nn_addr>:<ipc_port> start
  3. to check status of the most recent reconfigure operation, run hdfs dfsadmin -reconfig namenode <nn_addr>:<ipc_port> status
  4. to query a list reconfigurable properties on NN, run hdfs dfsadmin -reconfig namenode <nn_addr>:<ipc_port> properties

  • YARN-4784 | Major | Fairscheduler: defaultQueueSchedulingPolicy should not accept FIFO

Clusters cannot use FIFO policy as the defaultQueueSchedulingPolicy. Clusters with a single level of queues will have to explicitly set the policy to FIFO if that is desired.


  • HADOOP-10694 | Major | Remove synchronized input streams from Writable deserialization

Remove invisible synchronization primitives from DataInputBuffer


  • HADOOP-12782 | Major | Faster LDAP group name resolution with ActiveDirectory

If the user object returned by LDAP server has the user's group object DN (supported by Active Directory), Hadoop can reduce LDAP group mapping latency by setting hadoop.security.group.mapping.ldap.search.attr.memberof to memberOf.


  • HDFS-10328 | Minor | Add per-cache-pool default replication num configuration

Add per-cache-pool default replication num configuration


  • YARN-2928 | Critical | YARN Timeline Service v.2: alpha 1

We are introducing an early preview (alpha 1) of a major revision of YARN Timeline Service: v.2. YARN Timeline Service v.2 addresses two major challenges: improving scalability and reliability of Timeline Service, and enhancing usability by introducing flows and aggregation.

YARN Timeline Service v.2 alpha 1 is provided so that users and developers can test it and provide feedback and suggestions for making it a ready replacement for Timeline Service v.1.x. It should be used only in a test capacity. Most importantly, security is not enabled. Do not set up or use Timeline Service v.2 until security is implemented if security is a critical requirement.

More details are available in the YARN Timeline Service v.2 documentation.


  • HADOOP-13354 | Major | Update WASB driver to use the latest version (4.2.0) of SDK for Microsoft Azure Storage Clients

The WASB FileSystem now uses version 4.2.0 of the Azure Storage SDK.


  • HADOOP-13403 | Major | AzureNativeFileSystem rename/delete performance improvements

WASB has added an optional capability to execute certain FileSystem operations in parallel on multiple threads for improved performance. Please refer to the Azure Blob Storage documentation page for more information on how to enable and control the feature.


  • HADOOP-12747 | Major | support wildcard in libjars argument

It is now possible to specify multiple jar files for the libjars argument using a wildcard. For example, you can specify "-libjars 'libs/*'" as a shorthand for all jars in the libs directory.


  • YARN-5137 | Major | Make DiskChecker pluggable in NodeManager

Added new plugin property yarn.nodemanager.disk-validator to allow the NodeManager to use an alternate class for checking whether a disk is good or not.


  • HDFS-8312 | Critical | Trash does not descent into child directories to check for permissions

Permissions are now checked when moving a file to Trash.


  • HADOOP-13588 | Major | ConfServlet should respect Accept request header

Conf HTTP service should set response's content type according to the Accept header in the request.


  • MAPREDUCE-6776 | Major | yarn.app.mapreduce.client.job.max-retries should have a more useful default

The default value of yarn.app.mapreduce.client.job.max-retries has been changed from 0 to 3. This will help protect clients from failures that are transient. True failures may take slightly longer now due to the retries.


  • HADOOP-13522 | Major | Add %A and %a formats for fs -stat command to print permissions

Added permissions to the fs stat command. They are now available as symbolic (%A) and octal (%a) formats, which are in line with Linux.


  • YARN-5388 | Critical | Deprecate and remove DockerContainerExecutor

DockerContainerExecutor is deprecated starting 2.9.0 and removed from 3.0.0. Please use LinuxContainerExecutor with the DockerRuntime to run Docker containers on YARN clusters.


  • HDFS-10756 | Major | Expose getTrashRoot to HTTPFS and WebHDFS

"getTrashRoot" returns a trash root for a path. Currently in DFS if the path "/foo" is a normal path, it returns "/user/$USER/.Trash" for "/foo" and if "/foo" is an encrypted zone, it returns "/foo/.Trash/$USER" for the child file/dir of "/foo". This patch is about to override the old "getTrashRoot" of httpfs and webhdfs, so that the behavior of returning trash root in httpfs and webhdfs are consistent with DFS.


We are sorry for causing pain for everyone for whom this Jackson update causes problems, but it was proving impossible to stay on the older version: too much code had moved past it, and by staying back we were limiting what Hadoop could do, and giving everyone who wanted an up to date version of Jackson a different set of problems. We've selected Jackson 2.7.8 as it fixed fix a security issue in XML parsing, yet proved compatible at the API level with the Hadoop codebase --and hopefully everything downstream.


The dependency on the AWS SDK has been bumped to 1.11.45.


  • HADOOP-13953 | Major | Make FTPFileSystem's data connection mode and transfer mode configurable

Added two configuration key fs.ftp.data.connection.mode and fs.ftp.transfer.mode, and configure FTP data connection mode and transfer mode accordingly.


  • HADOOP-14003 | Major | Make additional KMS tomcat settings configurable
<!-- markdown -->

The KMS can now be configured with the additional environment variables KMS_PROTOCOL, KMS_ACCEPT_COUNT, and KMS_ACCEPTOR_THREAD_COUNT. See kms-env.sh for more information about these variables.


  • MAPREDUCE-6404 | Major | Allow AM to specify a port range for starting its webapp

Add a new configuration - "yarn.app.mapreduce.am.webapp.port-range" to specify port-range for webapp launched by AM.


  • HADOOP-13119 | Major | Add ability to secure log servlet using proxy users

WARNING: No release note provided for this change.


  • HADOOP-13075 | Major | Add support for SSE-KMS and SSE-C in s3a filesystem

The new encryption options SSE-KMS and especially SSE-C must be considered experimental at present. If you are using SSE-C, problems may arise if the bucket mixes encrypted and unencrypted files. For SSE-KMS, there may be extra throttling of IO, especially with the fadvise=random option. You may wish to request an increase in your KMS IOPs limits.


  • YARN-6177 | Major | Yarn client should exit with an informative error message if an incompatible Jersey library is used at client

Let yarn client exit with an informative error message if an incompatible Jersey library is used from client side.


  • HADOOP-13817 | Minor | Add a finite shell command timeout to ShellBasedUnixGroupsMapping

A new introduced configuration key "hadoop.security.groups.shell.command.timeout" allows applying a finite wait timeout over the 'id' commands launched by the ShellBasedUnixGroupsMapping plugin. Values specified can be in any valid time duration units: https://hadoop.apache.org/docs/current/api/org/apache/hadoop/conf/Configuration.html#getTimeDuration-java.lang.String-long-java.util.concurrent.TimeUnit-

Value defaults to 0, indicating infinite wait (preserving existing behaviour).


  • HADOOP-6801 | Minor | io.sort.mb and io.sort.factor were renamed and moved to mapreduce but are still in CommonConfigurationKeysPublic.java and used in SequenceFile.java

Two new configuration keys, seq.io.sort.mb and seq.io.sort.factor have been introduced for the SequenceFile's Sorter feature to replace older, deprecated property keys of io.sort.mb and io.sort.factor.

This only affects direct users of the org.apache.hadoop.io.SequenceFile.Sorter Java class. For controlling MR2's internal sorting instead, use the existing config keys of mapreduce.task.io.sort.mb and mapreduce.task.io.sort.factor.


  • HDFS-11499 | Major | Decommissioning stuck because of failing recovery

Allow a block to complete if the number of replicas on live nodes, decommissioning nodes and nodes in maintenance mode satisfies minimum replication factor. The fix prevents block recovery failure if replica of last block is being decommissioned. Vice versa, the decommissioning will be stuck, waiting for the last block to be completed. In addition, file close() operation will not fail due to last block being decommissioned.


  • HADOOP-14213 | Major | Move Configuration runtime check for hadoop-site.xml to initialization

Move the check for hadoop-site.xml to static initialization of the Configuration class.


<!-- markdown -->
  • Properties {{dfs.adls.}} are renamed {{fs.adl.}}
  • Property {{adl.dfs.enable.client.latency.tracker}} is renamed {{adl.enable.client.latency.tracker}}
  • Old properties are still supported

  • HADOOP-14174 | Major | Set default ADLS access token provider type to ClientCredential

Switch the default ADLS access token provider type from Custom to ClientCredential.


  • HADOOP-14301 | Major | Deprecate SharedInstanceProfileCredentialsProvider in branch-2.

SharedInstanceProfileCredentialsProvider has been deprecated. Users should use InstanceProfileCredentialsProvider provided by AWS SDK instead, which itself enforces a singleton instance to reduce calls to AWS EC2 Instance Metadata Service. SharedInstanceProfileCredentialsProvider will be removed permanently in a future release.


  • HADOOP-11794 | Major | Enable distcp to copy blocks in parallel

If a positive value is passed to command line switch -blocksperchunk, files with more blocks than this value will be split into chunks of `<blocksperchunk>` blocks to be transferred in parallel, and reassembled on the destination. By default, `<blocksperchunk>` is 0 and the files will be transmitted in their entirety without splitting. This switch is only applicable when both the source file system supports getBlockLocations and target supports concat.


  • HDFS-11402 | Major | HDFS Snapshots should capture point-in-time copies of OPEN files

When the config param "dfs.namenode.snapshot.capture.openfiles" is enabled, HDFS snapshots taken will additionally capture point-in-time copies of the open files that have valid leases. Even when the current version open files grow or shrink in size, the snapshot will always retain the immutable versions of these open files, just as in for all other closed files. Note: The file length captured for open files in the snapshot was the one recorded in NameNode at the time of snapshot and it may be shorter than what the client has written till then. In order to capture the latest length, the client can call hflush/hsync with the flag SyncFlag.UPDATE_LENGTH on the open files handles.


  • YARN-2962 | Critical | ZKRMStateStore: Limit the number of znodes under a znode

WARNING: No release note provided for this change.


  • HDFS-9016 | Major | Display upgrade domain information in fsck

New fsck option "-upgradedomains" has been added to display upgrade domains of any block.


  • HADOOP-14419 | Minor | Remove findbugs report from docs profile

Findbugs report is no longer part of the documentation.


  • HDFS-11661 | Blocker | GetContentSummary uses excessive amounts of memory

Reverted HDFS-10797 to fix a scalability regression brought by the commit.


  • HADOOP-14407 | Major | DistCp - Introduce a configurable copy buffer size

The copy buffer size can be configured via the new parameter <copybuffersize>. By default the <copybuffersize> is set to 8KB.


  • YARN-6127 | Major | Add support for work preserving NM restart when AMRMProxy is enabled

This breaks rolling upgrades because it changes the major version of the NM state store schema. Therefore when a new NM comes up on an old state store it crashes.

The state store versions for this change have been updated in YARN-6798.


  • HADOOP-14536 | Major | Update azure-storage sdk to version 5.3.0

The WASB FileSystem now uses version 5.3.0 of the Azure Storage SDK.


  • HADOOP-14546 | Major | Azure: Concurrent I/O does not work when secure.mode is enabled

Fix to wasb:// (Azure) file system that allows the concurrent I/O feature to be used with the secure mode feature.


  • HADOOP-14535 | Major | wasb: implement high-performance random access and seek of block blobs

Random access and seek improvements for the wasb:// (Azure) file system.


  • YARN-6798 | Major | Fix NM startup failure with old state store due to version mismatch
<!-- markdown -->

This fixes the LevelDB state store for the NodeManager. As of this patch, the state store versions now correspond to the following table.

  • Previous Patch: YARN-5049
    • LevelDB Key: queued
    • Hadoop Versions: 2.9.0, 3.0.0-alpha1
    • Corresponding LevelDB Version: 1.2
  • Previous Patch: YARN-6127
    • LevelDB Key: AMRMProxy/NextMasterKey
    • Hadoop Versions: 2.9.0, 3.0.0-alpha4
    • Corresponding LevelDB Version: 1.1

  • HADOOP-14539 | Major | Move commons logging APIs over to slf4j in hadoop-common

In Hadoop common, fatal log level is changed to error because slf4j API does not support fatal log level.


  • HADOOP-14518 | Minor | Customize User-Agent header sent in HTTP/HTTPS requests by WASB.

WASB now includes the current Apache Hadoop version in the User-Agent string passed to Azure Blob service. Users also may include optional additional information to identify their application. See the documentation of configuration property fs.wasb.user.agent.id for further details.


  • HADOOP-14722 | Major | Azure: BlockBlobInputStream position incorrect after seek

Bug fix to Azure Filesystem related to HADOOP-14535.


  • HADOOP-14680 | Minor | Azure: IndexOutOfBoundsException in BlockBlobInputStream

Bug fix to Azure Filesystem related to HADOOP-14535


  • HDFS-10326 | Major | Disable setting tcp socket send/receive buffers for write pipelines

The size of the TCP socket buffers are no longer hardcoded by default. Instead the OS now will automatically tune the size for the buffer.


  • HADOOP-14260 | Major | Configuration.dumpConfiguration should redact sensitive information
<!-- markdown -->

Configuration.dumpConfiguration no longer prints out the clear text values for the sensitive keys listed in hadoop.security.sensitive-config-keys. Callers can override the default list of sensitive keys either to redact more keys or print the clear text values for a few extra keys for debugging purpose.


  • HADOOP-14660 | Major | wasb: improve throughput by 34% when account limit exceeded

Up to 34% throughput improvement for the wasb:// (Azure) file system when fs.azure.selfthrottling.enable is false fs.azure.autothrottling.enable is true.


  • HADOOP-14769 | Major | WASB: delete recursive should not fail if a file is deleted

Recursive directory delete improvement for the wasb filesystem.


  • YARN-5355 | Critical | YARN Timeline Service v.2: alpha 2

We are releasing the alpha2 version of a major revision of YARN Timeline Service: v.2. YARN Timeline Service v.2 addresses two major challenges: improving scalability and reliability of Timeline Service, and enhancing usability by introducing flows and aggregation.

YARN Timeline Service v.2 alpha1 was introduced in 3.0.0-alpha1 via YARN-2928.

YARN Timeline Service v.2 alpha2 is now being provided so that users and developers can test it and provide feedback and suggestions for making it a ready replacement for Timeline Service v.1.x. Security is provided via Kerberos Authentication and delegation tokens. There is also a simple read level authorization provided via whitelists.

Some of the notable improvements since alpha-1 are:

  • Security via Kerberos Authentication and delegation tokens
  • Read side simple authorization via whitelist
  • Client configurable entity sort ordering
  • New REST APIs for apps, app attempts, containers, fetching metrics by timerange, pagination, sub-app entities
  • Support for storing sub-application entities (entities that exist outside the scope of an application)
  • Configurable TTLs (time-to-live) for tables, configurable table prefixes, configurable hbase cluster
  • Flow level aggregations done as dynamic (table level) coprocessors
  • Uses latest stable HBase release 1.2.6

More details are available in the YARN Timeline Service v.2 documentation.


  • HADOOP-13345 | Major | S3Guard: Improved Consistency for S3A

S3Guard (pronounced see-guard) is a new feature for the S3A connector to Amazon S3, which uses DynamoDB for a high performance and consistent metadata repository. Essentially: S3Guard caches directory information, so your S3A clients get faster lookups and resilience to inconsistency between S3 list operations and the status of objects. When files are created, with S3Guard, they'll always be found.

S3Guard does not address update consistency: if a file is updated, while the directory information will be updated, calling open() on the path may still return the old data. Similarly, deleted objects may also potentially be opened.

Please consult the S3Guard documentation in the Amazon S3 section of our documentation.

Note: part of this update includes moving to a new version of the AWS SDK 1.11, one which includes the Dynamo DB client and its a shaded version of Jackson 2. The large aws-sdk-bundle JAR is needed to use the S3A client with or without S3Guard enabled. The good news: because Jackson is shaded, there will be no conflict between any Jackson version used in your application and that which the AWS SDK needs.


  • HADOOP-14520 | Major | WASB: Block compaction for Azure Block Blobs

Block Compaction for Azure Block Blobs. When the number of blocks in a block blob is above 32000, the process of compaction replaces a sequence of small blocks with with one big block.


  • HDFS-11799 | Major | Introduce a config to allow setting up write pipeline with fewer nodes than replication factor

Added new configuration "dfs.client.block.write.replace-datanode-on-failure.min-replication".

The minimum number of replications that are needed to not to fail
  the write pipeline if new datanodes can not be found to replace
  failed datanodes (could be due to network failure) in the write pipeline.
  If the number of the remaining datanodes in the write pipeline is greater
  than or equal to this property value, continue writing to the remaining nodes.
  Otherwise throw exception.

  If this is set to 0, an exception will be thrown, when a replacement
  can not be found.

  • YARN-2915 | Major | Enable YARN RM scale out via federation using multiple RM's

A federation-based approach to transparently scale a single YARN cluster to tens of thousands of nodes, by federating multiple YARN standalone clusters (sub-clusters). The applications running in this federated environment will see a single massive YARN cluster and will be able to schedule tasks on any node of the federated cluster. Under the hood, the federation system will negotiate with sub-clusters ResourceManagers and provide resources to the application. The goal is to allow an individual job to “span” sub-clusters seamlessly.


  • YARN-1492 | Major | truly shared cache for jars (jobjar/libjar)

The YARN Shared Cache provides the facility to upload and manage shared application resources to HDFS in a safe and scalable manner. YARN applications can leverage resources uploaded by other applications or previous runs of the same application without having to re-­upload and localize identical files multiple times. This will save network resources and reduce YARN application startup time.


  • HDFS-10467 | Major | Router-based HDFS federation

HDFS Router-based Federation adds a RPC routing layer that provides a federated view of multiple HDFS namespaces. This is similar to the existing ViewFS and HDFS federation functionality, except the mount table is managed on the server-side by the routing layer rather than on the client. This simplifies access to a federated cluster for existing HDFS clients.

See HDFS-10467 and the HDFS Router-based Federation documentation for more details.


  • YARN-5734 | Major | OrgQueue for easy CapacityScheduler queue configuration management
<!-- markdown -->

The OrgQueue extension to the capacity scheduler provides a programmatic way to change configurations by providing a REST API that users can call to modify queue configurations. This enables automation of queue configuration management by administrators in the queue's administer_queue ACL.


MapReduce support for the YARN shared cache allows MapReduce jobs to take advantage of additional resource caching. This saves network bandwidth between the job submission client as well as within the YARN cluster itself. This will reduce job submission time and overall job runtime.


  • YARN-6623 | Blocker | Add support to turn off launching privileged containers in the container-executor

A change in configuration for launching Docker containers under YARN. Docker container capabilities, mounts, networks and allowing privileged container have to specified in the container-executor.cfg. By default, all of the above are turned off. This change will break existing setups launching Docker containers under YARN. Please refer to the Docker containers under YARN documentation for more information.


  • HADOOP-14840 | Major | Tool to estimate resource requirements of an application pipeline based on prior executions

The first version of Resource Estimator service, a tool that captures the historical resource usage of an app and predicts its future resource requirement.


  • YARN-2877 | Major | Extend YARN to support distributed scheduling

With this JIRA we are introducing distributed scheduling in YARN. In particular, we make the following contributions:

  • Introduce the notion of container types. GUARANTEED containers follow the semantics of the existing YARN containers. OPPORTUNISTIC ones can be seen as lower priority containers, and can be preempted in order to make space for GUARANTEED containers to run.
  • Queuing of tasks at the NMs. This enables us to send more containers in an NM than its available resources. At the moment we are allowing queuing of OPPORTUNISTIC containers. Once resources become available at the NM, such containers can immediately start their execution.
  • Introduce the AMRMProxy. This is a service running at each node, intercepting the requests between the AM and the RM. It is instrumental for both distributed scheduling and YARN Federation (YARN-2915).
  • Enable distributed scheduling. To minimize their allocation latency, OPPORTUNISTIC containers are dispatched immediately to NMs in a distributed fashion by using the AMRMProxy of the node where the corresponding AM resides, without needing to go through the ResourceManager.

All the functionality introduced in this JIRA is disabled by default, so it will not affect the behavior of existing applications. We have introduced parameters in YarnConfiguration to enable NM queuing (yarn.nodemanager.container-queuing-enabled), distributed scheduling (yarn.distributed-scheduling.enabled) and the AMRMProxy service (yarn.nodemanager.amrmproxy.enable). AMs currently need to specify the type of container to be requested for each task. We are in the process of adding in the MapReduce AM the ability to randomly request OPPORTUNISTIC containers for a specified percentage of a job's tasks, so that users can experiment with the new features.


  • YARN-5220 | Major | Scheduling of OPPORTUNISTIC containers through YARN RM

This extends the centralized YARN RM in to enable the scheduling of OPPORTUNISTIC containers in a centralized fashion. This way, users can use OPPORTUNISTIC containers to improve the cluster's utilization, without needing to enable distributed scheduling.


  • YARN-5085 | Major | Add support for change of container ExecutionType

This allows the Application Master to ask the Scheduler to change the ExecutionType of a running/allocated container.


  • YARN-5049 | Major | Extend NMStateStore to save queued container information

This breaks rolling upgrades because it changes the major version of the NM state store schema. Therefore when a new NM comes up on an old state store it crashes.

The state store versions for this change have been updated in YARN-6798.