Merge pull request #27976 from Tyorden:master
PiperOrigin-RevId: 244755878
This commit is contained in:
commit
c25c87f426
20
README.md
20
README.md
@ -85,7 +85,7 @@ guidelines](CONTRIBUTING.md). This project adheres to TensorFlow's
|
||||
uphold this code.**
|
||||
|
||||
**We use [GitHub issues](https://github.com/tensorflow/tensorflow/issues) for
|
||||
tracking requests and bugs, so please see
|
||||
tracking requests and bugs, please see
|
||||
[TensorFlow Discuss](https://groups.google.com/a/tensorflow.org/forum/#!forum/discuss)
|
||||
for general questions and discussion, and please direct specific questions to
|
||||
[Stack Overflow](https://stackoverflow.com/questions/tagged/tensorflow).**
|
||||
@ -114,15 +114,15 @@ The TensorFlow project strives to abide by generally accepted best practices in
|
||||
|
||||
### Community Supported Builds
|
||||
|
||||
Build Type | Status | Artifacts
|
||||
-------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------
|
||||
**IBM s390x** | [](http://ibmz-ci.osuosl.org/job/TensorFlow_IBMZ_CI/) | TBA
|
||||
**Linux ppc64le CPU** Nightly | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Build/) | [Nightly](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Nightly_Artifact/)
|
||||
**Linux ppc64le CPU** Stable Release | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Release_Build/) | [Release](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Release_Build/)
|
||||
**Linux ppc64le GPU** Nightly | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Build/) | [Nightly](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Nightly_Artifact/)
|
||||
**Linux ppc64le GPU** Stable Release | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Release_Build/) | [Release](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Release_Build/)
|
||||
**Linux CPU with Intel® MKL-DNN** Nightly | [](https://tensorflow-ci.intel.com/job/tensorflow-mkl-linux-cpu/) | [Nightly](https://tensorflow-ci.intel.com/job/tensorflow-mkl-build-whl-nightly/)
|
||||
**Linux CPU with Intel® MKL-DNN** <br> **Supports Python 2.7, 3.4, 3.5 and 3.6** | [](https://tensorflow-ci.intel.com/job/tensorflow-mkl-build-release-whl/lastStableBuild) | [1.13.1 pypi](https://pypi.org/project/intel-tensorflow/)
|
||||
Build Type | Status | Artifacts
|
||||
--------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------
|
||||
**IBM s390x** | [](http://ibmz-ci.osuosl.org/job/TensorFlow_IBMZ_CI/) | TBA
|
||||
**Linux ppc64le CPU** Nightly | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Build/) | [Nightly](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Nightly_Artifact/)
|
||||
**Linux ppc64le CPU** Stable Release | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Release_Build/) | [Release](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_CPU_Release_Build/)
|
||||
**Linux ppc64le GPU** Nightly | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Build/) | [Nightly](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Nightly_Artifact/)
|
||||
**Linux ppc64le GPU** Stable Release | [](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Release_Build/) | [Release](https://powerci.osuosl.org/job/TensorFlow_PPC64LE_GPU_Release_Build/)
|
||||
**Linux CPU with Intel® MKL-DNN** Nightly | [](https://tensorflow-ci.intel.com/job/tensorflow-mkl-linux-cpu/) | [Nightly](https://tensorflow-ci.intel.com/job/tensorflow-mkl-build-whl-nightly/)
|
||||
**Linux CPU with Intel® MKL-DNN** <br> **Supports Python 2.7, 3.4, 3.5, and 3.6** | [](https://tensorflow-ci.intel.com/job/tensorflow-mkl-build-release-whl/lastStableBuild) | [1.13.1 pypi](https://pypi.org/project/intel-tensorflow/)
|
||||
|
||||
## For more information
|
||||
|
||||
|
271
RELEASE.md
271
RELEASE.md
@ -21,98 +21,185 @@
|
||||
|
||||
## Bug Fixes and Other Changes
|
||||
|
||||
* Documentation
|
||||
* Update the doc with the details about the rounding mode used in quantize_and_dequantize_v2.
|
||||
* Clarify that tensorflow::port::InitMain() _should_ be called before using the TensorFlow library. Programs failing to do this are not portable to all platforms.
|
||||
* Deprecations and Symbol renames.
|
||||
* Removing deprecations for the following endpoints: `tf.acos`, `tf.acosh`, `tf.add`, `tf.as_string`, `tf.asin`, `tf.asinh`, `tf.atan`, `tf.atan2`, `tf.atanh`, `tf.cos`, `tf.cosh`, `tf.equal`, `tf.exp`, `tf.floor`, `tf.greater`, `tf.greater_equal`, `tf.less`, `tf.less_equal`, `tf.log`, `tf.logp1`, `tf.logical_and`, `tf.logical_not`, `tf.logical_or`, `tf.maximum`, `tf.minimum`, `tf.not_equal`, `tf.sin`, `tf.sinh`, `tf.tan`
|
||||
* Deprecate `tf.data.Dataset.shard`.
|
||||
* Deprecate `saved_model.loader.load` which is replaced by `saved_model.load` and `saved_model.main_op`, which will be replaced by `saved_model.main_op` in V2.
|
||||
* Deprecate tf.QUANTIZED_DTYPES. The official new symbol is tf.dtypes.QUANTIZED_DTYPES.
|
||||
* Update sklearn imports for deprecated packages.
|
||||
* Deprecate `Variable.count_up_to` and `tf.count_up_to` in favor of `Dataset.range`.
|
||||
* Export `confusion_matrix` op as `tf.math.confusion_matrix` instead of `tf.train.confusion_matrix`.
|
||||
* Add `tf.dtypes.` endpoint for every constant in dtypes.py; moving endpoints in versions.py to corresponding endpoints in `tf.sysconfig.` and `tf.version.`; moving all constants under `tf.saved_model` submodules to `tf.saved_model` module. New endpoints are added in V1 and V2 but existing endpoint removals are only applied in V2.
|
||||
* Deprecates behavior where device assignment overrides collocation constraints inside a collocation context manager.
|
||||
* Keras & Python API
|
||||
* Add to Keras functionality analogous to `tf.register_tensor_conversion_function`.
|
||||
* Subclassed Keras models can now be saved through `tf.contrib.saved_model.save_keras_model`.
|
||||
* `LinearOperator.matmul` now returns a new `LinearOperator`.
|
||||
* New ops and improved op functionality
|
||||
* Add a Nearest Neighbor Resize op.
|
||||
* Add an `ignore_unknown` argument to `parse_values` which suppresses ValueError for unknown hyperparameter types. Such * Add `tf.linalg.matvec` convenience function.
|
||||
* `tf.einsum()`raises `ValueError` for unsupported equations like `"ii->"`.
|
||||
* Add DCT-I and IDCT-I in `tf.signal.dct` and `tf.signal.idct`.
|
||||
* Add LU decomposition op.
|
||||
* Add quantile loss to gradient boosted trees in estimator.
|
||||
* Add `round_mode` to `QuantizeAndDequantizeV2` op to select rounding algorithm.
|
||||
* Add `unicode_encode`, `unicode_decode`, `unicode_decode_with_offsets`, `unicode_split`, `unicode_split_with_offset`, and `unicode_transcode` ops. Amongst other things, this Op adds the ability to encode, decode, and transcode a variety of input text encoding formats into the main Unicode encodings (UTF-8, UTF-16-BE, UTF-32-BE)
|
||||
* Add "unit" attribute to the substr op, which allows obtaining the substring of a string containing unicode characters.
|
||||
* Broadcasting support for Ragged Tensors.
|
||||
* `SpaceToDepth` supports uint8 data type.
|
||||
* Support multi-label quantile regression in estimator.
|
||||
* We now use "div" as the default partition_strategy in `tf.nn.safe_embedding_lookup_sparse`, `tf.nn.sampled_softmax` and `tf.nn.nce_loss`.
|
||||
hyperparameter are ignored.
|
||||
* Performance
|
||||
* Improve performance of GPU cumsum/cumprod by up to 300x.
|
||||
* Added support for weight decay in most TPU embedding optimizers, including AdamW and MomentumW.
|
||||
* TensorFlow 2.0 Development
|
||||
* Add a command line tool to convert to TF2.0, tf_upgrade_v2
|
||||
* Merge `tf.spectral` into `tf.signal` for TensorFlow 2.0.
|
||||
* Change the default recurrent activation function for LSTM from 'hard_sigmoid' to 'sigmoid' in 2.0. Historically recurrent activation is 'hard_sigmoid' since it is fast than 'sigmoid'. With new unified backend between CPU and GPU mode, since the CuDNN kernel is using sigmoid, we change the default for CPU mode to sigmoid as well. With that, the default LSTM will be compatible with both CPU and GPU kernel. This will enable user with GPU to use CuDNN kernel by default and get a 10x performance boost in training. Note that this is checkpoint breaking change. If user want to use their 1.x pre-trained checkpoint, please construct the layer with LSTM(recurrent_activation='hard_sigmoid') to fallback to 1.x behavior.
|
||||
* TensorFlow Lite
|
||||
* Move from `tensorflow/contrib/lite` to `tensorflow/lite`.
|
||||
* Add experimental Java API for injecting TensorFlow Lite delegates
|
||||
* Add support for strings in TensorFlow Lite Java API.
|
||||
* `tf.contrib`:
|
||||
* Add Apache Ignite Filesystem plugin to support accessing Apache IGFS.
|
||||
* Dropout now takes `rate` argument, `keep_prob` is deprecated.
|
||||
* Estimator occurrences references `tf.contrib.estimator` were changed to `tf.estimator`:
|
||||
* `tf.contrib.estimator.BaselineEstimator` with `tf.estimator.BaselineEstimator`
|
||||
* `tf.contrib.estimator.DNNLinearCombinedEstimator` with `tf.estimator.DNNLinearCombinedEstimator`
|
||||
* `tf.contrib.estimator.DNNEstimator` with `tf.estimator.DNNEstimator`
|
||||
* `tf.contrib.estimator.LinearEstimator` with `tf.estimator.LinearEstimator`
|
||||
* `tf.contrib.estimator.InMemoryEvaluatorHook` and tf.estimator.experimental.InMemoryEvaluatorHook`.
|
||||
* `tf.contrib.estimator.make_stop_at_checkpoint_step_hook` with `tf.estimator.experimental.make_stop_at_checkpoint_step_hook`.
|
||||
* Expose `tf.distribute.Strategy as the new name for tf.contrib.distribute.DistributionStrategy.
|
||||
* Migrate linear optimizer from contrib to core.
|
||||
* Move `tf.contrib.signal` to `tf.signal` (preserving aliases in tf.contrib.signal).
|
||||
* Users of `tf.contrib.estimator.export_all_saved_models` and related should switch to `tf.estimator.Estimator.experimental_export_all_saved_models`.
|
||||
* tf.data:
|
||||
* Add `tf.data.experimental.StatsOptions()`, to configure options to collect statistics from `tf.data.Dataset` pipeline using `StatsAggregator`. Add nested option, `experimental_stats` (which takes a `tf.data.experimen tal.StatsOptions` object), to `tf.data.Options`. Deprecates `tf.data.experimental.set_stats_agregator`.
|
||||
* Performance optimizations:
|
||||
* Add `tf.data.experimental.OptimizationOptions()`, to configure options to enable `tf.data` performance optimizations. Add nested option, `experimental_optimization` (which takes a `tf.data.experimental.OptimizationOptions` object), to `tf.data.Options`. Remove performance optimization options from `tf.data.Options`, and add them under `tf.data.experimental.OptimizationOptions` instead.
|
||||
* Enable `map_and_batch_fusion` and `noop_elimination` optimizations by default. They can be disabled by configuring `tf.data.experimental.OptimizationOptions` to set `map_and_batch = False` or `noop_elimination = False` respectively. To disable all default optimizations, set `apply_default_optimizations = False`.
|
||||
* Support parallel map in `map_and_filter_fusion`.
|
||||
* Disable static optimizations for input pipelines that use non-resource `tf.Variable`s.
|
||||
* Add NUMA-aware MapAndBatch dataset.
|
||||
* Deprecate `tf.data.Dataset.make_one_shot_iterator()` in V1, removed it from V2, and added tf.compat.v1.data.make_one_shot_iterator()`.
|
||||
* Deprecate `tf.data.Dataset.make_initializable_iterator()` in V1, removed it from V2, and added `tf.compat.v1.data.make_initializable_iterator()`.
|
||||
* Enable nested dataset support in core `tf.data` transformations.
|
||||
* For `tf.data.Dataset` implementers: Added `tf.data.Dataset._element_structured property` to replace `Dataset.output_{types,shapes,classes}`.
|
||||
* Make `num_parallel_calls` of `tf.data.Dataset.interleave` and `tf.data.Dataset.map` work in Eager mode.
|
||||
* Toolchains
|
||||
* Fixed OpenSSL compatibility by avoiding `EVP_MD_CTX_destroy`.
|
||||
* Added bounds checking to printing deprecation warnings.
|
||||
* Upgraded CUDA dependency to 10.0
|
||||
* To build with Android NDK r14b, add "#include <linux/compiler.h>" to android-ndk-r14b/platforms/android-14/arch-*/usr/include/linux/futex.h
|
||||
* Removed `:android_tensorflow_lib_selective_registration*` targets, use `:android_tensorflow_lib_lite*` targets instead.
|
||||
* XLA
|
||||
* Move `RoundToEven` function to xla/client/lib/math.h.
|
||||
* A new environment variable `TF_XLA_DEBUG_OPTIONS_PASSTHROUGH` set to "1" or "true" allows the debug options passed within an XRTCompile op to be passed directly to the XLA compilation backend. If such variable is not set (service side), only a restricted set will be passed through.
|
||||
* Allow the XRTCompile op to return the ProgramShape resulted form the XLA compilation as a second return argument.
|
||||
* XLA HLO graphs can now be rendered as SVG/HTML.
|
||||
* Estimator
|
||||
* Replace all occurences of `tf.contrib.estimator.BaselineEstimator` with `tf.estimator.BaselineEstimator`
|
||||
* Replace all occurences of `tf.contrib.estimator.DNNLinearCombinedEstimator` with `tf.estimator.DNNLinearCombinedEstimator`
|
||||
* Replace all occurrences of `tf.contrib.estimator.DNNEstimator` with `tf.estimator.DNNEstimator`
|
||||
* Replace all occurrences of `tf.contrib.estimator.LinearEstimator` with `tf.estimator.LinearEstimator`
|
||||
* Users of `tf.contrib.estimator.export_all_saved_models` and related should switch to `tf.estimator.Estimator.experimental_export_all_saved_models`.
|
||||
* Update `regression_head` to the new Head API for Canned Estimator V2.
|
||||
* Switch `multi_class_head` to Head API for Canned Estimator V2.
|
||||
* Replace all occurences of `tf.contrib.estimator.InMemoryEvaluatorHook` and `tf.contrib.estimator.make_stop_at_checkpoint_step_hook` with `tf.estimator.experimental.InMemoryEvaluatorHook` and `tf.estimator.experimental.make_stop_at_checkpoint_step_hook`
|
||||
* Migrate linear optimizer from contrib to core.
|
||||
|
||||
* Documentation
|
||||
* Update the doc with the details about the rounding mode used in
|
||||
quantize_and_dequantize_v2.
|
||||
* Clarify that tensorflow::port::InitMain() _should_ be called before
|
||||
using the TensorFlow library. Programs failing to do this are not
|
||||
portable to all platforms.
|
||||
* Deprecations and Symbol renames.
|
||||
* Removing deprecations for the following endpoints: `tf.acos`,
|
||||
`tf.acosh`, `tf.add`, `tf.as_string`, `tf.asin`, `tf.asinh`, `tf.atan`,
|
||||
`tf.atan2`, `tf.atanh`, `tf.cos`, `tf.cosh`, `tf.equal`, `tf.exp`,
|
||||
`tf.floor`, `tf.greater`, `tf.greater_equal`, `tf.less`,
|
||||
`tf.less_equal`, `tf.log`, `tf.logp1`, `tf.logical_and`,
|
||||
`tf.logical_not`, `tf.logical_or`, `tf.maximum`, `tf.minimum`,
|
||||
`tf.not_equal`, `tf.sin`, `tf.sinh`, `tf.tan`
|
||||
* Deprecate `tf.data.Dataset.shard`.
|
||||
* Deprecate `saved_model.loader.load` which is replaced by
|
||||
`saved_model.load` and `saved_model.main_op`, which will be replaced by
|
||||
`saved_model.main_op` in V2.
|
||||
* Deprecate tf.QUANTIZED_DTYPES. The official new symbol is
|
||||
tf.dtypes.QUANTIZED_DTYPES.
|
||||
* Update sklearn imports for deprecated packages.
|
||||
* Deprecate `Variable.count_up_to` and `tf.count_up_to` in favor of
|
||||
`Dataset.range`.
|
||||
* Export `confusion_matrix` op as `tf.math.confusion_matrix` instead of
|
||||
`tf.train.confusion_matrix`.
|
||||
* Add `tf.dtypes.` endpoint for every constant in dtypes.py. Moving
|
||||
endpoints in versions.py to corresponding endpoints in `tf.sysconfig.`
|
||||
and `tf.version.`. Moving all constants under `tf.saved_model`
|
||||
submodules to `tf.saved_model` module. New endpoints are added in V1 and
|
||||
V2 but existing endpoint removals are only applied in V2.
|
||||
* Deprecates behavior where device assignment overrides collocation
|
||||
constraints inside a collocation context manager.
|
||||
* Keras & Python API
|
||||
* Add to Keras functionality analogous to
|
||||
`tf.register_tensor_conversion_function`.
|
||||
* Subclassed Keras models can now be saved through
|
||||
`tf.contrib.saved_model.save_keras_model`.
|
||||
* `LinearOperator.matmul` now returns a new `LinearOperator`.
|
||||
* New ops and improved op functionality
|
||||
* Add a Nearest Neighbor Resize op.
|
||||
* Add an `ignore_unknown` argument to `parse_values` which suppresses
|
||||
ValueError for unknown hyperparameter types. Such * Add
|
||||
`tf.linalg.matvec` convenience function.
|
||||
* `tf.einsum()`raises `ValueError` for unsupported equations like
|
||||
`"ii->"`.
|
||||
* Add DCT-I and IDCT-I in `tf.signal.dct` and `tf.signal.idct`.
|
||||
* Add LU decomposition op.
|
||||
* Add quantile loss to gradient boosted trees in estimator.
|
||||
* Add `round_mode` to `QuantizeAndDequantizeV2` op to select rounding
|
||||
algorithm.
|
||||
* Add `unicode_encode`, `unicode_decode`, `unicode_decode_with_offsets`,
|
||||
`unicode_split`, `unicode_split_with_offset`, and `unicode_transcode`
|
||||
ops. Amongst other things, this Op adds the ability to encode, decode,
|
||||
and transcode a variety of input text encoding formats into the main
|
||||
Unicode encodings (UTF-8, UTF-16-BE, UTF-32-BE)
|
||||
* Add "unit" attribute to the substr op, which allows obtaining the
|
||||
substring of a string containing unicode characters.
|
||||
* Broadcasting support for Ragged Tensors.
|
||||
* `SpaceToDepth` supports uint8 data type.
|
||||
* Support multi-label quantile regression in estimator.
|
||||
* We now use "div" as the default partition_strategy in
|
||||
`tf.nn.safe_embedding_lookup_sparse`, `tf.nn.sampled_softmax` and
|
||||
`tf.nn.nce_loss`. hyperparameter are ignored.
|
||||
* Performance
|
||||
* Improve performance of GPU cumsum/cumprod by up to 300x.
|
||||
* Added support for weight decay in most TPU embedding optimizers,
|
||||
including AdamW and MomentumW.
|
||||
* TensorFlow 2.0 Development
|
||||
* Add a command line tool to convert to TF2.0, tf_upgrade_v2
|
||||
* Merge `tf.spectral` into `tf.signal` for TensorFlow 2.0.
|
||||
* Change the default recurrent activation function for LSTM from
|
||||
'hard_sigmoid' to 'sigmoid' in 2.0. Historically recurrent activation is
|
||||
'hard_sigmoid' since it is fast than 'sigmoid'. With new unified backend
|
||||
between CPU and GPU mode, since the CuDNN kernel is using sigmoid, we
|
||||
change the default for CPU mode to sigmoid as well. With that, the
|
||||
default LSTM will be compatible with both CPU and GPU kernel. This will
|
||||
enable user with GPU to use CuDNN kernel by default and get a 10x
|
||||
performance boost in training. Note that this is checkpoint breaking
|
||||
change. If user want to use their 1.x pre-trained checkpoint, please
|
||||
construct the layer with LSTM(recurrent_activation='hard_sigmoid') to
|
||||
fallback to 1.x behavior.
|
||||
* TensorFlow Lite
|
||||
* Move from `tensorflow/contrib/lite` to `tensorflow/lite`.
|
||||
* Add experimental Java API for injecting TensorFlow Lite delegates
|
||||
* Add support for strings in TensorFlow Lite Java API.
|
||||
* `tf.contrib`:
|
||||
* Add Apache Ignite Filesystem plugin to support accessing Apache IGFS.
|
||||
* Dropout now takes `rate` argument, `keep_prob` is deprecated.
|
||||
* Estimator occurrences references `tf.contrib.estimator` were changed to
|
||||
`tf.estimator`:
|
||||
* `tf.contrib.estimator.BaselineEstimator` with
|
||||
`tf.estimator.BaselineEstimator`
|
||||
* `tf.contrib.estimator.DNNLinearCombinedEstimator` with
|
||||
`tf.estimator.DNNLinearCombinedEstimator`
|
||||
* `tf.contrib.estimator.DNNEstimator` with `tf.estimator.DNNEstimator`
|
||||
* `tf.contrib.estimator.LinearEstimator` with
|
||||
`tf.estimator.LinearEstimator`
|
||||
* `tf.contrib.estimator.InMemoryEvaluatorHook` and
|
||||
tf.estimator.experimental.InMemoryEvaluatorHook`.
|
||||
* `tf.contrib.estimator.make_stop_at_checkpoint_step_hook` with
|
||||
`tf.estimator.experimental.make_stop_at_checkpoint_step_hook`.
|
||||
* Expose `tf.distribute.Strategy as the new name for
|
||||
tf.contrib.distribute.DistributionStrategy.
|
||||
* Migrate linear optimizer from contrib to core.
|
||||
* Move `tf.contrib.signal` to `tf.signal` (preserving aliases in
|
||||
tf.contrib.signal).
|
||||
* Users of `tf.contrib.estimator.export_all_saved_models` and related
|
||||
should switch to
|
||||
`tf.estimator.Estimator.experimental_export_all_saved_models`.
|
||||
* tf.data:
|
||||
* Add `tf.data.experimental.StatsOptions()`, to configure options to
|
||||
collect statistics from `tf.data.Dataset` pipeline using
|
||||
`StatsAggregator`. Add nested option, `experimental_stats` (which takes
|
||||
a `tf.data.experimen tal.StatsOptions` object), to `tf.data.Options`.
|
||||
Deprecates `tf.data.experimental.set_stats_agregator`.
|
||||
* Performance optimizations:
|
||||
* Add `tf.data.experimental.OptimizationOptions()`, to configure options
|
||||
to enable `tf.data` performance optimizations. Add nested option,
|
||||
`experimental_optimization` (which takes a
|
||||
`tf.data.experimental.OptimizationOptions` object), to
|
||||
`tf.data.Options`. Remove performance optimization options from
|
||||
`tf.data.Options`, and add them under
|
||||
`tf.data.experimental.OptimizationOptions` instead.
|
||||
* Enable `map_and_batch_fusion` and `noop_elimination` optimizations by
|
||||
default. They can be disabled by configuring
|
||||
`tf.data.experimental.OptimizationOptions` to set `map_and_batch =
|
||||
False` or `noop_elimination = False` respectively. To disable all
|
||||
default optimizations, set `apply_default_optimizations = False`.
|
||||
* Support parallel map in `map_and_filter_fusion`.
|
||||
* Disable static optimizations for input pipelines that use non-resource
|
||||
`tf.Variable`s.
|
||||
* Add NUMA-aware MapAndBatch dataset.
|
||||
* Deprecate `tf.data.Dataset.make_one_shot_iterator()` in V1, removed it
|
||||
from V2, and added tf.compat.v1.data.make_one_shot_iterator()`.
|
||||
* Deprecate `tf.data.Dataset.make_initializable_iterator()` in V1, removed
|
||||
it from V2, and added `tf.compat.v1.data.make_initializable_iterator()`.
|
||||
* Enable nested dataset support in core `tf.data` transformations.
|
||||
* For `tf.data.Dataset` implementers: Added
|
||||
`tf.data.Dataset._element_structured property` to replace
|
||||
`Dataset.output_{types,shapes,classes}`.
|
||||
* Make `num_parallel_calls` of `tf.data.Dataset.interleave` and
|
||||
`tf.data.Dataset.map` work in Eager mode.
|
||||
* Toolchains
|
||||
* Fixed OpenSSL compatibility by avoiding `EVP_MD_CTX_destroy`.
|
||||
* Added bounds checking to printing deprecation warnings.
|
||||
* Upgraded CUDA dependency to 10.0
|
||||
* To build with Android NDK r14b, add "#include <linux/compiler.h>" to
|
||||
android-ndk-r14b/platforms/android-14/arch-*/usr/include/linux/futex.h
|
||||
* Removed `:android_tensorflow_lib_selective_registration*` targets, use
|
||||
`:android_tensorflow_lib_lite*` targets instead.
|
||||
* XLA
|
||||
* Move `RoundToEven` function to xla/client/lib/math.h.
|
||||
* A new environment variable `TF_XLA_DEBUG_OPTIONS_PASSTHROUGH` set to "1"
|
||||
or "true" allows the debug options passed within an XRTCompile op to be
|
||||
passed directly to the XLA compilation backend. If such variable is not
|
||||
set (service side), only a restricted set will be passed through.
|
||||
* Allow the XRTCompile op to return the ProgramShape resulted form the XLA
|
||||
compilation as a second return argument.
|
||||
* XLA HLO graphs can now be rendered as SVG/HTML.
|
||||
* Estimator
|
||||
* Replace all occurences of `tf.contrib.estimator.BaselineEstimator` with
|
||||
`tf.estimator.BaselineEstimator`
|
||||
* Replace all occurences of
|
||||
`tf.contrib.estimator.DNNLinearCombinedEstimator` with
|
||||
`tf.estimator.DNNLinearCombinedEstimator`
|
||||
* Replace all occurrences of `tf.contrib.estimator.DNNEstimator` with
|
||||
`tf.estimator.DNNEstimator`
|
||||
* Replace all occurrences of `tf.contrib.estimator.LinearEstimator` with
|
||||
`tf.estimator.LinearEstimator`
|
||||
* Users of `tf.contrib.estimator.export_all_saved_models` and related
|
||||
should switch to
|
||||
`tf.estimator.Estimator.experimental_export_all_saved_models`.
|
||||
* Update `regression_head` to the new Head API for Canned Estimator V2.
|
||||
* Switch `multi_class_head` to Head API for Canned Estimator V2.
|
||||
* Replace all occurences of `tf.contrib.estimator.InMemoryEvaluatorHook`
|
||||
and `tf.contrib.estimator.make_stop_at_checkpoint_step_hook` with
|
||||
`tf.estimator.experimental.InMemoryEvaluatorHook` and
|
||||
`tf.estimator.experimental.make_stop_at_checkpoint_step_hook`
|
||||
* Migrate linear optimizer from contrib to core.
|
||||
|
||||
## Thanks to our Contributors
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user