From 75ae7742abc027b001a5f3d7c020bb4504cc0f78 Mon Sep 17 00:00:00 2001 From: Ran Chen Date: Tue, 24 Mar 2020 13:28:21 -0700 Subject: [PATCH] Rename all_reduce_sum_gradients to experimental_aggregate_gradients For some strategies we don't do all reduce, so all_reduce_sum_gradients can be misleading. The parameter is also changed to experimental because of issues with CentralStorageStrategy. PiperOrigin-RevId: 302734837 Change-Id: Ic30e2f81ab61eef568ee68e5752015f950117d47 --- .../python/distribute/collective_util.py | 3 ++- .../custom_training_loop_optimizer_test.py | 18 +++++++------ tensorflow/python/keras/engine/training.py | 9 ++++--- .../python/keras/engine/training_test.py | 6 ++--- .../experimental/loss_scale_optimizer.py | 21 ++++++++------- .../experimental/loss_scale_optimizer_test.py | 8 +++--- .../python/keras/optimizer_v2/adadelta.py | 2 +- .../python/keras/optimizer_v2/adagrad.py | 2 +- tensorflow/python/keras/optimizer_v2/adam.py | 4 +-- .../python/keras/optimizer_v2/adamax.py | 2 +- .../keras/optimizer_v2/gradient_descent.py | 2 +- tensorflow/python/keras/optimizer_v2/nadam.py | 2 +- .../python/keras/optimizer_v2/optimizer_v2.py | 27 ++++++++++--------- .../keras/optimizer_v2/optimizer_v2_test.py | 8 +++--- .../python/keras/optimizer_v2/rmsprop.py | 2 +- tensorflow/python/keras/optimizers.py | 4 +-- ...n.experimental.-loss-scale-optimizer.pbtxt | 2 +- ...ensorflow.keras.optimizers.-adadelta.pbtxt | 2 +- ...tensorflow.keras.optimizers.-adagrad.pbtxt | 2 +- .../tensorflow.keras.optimizers.-adam.pbtxt | 2 +- .../tensorflow.keras.optimizers.-adamax.pbtxt | 2 +- .../tensorflow.keras.optimizers.-ftrl.pbtxt | 2 +- .../tensorflow.keras.optimizers.-nadam.pbtxt | 2 +- ...nsorflow.keras.optimizers.-optimizer.pbtxt | 2 +- ...nsorflow.keras.optimizers.-r-m-sprop.pbtxt | 2 +- .../tensorflow.keras.optimizers.-s-g-d.pbtxt | 2 +- ...n.experimental.-loss-scale-optimizer.pbtxt | 2 +- ...ensorflow.keras.optimizers.-adadelta.pbtxt | 2 +- ...tensorflow.keras.optimizers.-adagrad.pbtxt | 2 +- .../tensorflow.keras.optimizers.-adam.pbtxt | 2 +- .../tensorflow.keras.optimizers.-adamax.pbtxt | 2 +- .../tensorflow.keras.optimizers.-ftrl.pbtxt | 2 +- .../tensorflow.keras.optimizers.-nadam.pbtxt | 2 +- ...nsorflow.keras.optimizers.-optimizer.pbtxt | 2 +- ...nsorflow.keras.optimizers.-r-m-sprop.pbtxt | 2 +- .../tensorflow.keras.optimizers.-s-g-d.pbtxt | 2 +- .../v2/tensorflow.optimizers.-adadelta.pbtxt | 2 +- .../v2/tensorflow.optimizers.-adagrad.pbtxt | 2 +- .../v2/tensorflow.optimizers.-adam.pbtxt | 2 +- .../v2/tensorflow.optimizers.-adamax.pbtxt | 2 +- .../v2/tensorflow.optimizers.-ftrl.pbtxt | 2 +- .../v2/tensorflow.optimizers.-nadam.pbtxt | 2 +- .../v2/tensorflow.optimizers.-optimizer.pbtxt | 2 +- .../v2/tensorflow.optimizers.-r-m-sprop.pbtxt | 2 +- .../v2/tensorflow.optimizers.-s-g-d.pbtxt | 2 +- 45 files changed, 94 insertions(+), 84 deletions(-) diff --git a/tensorflow/python/distribute/collective_util.py b/tensorflow/python/distribute/collective_util.py index fb7008d1636..f79b953eb4a 100644 --- a/tensorflow/python/distribute/collective_util.py +++ b/tensorflow/python/distribute/collective_util.py @@ -41,7 +41,8 @@ class Hints(object): bytes_per_pack=50 * 1024 * 1024) grads = tf.distribute.get_replica_context().all_reduce( 'sum', grads, experimental_hints=hints) - optimizer.apply_gradients(zip(grads, vars), all_reduce_sum_gradients=False) + optimizer.apply_gradients(zip(grads, vars), + experimental_aggregate_gradients=False) ``` """ diff --git a/tensorflow/python/distribute/custom_training_loop_optimizer_test.py b/tensorflow/python/distribute/custom_training_loop_optimizer_test.py index 5f39efc9edc..942f83ed01d 100644 --- a/tensorflow/python/distribute/custom_training_loop_optimizer_test.py +++ b/tensorflow/python/distribute/custom_training_loop_optimizer_test.py @@ -40,14 +40,14 @@ class OptimizerTest(test.TestCase, parameterized.TestCase): ), combinations.concat( combinations.combine( - all_reduce_sum_gradients=True, + experimental_aggregate_gradients=True, expected=[[[-0.3, -0.3], [-0.3, -0.3]]]), combinations.combine( - all_reduce_sum_gradients=False, + experimental_aggregate_gradients=False, expected=[[[-0.1, -0.1], [-0.2, -0.2]]]), ))) - def test_custom_aggregation(self, distribution, all_reduce_sum_gradients, - expected): + def test_custom_aggregation(self, distribution, + experimental_aggregate_gradients, expected): with distribution.scope(): v = variables.Variable([0., 0.]) @@ -62,7 +62,8 @@ class OptimizerTest(test.TestCase, parameterized.TestCase): def step_fn(grads): optimizer.apply_gradients( - [(grads, v)], all_reduce_sum_gradients=all_reduce_sum_gradients) + [(grads, v)], + experimental_aggregate_gradients=experimental_aggregate_gradients) return v.read_value() return distribution.experimental_local_results( @@ -74,9 +75,9 @@ class OptimizerTest(test.TestCase, parameterized.TestCase): combinations.combine( distribution=strategy_combinations.one_device_strategy, mode=["eager"], - all_reduce_sum_gradients=[True, False])) + experimental_aggregate_gradients=[True, False])) def test_custom_aggregation_one_device(self, distribution, - all_reduce_sum_gradients): + experimental_aggregate_gradients): with distribution.scope(): v = variables.Variable([0., 0.]) @@ -88,7 +89,8 @@ class OptimizerTest(test.TestCase, parameterized.TestCase): def step_fn(grads): optimizer.apply_gradients( - [(grads, v)], all_reduce_sum_gradients=all_reduce_sum_gradients) + [(grads, v)], + experimental_aggregate_gradients=experimental_aggregate_gradients) return v.read_value() return distribution.experimental_local_results( diff --git a/tensorflow/python/keras/engine/training.py b/tensorflow/python/keras/engine/training.py index ece563ba9d9..a39e4768ec6 100644 --- a/tensorflow/python/keras/engine/training.py +++ b/tensorflow/python/keras/engine/training.py @@ -1794,7 +1794,7 @@ def _minimize(tape, optimizer, loss, trainable_variables): gradients = tape.gradient(loss, trainable_variables) - if optimizer._HAS_ALL_REDUCE_SUM_GRAD: # pylint: disable=protected-access + if optimizer._HAS_AGGREGATE_GRAD: # pylint: disable=protected-access # We aggregate gradients before unscaling them, in case a subclass of # LossScaleOptimizer all-reduces in fp16. All-reducing in fp16 can only be # done on scaled gradients, not unscaled gradients, for numeric stability. @@ -1804,9 +1804,10 @@ def _minimize(tape, optimizer, loss, trainable_variables): gradients = optimizer.get_unscaled_gradients(gradients) gradients = optimizer._clip_gradients(gradients) # pylint: disable=protected-access if trainable_variables: - if optimizer._HAS_ALL_REDUCE_SUM_GRAD: # pylint: disable=protected-access - optimizer.apply_gradients(zip(gradients, trainable_variables), - all_reduce_sum_gradients=False) + if optimizer._HAS_AGGREGATE_GRAD: # pylint: disable=protected-access + optimizer.apply_gradients( + zip(gradients, trainable_variables), + experimental_aggregate_gradients=False) else: optimizer.apply_gradients(zip(gradients, trainable_variables)) diff --git a/tensorflow/python/keras/engine/training_test.py b/tensorflow/python/keras/engine/training_test.py index 22125df6512..d8d44d18033 100644 --- a/tensorflow/python/keras/engine/training_test.py +++ b/tensorflow/python/keras/engine/training_test.py @@ -1342,7 +1342,7 @@ class TrainingTest(keras_parameterized.TestCase): class _Optimizer(optimizer_v2.gradient_descent.SGD): """Mock optimizer to check if _aggregate_gradient is called.""" - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self): self.aggregate_gradients_called = False @@ -1367,10 +1367,10 @@ class TrainingTest(keras_parameterized.TestCase): """Override apply_gradients. To test the case where the optimizer does not define the - all_reduce_sum_gradients parameter. + experimental_aggregate_gradients parameter. """ - _HAS_ALL_REDUCE_SUM_GRAD = False + _HAS_AGGREGATE_GRAD = False def apply_gradients(self, grads_and_vars, name=None): # pylint: disable=useless-super-delegation return super(_OptimizerOverrideApplyGradients, diff --git a/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer.py b/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer.py index 91e9ea6c587..fdaf971a85f 100644 --- a/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer.py +++ b/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer.py @@ -225,17 +225,19 @@ class LossScaleOptimizer(optimizer_v2.OptimizerV2): grads = self._optimizer.get_gradients(loss, params) return self.get_unscaled_gradients(grads) - def apply_gradients(self, grads_and_vars, name=None, - all_reduce_sum_gradients=True): + def apply_gradients(self, + grads_and_vars, + name=None, + experimental_aggregate_gradients=True): if distribution_strategy_context.in_cross_replica_context(): raise ValueError('apply_gradients() must be called in a replica context.') grads_and_vars = tuple(grads_and_vars) return distribution_strategy_context.get_replica_context().merge_call( self._apply_gradients_cross_replica, - args=(grads_and_vars, name, all_reduce_sum_gradients)) + args=(grads_and_vars, name, experimental_aggregate_gradients)) def _apply_gradients_cross_replica(self, distribution, grads_and_vars, name, - all_reduce_sum_gradients): + experimental_aggregate_gradients): grads = [g for g, _ in grads_and_vars] loss_scale_update_op, should_apply_grads = self._loss_scale.update(grads) @@ -247,8 +249,8 @@ class LossScaleOptimizer(optimizer_v2.OptimizerV2): # MirroredVariables. wrapped_vars = _UnwrapPreventer([v for _, v in grads_and_vars]) return distribution.extended.call_for_each_replica( - self._apply_gradients, args=(grads, wrapped_vars, name, - all_reduce_sum_gradients)) + self._apply_gradients, + args=(grads, wrapped_vars, name, experimental_aggregate_gradients)) # Note: We must call this cond() in a cross-replica context. # DistributionStrategy does not support having a cond in a replica context @@ -260,9 +262,10 @@ class LossScaleOptimizer(optimizer_v2.OptimizerV2): return control_flow_ops.group(maybe_apply_op, loss_scale_update_op) def _apply_gradients(self, grads, wrapped_vars, name, - all_reduce_sum_gradients): - return self._optimizer.apply_gradients(list(zip(grads, wrapped_vars.value)), - name, all_reduce_sum_gradients) + experimental_aggregate_gradients): + return self._optimizer.apply_gradients( + list(zip(grads, wrapped_vars.value)), name, + experimental_aggregate_gradients) def get_config(self): serialized_optimizer = optimizers.serialize(self._optimizer) diff --git a/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer_test.py b/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer_test.py index 8d13c6a37a3..8b4b7b9f7ec 100644 --- a/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer_test.py +++ b/tensorflow/python/keras/mixed_precision/experimental/loss_scale_optimizer_test.py @@ -375,13 +375,15 @@ class LossScaleOptimizerTest(test.TestCase, parameterized.TestCase): class MyOptimizer(gradient_descent.SGD): - def apply_gradients(self, grads_and_vars, name=None, - all_reduce_sum_gradients=True): + def apply_gradients(self, + grads_and_vars, + name=None, + experimental_aggregate_gradients=True): for grad, _ in grads_and_vars: outer_self.assertIsInstance(grad, ops.Tensor) return super(MyOptimizer, self).apply_gradients(grads_and_vars, name, - all_reduce_sum_gradients) + experimental_aggregate_gradients) with create_mirrored_strategy().scope() as strategy: var = variables.Variable([5.0]) diff --git a/tensorflow/python/keras/optimizer_v2/adadelta.py b/tensorflow/python/keras/optimizer_v2/adadelta.py index dfed74d8ab8..9d67ed25c66 100644 --- a/tensorflow/python/keras/optimizer_v2/adadelta.py +++ b/tensorflow/python/keras/optimizer_v2/adadelta.py @@ -71,7 +71,7 @@ class Adadelta(optimizer_v2.OptimizerV2): ([pdf](http://arxiv.org/pdf/1212.5701v1.pdf)) """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.001, diff --git a/tensorflow/python/keras/optimizer_v2/adagrad.py b/tensorflow/python/keras/optimizer_v2/adagrad.py index ca1cbd6d3fb..4e4ffd8e856 100644 --- a/tensorflow/python/keras/optimizer_v2/adagrad.py +++ b/tensorflow/python/keras/optimizer_v2/adagrad.py @@ -61,7 +61,7 @@ class Adagrad(optimizer_v2.OptimizerV2): (https://ppasupat.github.io/a9online/uploads/proximal_notes.pdf). """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.001, diff --git a/tensorflow/python/keras/optimizer_v2/adam.py b/tensorflow/python/keras/optimizer_v2/adam.py index 94eab7db6eb..67152e4b537 100644 --- a/tensorflow/python/keras/optimizer_v2/adam.py +++ b/tensorflow/python/keras/optimizer_v2/adam.py @@ -103,7 +103,7 @@ class Adam(optimizer_v2.OptimizerV2): 9.9 """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.001, @@ -354,7 +354,7 @@ class NonFusedAdam(optimizer_v2.OptimizerV2): 9.9 """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.001, diff --git a/tensorflow/python/keras/optimizer_v2/adamax.py b/tensorflow/python/keras/optimizer_v2/adamax.py index 9166f637c1e..9a7e1e28a89 100644 --- a/tensorflow/python/keras/optimizer_v2/adamax.py +++ b/tensorflow/python/keras/optimizer_v2/adamax.py @@ -73,7 +73,7 @@ class Adamax(optimizer_v2.OptimizerV2): ([pdf](http://arxiv.org/pdf/1412.6980.pdf)). """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.001, diff --git a/tensorflow/python/keras/optimizer_v2/gradient_descent.py b/tensorflow/python/keras/optimizer_v2/gradient_descent.py index 539443aef60..32547b95a52 100644 --- a/tensorflow/python/keras/optimizer_v2/gradient_descent.py +++ b/tensorflow/python/keras/optimizer_v2/gradient_descent.py @@ -74,7 +74,7 @@ class SGD(optimizer_v2.OptimizerV2): http://jmlr.org/proceedings/papers/v28/sutskever13.pdf). """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.01, diff --git a/tensorflow/python/keras/optimizer_v2/nadam.py b/tensorflow/python/keras/optimizer_v2/nadam.py index a572e225b62..f22fbaaae3c 100644 --- a/tensorflow/python/keras/optimizer_v2/nadam.py +++ b/tensorflow/python/keras/optimizer_v2/nadam.py @@ -61,7 +61,7 @@ class Nadam(optimizer_v2.OptimizerV2): See [Dozat, T., 2015](http://cs229.stanford.edu/proj2015/054_report.pdf). """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.001, diff --git a/tensorflow/python/keras/optimizer_v2/optimizer_v2.py b/tensorflow/python/keras/optimizer_v2/optimizer_v2.py index 3026816de8f..98b42cbad97 100644 --- a/tensorflow/python/keras/optimizer_v2/optimizer_v2.py +++ b/tensorflow/python/keras/optimizer_v2/optimizer_v2.py @@ -160,8 +160,8 @@ class OptimizerV2(trackable.Trackable): `tf.keras.losses.Reduction.SUM` for not. To aggregate gradients yourself, call `apply_gradients` with - `all_reduce_sum_gradients` set to False. This is useful if you need to process - aggregated gradients. + `experimental_aggregate_gradients` set to False. This is useful if you need to + process aggregated gradients. If you are not using these and you want to average gradients, you should use `tf.math.reduce_sum` to add up your per-example losses and then divide by the @@ -230,13 +230,13 @@ class OptimizerV2(trackable.Trackable): """ # Subclasses should set this to True unless they override `apply_gradients` - # with a version that does not have the `all_reduce_sum_gradients` argument. - # Older versions of Keras did not have this argument so custom optimizers may - # have overridden `apply_gradients` without the `all_reduce_sum_gradients` - # argument. Keras only passes `all_reduce_sum_gradients` if this attribute is - # True. + # with a version that does not have the `experimental_aggregate_gradients` + # argument. Older versions of Keras did not have this argument so custom + # optimizers may have overridden `apply_gradients` without the + # `experimental_aggregate_gradients` argument. Keras only passes + # `experimental_aggregate_gradients` if this attribute is True. # Note: This attribute will likely be removed in an upcoming release. - _HAS_ALL_REDUCE_SUM_GRAD = False + _HAS_AGGREGATE_GRAD = False def __init__(self, name, **kwargs): """Create a new Optimizer. @@ -433,7 +433,7 @@ class OptimizerV2(trackable.Trackable): def apply_gradients(self, grads_and_vars, name=None, - all_reduce_sum_gradients=True): + experimental_aggregate_gradients=True): """Apply gradients to variables. This is the second part of `minimize()`. It returns an `Operation` that @@ -441,7 +441,7 @@ class OptimizerV2(trackable.Trackable): The method sums gradients from all replicas in the presence of `tf.distribute.Strategy` by default. You can aggregate gradients yourself by - passing `all_reduce_sum_gradients=False`. + passing `experimental_aggregate_gradients=False`. Example: @@ -449,7 +449,8 @@ class OptimizerV2(trackable.Trackable): grads = tape.gradient(loss, vars) grads = tf.distribute.get_replica_context().all_reduce('sum', grads) # Processing aggregated gradients. - optimizer.apply_gradients(zip(grads, vars), all_reduce_sum_gradients=False) + optimizer.apply_gradients(zip(grads, vars), + experimental_aggregate_gradients=False) ``` @@ -457,7 +458,7 @@ class OptimizerV2(trackable.Trackable): grads_and_vars: List of (gradient, variable) pairs. name: Optional name for the returned operation. Default to the name passed to the `Optimizer` constructor. - all_reduce_sum_gradients: Whether to sum gradients from different + experimental_aggregate_gradients: Whether to sum gradients from different replicas in the presense of `tf.distribute.Strategy`. If False, it's user responsibility to aggregate the gradients. Default to True. @@ -491,7 +492,7 @@ class OptimizerV2(trackable.Trackable): "context.") apply_state = self._prepare(var_list) - if all_reduce_sum_gradients: + if experimental_aggregate_gradients: reduced_grads = self._aggregate_gradients(grads_and_vars) var_list = [v for _, v in grads_and_vars] grads_and_vars = list(zip(reduced_grads, var_list)) diff --git a/tensorflow/python/keras/optimizer_v2/optimizer_v2_test.py b/tensorflow/python/keras/optimizer_v2/optimizer_v2_test.py index b6e26f44ffa..9f99b43ad20 100644 --- a/tensorflow/python/keras/optimizer_v2/optimizer_v2_test.py +++ b/tensorflow/python/keras/optimizer_v2/optimizer_v2_test.py @@ -627,7 +627,7 @@ class OptimizerTest(test.TestCase, parameterized.TestCase): @combinations.generate(combinations.combine(mode=['graph', 'eager'])) def testAggregationTrue(self): - # Test that all_reduce_sum_gradients=True works without distributed + # Test that experimental_aggregate_gradients=True works without distributed # strategy. var = resource_variable_ops.ResourceVariable([1., 2.]) opt = gradient_descent.SGD(3.0) @@ -635,14 +635,14 @@ class OptimizerTest(test.TestCase, parameterized.TestCase): self.evaluate(variables.global_variables_initializer()) self.assertAllClose([1., 2.], self.evaluate(var)) opt_op = opt.apply_gradients([([0.1, 0.1], var)], - all_reduce_sum_gradients=True) + experimental_aggregate_gradients=True) self.evaluate(variables.global_variables_initializer()) self.evaluate(opt_op) self.assertAllClose([0.7, 1.7], self.evaluate(var)) @combinations.generate(combinations.combine(mode=['graph', 'eager'])) def testAggregationFalse(self): - # Test that all_reduce_sum_gradients=False works without distributed + # Test that experimental_aggregate_gradients=False works without distributed # strategy. var = resource_variable_ops.ResourceVariable([1., 2.]) opt = gradient_descent.SGD(3.0) @@ -650,7 +650,7 @@ class OptimizerTest(test.TestCase, parameterized.TestCase): self.evaluate(variables.global_variables_initializer()) self.assertAllClose([1., 2.], self.evaluate(var)) opt_op = opt.apply_gradients([([0.1, 0.1], var)], - all_reduce_sum_gradients=False) + experimental_aggregate_gradients=False) self.evaluate(variables.global_variables_initializer()) self.evaluate(opt_op) self.assertAllClose([0.7, 1.7], self.evaluate(var)) diff --git a/tensorflow/python/keras/optimizer_v2/rmsprop.py b/tensorflow/python/keras/optimizer_v2/rmsprop.py index bf08870bee5..d2e932a6564 100644 --- a/tensorflow/python/keras/optimizer_v2/rmsprop.py +++ b/tensorflow/python/keras/optimizer_v2/rmsprop.py @@ -91,7 +91,7 @@ class RMSprop(optimizer_v2.OptimizerV2): http://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf). """ - _HAS_ALL_REDUCE_SUM_GRAD = True + _HAS_AGGREGATE_GRAD = True def __init__(self, learning_rate=0.001, diff --git a/tensorflow/python/keras/optimizers.py b/tensorflow/python/keras/optimizers.py index 3fd9fc22023..393a4667c8f 100644 --- a/tensorflow/python/keras/optimizers.py +++ b/tensorflow/python/keras/optimizers.py @@ -72,8 +72,8 @@ class Optimizer(object): self.weights = [] # Set this to False, indicating `apply_gradients` does not take the - # `all_reduce_sum_gradients` argument. - _HAS_ALL_REDUCE_SUM_GRAD = False + # `experimental_aggregate_gradients` argument. + _HAS_AGGREGATE_GRAD = False def get_updates(self, loss, params): raise NotImplementedError diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt index de249627514..b486a58fc0b 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt @@ -38,7 +38,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adadelta.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adadelta.pbtxt index aaf0e8cc131..3a4e965007e 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adadelta.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adadelta.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adagrad.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adagrad.pbtxt index 2abbf63ada3..2cbc8b84800 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adagrad.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adagrad.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adam.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adam.pbtxt index c7c04aa59cf..19e805bd51c 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adam.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adam.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adamax.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adamax.pbtxt index a507e04483b..9e8914aa432 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adamax.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-adamax.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-ftrl.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-ftrl.pbtxt index 53b091a553b..e3656cef3e7 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-ftrl.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-ftrl.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-nadam.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-nadam.pbtxt index 80a8e3a90db..ed7b6cf0eb8 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-nadam.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-nadam.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-optimizer.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-optimizer.pbtxt index e95145b1fc5..d5f60307028 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-optimizer.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-optimizer.pbtxt @@ -25,7 +25,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-r-m-sprop.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-r-m-sprop.pbtxt index 7238e24bf29..7bb57f46a34 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-r-m-sprop.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-r-m-sprop.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-s-g-d.pbtxt b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-s-g-d.pbtxt index e4bbdc3ec55..c14f081f444 100644 --- a/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-s-g-d.pbtxt +++ b/tensorflow/tools/api/golden/v1/tensorflow.keras.optimizers.-s-g-d.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt index de249627514..b486a58fc0b 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.mixed_precision.experimental.-loss-scale-optimizer.pbtxt @@ -38,7 +38,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adadelta.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adadelta.pbtxt index aaf0e8cc131..3a4e965007e 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adadelta.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adadelta.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adagrad.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adagrad.pbtxt index 2abbf63ada3..2cbc8b84800 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adagrad.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adagrad.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adam.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adam.pbtxt index c7c04aa59cf..19e805bd51c 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adam.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adam.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adamax.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adamax.pbtxt index a507e04483b..9e8914aa432 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adamax.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-adamax.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-ftrl.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-ftrl.pbtxt index 53b091a553b..e3656cef3e7 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-ftrl.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-ftrl.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-nadam.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-nadam.pbtxt index 80a8e3a90db..ed7b6cf0eb8 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-nadam.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-nadam.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-optimizer.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-optimizer.pbtxt index e95145b1fc5..d5f60307028 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-optimizer.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-optimizer.pbtxt @@ -25,7 +25,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-r-m-sprop.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-r-m-sprop.pbtxt index 7238e24bf29..7bb57f46a34 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-r-m-sprop.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-r-m-sprop.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-s-g-d.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-s-g-d.pbtxt index e4bbdc3ec55..c14f081f444 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-s-g-d.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.keras.optimizers.-s-g-d.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adadelta.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adadelta.pbtxt index 8db3a63c868..a1a7c59dc48 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adadelta.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adadelta.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adagrad.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adagrad.pbtxt index 8505aa299e6..95579a35612 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adagrad.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adagrad.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adam.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adam.pbtxt index 2014e181484..78e5f6af446 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adam.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adam.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adamax.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adamax.pbtxt index a30f2a9afa4..29e50bd90dc 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adamax.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-adamax.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-ftrl.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-ftrl.pbtxt index f83fcd959de..4ba2ee05ef4 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-ftrl.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-ftrl.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-nadam.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-nadam.pbtxt index a21c2d9790c..419378e2174 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-nadam.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-nadam.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-optimizer.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-optimizer.pbtxt index 611044aa9c6..a390e41c6ea 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-optimizer.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-optimizer.pbtxt @@ -25,7 +25,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-r-m-sprop.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-r-m-sprop.pbtxt index a49290a1227..588dce59e1c 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-r-m-sprop.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-r-m-sprop.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config" diff --git a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-s-g-d.pbtxt b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-s-g-d.pbtxt index 6ac6872477d..5c4bcc4c17d 100644 --- a/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-s-g-d.pbtxt +++ b/tensorflow/tools/api/golden/v2/tensorflow.optimizers.-s-g-d.pbtxt @@ -26,7 +26,7 @@ tf_class { } member_method { name: "apply_gradients" - argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'all_reduce_sum_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " + argspec: "args=[\'self\', \'grads_and_vars\', \'name\', \'experimental_aggregate_gradients\'], varargs=None, keywords=None, defaults=[\'None\', \'True\'], " } member_method { name: "from_config"