Re-enable tests with old optimizers + dist strat + keras.
PiperOrigin-RevId: 260619319
This commit is contained in:
parent
e327064034
commit
5dd7dbb233
@ -274,6 +274,10 @@ def strategy_and_optimizer_combinations():
|
||||
strategy_minus_tpu_combinations(),
|
||||
combinations.combine(
|
||||
optimizer=[
|
||||
strategy_combinations.adagrad_optimizer_v1_fn,
|
||||
strategy_combinations.adam_optimizer_v1_fn,
|
||||
strategy_combinations.gradient_descent_optimizer_v1_fn,
|
||||
strategy_combinations.rmsprop_optimizer_v1_fn,
|
||||
strategy_combinations.adagrad_optimizer_keras_v2_fn,
|
||||
strategy_combinations.adam_optimizer_keras_v2_fn,
|
||||
strategy_combinations.gradient_descent_optimizer_keras_v2_fn,
|
||||
|
@ -51,8 +51,6 @@ class DistributionStrategyEmbeddingModelCorrectnessTest(
|
||||
model.set_weights(initial_weights)
|
||||
|
||||
model.compile(
|
||||
# TODO(b/130808953): Switch back the V1 optimizer once global_step is
|
||||
# mirrored.
|
||||
optimizer=gradient_descent_keras.SGD(learning_rate=0.1),
|
||||
loss='sparse_categorical_crossentropy',
|
||||
metrics=['sparse_categorical_accuracy'],
|
||||
|
Loading…
x
Reference in New Issue
Block a user