Fork tracking/util_with_v1_optimizers_test to keras/tests.
PiperOrigin-RevId: 316496459 Change-Id: I9563ba20c0d3cd72bc0651e49ab9a48e0b355d34
This commit is contained in:
parent
e3b8a8e720
commit
a2dc78458b
@ -435,6 +435,37 @@ tf_py_test(
|
|||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
tf_py_test(
|
||||||
|
name = "tracking_util_with_v1_optimizers_test",
|
||||||
|
srcs = ["tracking_util_with_v1_optimizers_test.py"],
|
||||||
|
tags = [
|
||||||
|
"notsan", # b/74395663
|
||||||
|
],
|
||||||
|
deps = [
|
||||||
|
"//tensorflow/python:checkpoint_management",
|
||||||
|
"//tensorflow/python:constant_op",
|
||||||
|
"//tensorflow/python:dtypes",
|
||||||
|
"//tensorflow/python:extra_py_tests_deps",
|
||||||
|
"//tensorflow/python:framework_ops",
|
||||||
|
"//tensorflow/python:framework_test_lib",
|
||||||
|
"//tensorflow/python:resource_variable_ops",
|
||||||
|
"//tensorflow/python:saver",
|
||||||
|
"//tensorflow/python:state_ops",
|
||||||
|
"//tensorflow/python:training_lib",
|
||||||
|
"//tensorflow/python:training_util",
|
||||||
|
"//tensorflow/python/distribute:mirrored_strategy",
|
||||||
|
"//tensorflow/python/eager:backprop",
|
||||||
|
"//tensorflow/python/eager:context",
|
||||||
|
"//tensorflow/python/eager:def_function",
|
||||||
|
"//tensorflow/python/eager:test",
|
||||||
|
"//tensorflow/python/keras/engine",
|
||||||
|
"//tensorflow/python/keras/layers:core",
|
||||||
|
"//tensorflow/python/training/tracking",
|
||||||
|
"//tensorflow/python/training/tracking:graph_view",
|
||||||
|
"//tensorflow/python/training/tracking:util",
|
||||||
|
],
|
||||||
|
)
|
||||||
|
|
||||||
py_library(
|
py_library(
|
||||||
name = "get_config_samples",
|
name = "get_config_samples",
|
||||||
srcs = ["get_config_samples.py"],
|
srcs = ["get_config_samples.py"],
|
||||||
|
@ -0,0 +1,703 @@
|
|||||||
|
# Copyright 2017 The TensorFlow Authors. All Rights Reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
# ==============================================================================
|
||||||
|
"""Tests for object-based saving which use tf.train.* optimizers."""
|
||||||
|
from __future__ import absolute_import
|
||||||
|
from __future__ import division
|
||||||
|
from __future__ import print_function
|
||||||
|
|
||||||
|
import functools
|
||||||
|
import os
|
||||||
|
|
||||||
|
import six
|
||||||
|
|
||||||
|
from tensorflow.python.distribute import mirrored_strategy
|
||||||
|
from tensorflow.python.eager import backprop
|
||||||
|
from tensorflow.python.eager import context
|
||||||
|
from tensorflow.python.eager import def_function
|
||||||
|
from tensorflow.python.eager import test
|
||||||
|
from tensorflow.python.framework import constant_op
|
||||||
|
from tensorflow.python.framework import dtypes
|
||||||
|
from tensorflow.python.framework import ops
|
||||||
|
from tensorflow.python.framework import test_util
|
||||||
|
from tensorflow.python.keras.engine import training
|
||||||
|
from tensorflow.python.keras.layers import core
|
||||||
|
from tensorflow.python.ops import resource_variable_ops
|
||||||
|
from tensorflow.python.ops import state_ops
|
||||||
|
from tensorflow.python.training import adam
|
||||||
|
from tensorflow.python.training import checkpoint_management
|
||||||
|
from tensorflow.python.training import saver as saver_lib
|
||||||
|
from tensorflow.python.training import training_util
|
||||||
|
from tensorflow.python.training.tracking import graph_view
|
||||||
|
from tensorflow.python.training.tracking import tracking
|
||||||
|
from tensorflow.python.training.tracking import util as trackable_utils
|
||||||
|
|
||||||
|
|
||||||
|
class NonLayerTrackable(tracking.AutoTrackable):
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super(NonLayerTrackable, self).__init__()
|
||||||
|
self.a_variable = trackable_utils.add_variable(
|
||||||
|
self, name="a_variable", shape=[])
|
||||||
|
|
||||||
|
|
||||||
|
# pylint: disable=not-callable
|
||||||
|
class MyModel(training.Model):
|
||||||
|
"""A concrete Model for testing."""
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super(MyModel, self).__init__()
|
||||||
|
self._named_dense = core.Dense(1, use_bias=True)
|
||||||
|
self._second = core.Dense(1, use_bias=False)
|
||||||
|
# We can still track Trackables which aren't Layers.
|
||||||
|
self._non_layer = NonLayerTrackable()
|
||||||
|
|
||||||
|
def call(self, values):
|
||||||
|
ret = self._second(self._named_dense(values))
|
||||||
|
return ret
|
||||||
|
|
||||||
|
|
||||||
|
class CheckpointingTests(test.TestCase):
|
||||||
|
|
||||||
|
@test_util.run_in_graph_and_eager_modes(assert_no_eager_garbage=True)
|
||||||
|
def testNamingWithOptimizer(self):
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
model = MyModel()
|
||||||
|
# A nuisance Model using the same optimizer. Its slot variables should not
|
||||||
|
# go in the checkpoint, since it is never depended on.
|
||||||
|
other_model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
optimizer_step = training_util.get_or_create_global_step()
|
||||||
|
root_trackable = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
||||||
|
if context.executing_eagerly():
|
||||||
|
optimizer.minimize(
|
||||||
|
lambda: model(input_value),
|
||||||
|
global_step=optimizer_step)
|
||||||
|
optimizer.minimize(
|
||||||
|
lambda: other_model(input_value),
|
||||||
|
global_step=optimizer_step)
|
||||||
|
else:
|
||||||
|
train_op = optimizer.minimize(
|
||||||
|
model(input_value), global_step=optimizer_step)
|
||||||
|
optimizer.minimize(
|
||||||
|
other_model(input_value),
|
||||||
|
global_step=optimizer_step)
|
||||||
|
self.evaluate(trackable_utils.gather_initializers(
|
||||||
|
root_trackable))
|
||||||
|
self.evaluate(train_op)
|
||||||
|
named_variables, serialized_graph, _ = graph_view.ObjectGraphView(
|
||||||
|
root_trackable).serialize_object_graph()
|
||||||
|
expected_checkpoint_names = (
|
||||||
|
# Created in the root node, so no prefix.
|
||||||
|
"optimizer_step",
|
||||||
|
"model/_second/kernel",
|
||||||
|
"model/_named_dense/kernel",
|
||||||
|
"model/_named_dense/bias",
|
||||||
|
# non-Layer dependency of the model
|
||||||
|
"model/_non_layer/a_variable",
|
||||||
|
# The optimizer creates two non-slot variables
|
||||||
|
"optimizer/beta1_power",
|
||||||
|
"optimizer/beta2_power",
|
||||||
|
# Slot variables
|
||||||
|
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
||||||
|
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
||||||
|
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
||||||
|
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
||||||
|
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/m",
|
||||||
|
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/v",
|
||||||
|
)
|
||||||
|
suffix = "/.ATTRIBUTES/VARIABLE_VALUE"
|
||||||
|
expected_checkpoint_names = [
|
||||||
|
name + suffix for name in expected_checkpoint_names]
|
||||||
|
named_variables = {v.name: v for v in named_variables}
|
||||||
|
six.assertCountEqual(self, expected_checkpoint_names,
|
||||||
|
named_variables.keys())
|
||||||
|
# Check that we've mapped to the right variable objects (not exhaustive)
|
||||||
|
self.assertEqual(
|
||||||
|
"global_step",
|
||||||
|
named_variables["optimizer_step" + suffix].full_name)
|
||||||
|
self.assertEqual(
|
||||||
|
"my_model/dense_1/kernel",
|
||||||
|
named_variables["model/_second/kernel" + suffix].full_name)
|
||||||
|
self.assertEqual(
|
||||||
|
"my_model/dense/kernel",
|
||||||
|
named_variables["model/_named_dense/kernel" + suffix].full_name)
|
||||||
|
self.assertEqual(
|
||||||
|
"beta1_power",
|
||||||
|
named_variables["optimizer/beta1_power" + suffix].full_name)
|
||||||
|
self.assertEqual(
|
||||||
|
"beta2_power",
|
||||||
|
named_variables["optimizer/beta2_power" + suffix].full_name)
|
||||||
|
# Spot check the generated protocol buffers.
|
||||||
|
self.assertEqual("optimizer",
|
||||||
|
serialized_graph.nodes[0].children[1].local_name)
|
||||||
|
optimizer_node = serialized_graph.nodes[serialized_graph.nodes[0].children[
|
||||||
|
1].node_id]
|
||||||
|
self.assertEqual("beta1_power",
|
||||||
|
optimizer_node.children[0].local_name)
|
||||||
|
self.assertEqual("beta1_power",
|
||||||
|
serialized_graph.nodes[optimizer_node.children[0].node_id]
|
||||||
|
.attributes[0].full_name)
|
||||||
|
self.assertEqual(
|
||||||
|
"my_model/dense/kernel",
|
||||||
|
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
||||||
|
.original_variable_node_id]
|
||||||
|
.attributes[0].full_name)
|
||||||
|
# We strip off the :0 suffix, as variable.name-based saving does.
|
||||||
|
self.assertEqual(
|
||||||
|
"my_model/dense/kernel/Adam",
|
||||||
|
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
||||||
|
.slot_variable_node_id]
|
||||||
|
.attributes[0].full_name)
|
||||||
|
self.assertEqual(
|
||||||
|
"my_model/dense/kernel/Adam:0",
|
||||||
|
optimizer.get_slot(
|
||||||
|
var=model._named_dense.kernel,
|
||||||
|
name="m").name)
|
||||||
|
self.assertEqual(
|
||||||
|
"model/_named_dense/kernel" + suffix,
|
||||||
|
serialized_graph.nodes[
|
||||||
|
optimizer_node.slot_variables[0]
|
||||||
|
.original_variable_node_id].attributes[0].checkpoint_key)
|
||||||
|
self.assertEqual("m", optimizer_node.slot_variables[0].slot_name)
|
||||||
|
self.assertEqual(
|
||||||
|
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m" + suffix,
|
||||||
|
serialized_graph.nodes[
|
||||||
|
optimizer_node.slot_variables[0]
|
||||||
|
.slot_variable_node_id].attributes[0].checkpoint_key)
|
||||||
|
|
||||||
|
@test_util.run_in_graph_and_eager_modes
|
||||||
|
def testSaveRestore(self):
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root_trackable = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model)
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
if context.executing_eagerly():
|
||||||
|
optimizer.minimize(
|
||||||
|
lambda: model(input_value))
|
||||||
|
else:
|
||||||
|
train_op = optimizer.minimize(model(input_value))
|
||||||
|
# TODO(allenl): Make initialization more pleasant when graph building.
|
||||||
|
root_trackable.save_counter # pylint: disable=pointless-statement
|
||||||
|
self.evaluate(trackable_utils.gather_initializers(
|
||||||
|
root_trackable))
|
||||||
|
self.evaluate(train_op)
|
||||||
|
prefix = os.path.join(self.get_temp_dir(), "ckpt")
|
||||||
|
self.evaluate(state_ops.assign(model._named_dense.variables[1], [42.]))
|
||||||
|
m_bias_slot = optimizer.get_slot(model._named_dense.variables[1], "m")
|
||||||
|
self.evaluate(state_ops.assign(m_bias_slot, [1.5]))
|
||||||
|
save_path = root_trackable.save(file_prefix=prefix)
|
||||||
|
self.evaluate(state_ops.assign(model._named_dense.variables[1], [43.]))
|
||||||
|
self.evaluate(state_ops.assign(root_trackable.save_counter, 3))
|
||||||
|
optimizer_variables = self.evaluate(optimizer.variables())
|
||||||
|
self.evaluate(state_ops.assign(m_bias_slot, [-2.]))
|
||||||
|
# Immediate restoration
|
||||||
|
status = root_trackable.restore(save_path=save_path).assert_consumed()
|
||||||
|
status.run_restore_ops()
|
||||||
|
self.assertAllEqual([42.], self.evaluate(model._named_dense.variables[1]))
|
||||||
|
self.assertAllEqual(1, self.evaluate(root_trackable.save_counter))
|
||||||
|
self.assertAllEqual([1.5], self.evaluate(m_bias_slot))
|
||||||
|
if not context.executing_eagerly():
|
||||||
|
return # Restore-on-create is only supported when executing eagerly
|
||||||
|
on_create_model = MyModel()
|
||||||
|
on_create_optimizer = adam.AdamOptimizer(
|
||||||
|
0.001,
|
||||||
|
# Preserve beta1_power and beta2_power when applying gradients so we can
|
||||||
|
# test that they've been restored correctly.
|
||||||
|
beta1=1.0,
|
||||||
|
beta2=1.0)
|
||||||
|
on_create_root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=on_create_optimizer, model=on_create_model)
|
||||||
|
# Deferred restoration
|
||||||
|
status = on_create_root.restore(save_path=save_path)
|
||||||
|
status.assert_nontrivial_match()
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
status.assert_consumed()
|
||||||
|
on_create_model(constant_op.constant([[3.]])) # create variables
|
||||||
|
self.assertAllEqual(1, self.evaluate(on_create_root.save_counter))
|
||||||
|
self.assertAllEqual([42.],
|
||||||
|
self.evaluate(
|
||||||
|
on_create_model._named_dense.variables[1]))
|
||||||
|
on_create_m_bias_slot = on_create_optimizer.get_slot(
|
||||||
|
on_create_model._named_dense.variables[1], "m")
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
status.assert_consumed()
|
||||||
|
# Optimizer slot variables are created when the original variable is
|
||||||
|
# restored.
|
||||||
|
self.assertAllEqual([1.5], self.evaluate(on_create_m_bias_slot))
|
||||||
|
self.assertAllEqual(optimizer_variables[2:],
|
||||||
|
self.evaluate(on_create_optimizer.variables()))
|
||||||
|
dummy_var = resource_variable_ops.ResourceVariable([1.])
|
||||||
|
on_create_optimizer.minimize(loss=dummy_var.read_value)
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
status.assert_consumed()
|
||||||
|
beta1_power, beta2_power = on_create_optimizer._get_beta_accumulators()
|
||||||
|
self.assertAllEqual(optimizer_variables[0], self.evaluate(beta1_power))
|
||||||
|
self.assertAllEqual(optimizer_variables[1], self.evaluate(beta2_power))
|
||||||
|
|
||||||
|
# TODO(allenl): Debug garbage created by this test in python3.
|
||||||
|
def testDeferredRestorationUsageEager(self):
|
||||||
|
"""An idiomatic eager execution example."""
|
||||||
|
num_training_steps = 10
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
for training_continuation in range(3):
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model,
|
||||||
|
optimizer_step=training_util.get_or_create_global_step())
|
||||||
|
root.restore(checkpoint_management.latest_checkpoint(
|
||||||
|
checkpoint_directory))
|
||||||
|
for _ in range(num_training_steps):
|
||||||
|
# TODO(allenl): Use a Dataset and serialize/checkpoint it.
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
optimizer.minimize(
|
||||||
|
lambda: model(input_value), # pylint: disable=cell-var-from-loop
|
||||||
|
global_step=root.optimizer_step)
|
||||||
|
root.save(file_prefix=checkpoint_prefix)
|
||||||
|
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||||
|
root.optimizer_step.numpy())
|
||||||
|
|
||||||
|
def testEagerDistributionStrategy(self):
|
||||||
|
num_training_steps = 10
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
|
||||||
|
def _train_fn(optimizer, model):
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
optimizer.minimize(
|
||||||
|
functools.partial(model, input_value),
|
||||||
|
global_step=root.optimizer_step)
|
||||||
|
|
||||||
|
strategy = mirrored_strategy.MirroredStrategy()
|
||||||
|
with strategy.scope():
|
||||||
|
for training_continuation in range(3):
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer,
|
||||||
|
model=model,
|
||||||
|
optimizer_step=training_util.get_or_create_global_step())
|
||||||
|
root.restore(
|
||||||
|
checkpoint_management.latest_checkpoint(checkpoint_directory))
|
||||||
|
|
||||||
|
for _ in range(num_training_steps):
|
||||||
|
strategy.extended.call_for_each_replica(
|
||||||
|
functools.partial(_train_fn, optimizer, model))
|
||||||
|
root.save(file_prefix=checkpoint_prefix)
|
||||||
|
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||||
|
root.optimizer_step.numpy())
|
||||||
|
|
||||||
|
def testGraphDistributionStrategy(self):
|
||||||
|
self.skipTest("b/121381184")
|
||||||
|
num_training_steps = 10
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
|
||||||
|
def _train_fn(optimizer, model):
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
return optimizer.minimize(
|
||||||
|
functools.partial(model, input_value),
|
||||||
|
global_step=root.optimizer_step)
|
||||||
|
|
||||||
|
for training_continuation in range(3):
|
||||||
|
with ops.Graph().as_default():
|
||||||
|
strategy = mirrored_strategy.MirroredStrategy()
|
||||||
|
with strategy.scope():
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model,
|
||||||
|
optimizer_step=training_util.get_or_create_global_step())
|
||||||
|
status = root.restore(checkpoint_management.latest_checkpoint(
|
||||||
|
checkpoint_directory))
|
||||||
|
train_op = strategy.extended.call_for_each_replica(
|
||||||
|
functools.partial(_train_fn, optimizer, model))
|
||||||
|
with self.session() as session:
|
||||||
|
if training_continuation > 0:
|
||||||
|
status.assert_consumed()
|
||||||
|
status.initialize_or_restore()
|
||||||
|
for _ in range(num_training_steps):
|
||||||
|
session.run(train_op)
|
||||||
|
root.save(file_prefix=checkpoint_prefix)
|
||||||
|
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||||
|
root.optimizer_step.numpy())
|
||||||
|
|
||||||
|
def testUsageGraph(self):
|
||||||
|
"""Expected usage when graph building."""
|
||||||
|
with context.graph_mode():
|
||||||
|
num_training_steps = 10
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
for training_continuation in range(3):
|
||||||
|
with ops.Graph().as_default():
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root = trackable_utils.CheckpointV1(
|
||||||
|
optimizer=optimizer, model=model,
|
||||||
|
global_step=training_util.get_or_create_global_step())
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
train_op = optimizer.minimize(
|
||||||
|
model(input_value),
|
||||||
|
global_step=root.global_step)
|
||||||
|
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||||
|
checkpoint_directory)
|
||||||
|
with self.session(graph=ops.get_default_graph()) as session:
|
||||||
|
status = root.restore(save_path=checkpoint_path)
|
||||||
|
status.initialize_or_restore(session=session)
|
||||||
|
if checkpoint_path is None:
|
||||||
|
self.assertEqual(0, training_continuation)
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
status.assert_consumed()
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
else:
|
||||||
|
status.assert_consumed()
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
for _ in range(num_training_steps):
|
||||||
|
session.run(train_op)
|
||||||
|
root.save(file_prefix=checkpoint_prefix, session=session)
|
||||||
|
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||||
|
session.run(root.global_step))
|
||||||
|
self.assertEqual(training_continuation + 1,
|
||||||
|
session.run(root.save_counter))
|
||||||
|
|
||||||
|
@test_util.run_in_graph_and_eager_modes
|
||||||
|
def testAgnosticUsage(self):
|
||||||
|
"""Graph/eager agnostic usage."""
|
||||||
|
# Does create garbage when executing eagerly due to ops.Graph() creation.
|
||||||
|
num_training_steps = 10
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
for training_continuation in range(3):
|
||||||
|
with test_util.device(use_gpu=True):
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model,
|
||||||
|
global_step=training_util.get_or_create_global_step())
|
||||||
|
manager = checkpoint_management.CheckpointManager(
|
||||||
|
root, checkpoint_directory, max_to_keep=1)
|
||||||
|
status = root.restore(save_path=manager.latest_checkpoint)
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
train_fn = functools.partial(
|
||||||
|
optimizer.minimize,
|
||||||
|
functools.partial(model, input_value),
|
||||||
|
global_step=root.global_step)
|
||||||
|
if not context.executing_eagerly():
|
||||||
|
train_fn = functools.partial(self.evaluate, train_fn())
|
||||||
|
status.initialize_or_restore()
|
||||||
|
for _ in range(num_training_steps):
|
||||||
|
train_fn()
|
||||||
|
manager.save()
|
||||||
|
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||||
|
self.evaluate(root.global_step))
|
||||||
|
self.assertEqual(training_continuation + 1,
|
||||||
|
self.evaluate(root.save_counter))
|
||||||
|
|
||||||
|
# pylint: disable=cell-var-from-loop
|
||||||
|
@test_util.run_in_graph_and_eager_modes
|
||||||
|
def testWithDefun(self):
|
||||||
|
num_training_steps = 2
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
for training_continuation in range(3):
|
||||||
|
with test_util.device(use_gpu=True):
|
||||||
|
model = MyModel()
|
||||||
|
# Don't actually train so we can test variable values
|
||||||
|
optimizer = adam.AdamOptimizer(0.)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model,
|
||||||
|
global_step=training_util.get_or_create_global_step())
|
||||||
|
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||||
|
checkpoint_directory)
|
||||||
|
status = root.restore(save_path=checkpoint_path)
|
||||||
|
def train_fn():
|
||||||
|
@def_function.function
|
||||||
|
def _call_model(x):
|
||||||
|
return model(x)
|
||||||
|
with backprop.GradientTape() as tape:
|
||||||
|
loss = _call_model(constant_op.constant([[3.]]))
|
||||||
|
gradients = tape.gradient(loss, model.variables)
|
||||||
|
return optimizer.apply_gradients(zip(gradients, model.variables),
|
||||||
|
global_step=root.global_step)
|
||||||
|
if not context.executing_eagerly():
|
||||||
|
train_fn = functools.partial(
|
||||||
|
self.evaluate, train_fn())
|
||||||
|
status.initialize_or_restore()
|
||||||
|
for _ in range(num_training_steps):
|
||||||
|
train_fn()
|
||||||
|
if training_continuation > 0:
|
||||||
|
status.assert_consumed()
|
||||||
|
self.assertAllClose([[42.]], self.evaluate(model.variables[0]))
|
||||||
|
else:
|
||||||
|
self.evaluate(model.variables[0].assign([[42.]]))
|
||||||
|
root.save(file_prefix=checkpoint_prefix)
|
||||||
|
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||||
|
self.evaluate(root.global_step))
|
||||||
|
self.assertEqual(training_continuation + 1,
|
||||||
|
self.evaluate(root.save_counter))
|
||||||
|
# pylint: enable=cell-var-from-loop
|
||||||
|
|
||||||
|
def _get_checkpoint_name(self, name):
|
||||||
|
root = tracking.AutoTrackable()
|
||||||
|
trackable_utils.add_variable(
|
||||||
|
root, name=name, shape=[1, 2], dtype=dtypes.float64)
|
||||||
|
(named_variable,), _, _ = trackable_utils._serialize_object_graph(
|
||||||
|
root, saveables_cache=None)
|
||||||
|
with ops.name_scope("root/" + named_variable.name):
|
||||||
|
pass # Make sure we can use this as an op name if we prefix it.
|
||||||
|
return named_variable.name
|
||||||
|
|
||||||
|
def testAnonymousVarsInInit(self):
|
||||||
|
|
||||||
|
class Model(training.Model):
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
super(Model, self).__init__()
|
||||||
|
self.w = resource_variable_ops.ResourceVariable(0.0)
|
||||||
|
self.b = resource_variable_ops.ResourceVariable(0.0)
|
||||||
|
self.vars = [self.w, self.b]
|
||||||
|
|
||||||
|
def call(self, x):
|
||||||
|
return x * self.w + self.b
|
||||||
|
|
||||||
|
with context.eager_mode():
|
||||||
|
model = Model()
|
||||||
|
optimizer = adam.AdamOptimizer(learning_rate=0.05)
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
checkpoint = trackable_utils.Checkpoint(
|
||||||
|
model=model, optimizer=optimizer)
|
||||||
|
for _ in range(2):
|
||||||
|
checkpoint.save(checkpoint_prefix)
|
||||||
|
with backprop.GradientTape() as tape:
|
||||||
|
loss = (constant_op.constant(1.)
|
||||||
|
- model(constant_op.constant(1.))) ** 2
|
||||||
|
grad = tape.gradient(loss, model.vars)
|
||||||
|
optimizer.apply_gradients(
|
||||||
|
[(g, v) for g, v in zip(grad, model.vars)])
|
||||||
|
|
||||||
|
@test_util.run_in_graph_and_eager_modes
|
||||||
|
def test_initialize_if_not_restoring(self):
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
optimizer_only_prefix = os.path.join(checkpoint_directory, "opt")
|
||||||
|
with test_util.device(use_gpu=True):
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
model=model, # Do not save the optimizer with the checkpoint.
|
||||||
|
global_step=training_util.get_or_create_global_step())
|
||||||
|
optimizer_checkpoint = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer)
|
||||||
|
|
||||||
|
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||||
|
checkpoint_directory)
|
||||||
|
status = root.restore(save_path=checkpoint_path)
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
train_fn = functools.partial(
|
||||||
|
optimizer.minimize,
|
||||||
|
functools.partial(model, input_value),
|
||||||
|
global_step=root.global_step)
|
||||||
|
if not context.executing_eagerly():
|
||||||
|
train_fn = functools.partial(self.evaluate, train_fn())
|
||||||
|
status.initialize_or_restore()
|
||||||
|
self.evaluate([v.initializer for v in optimizer.variables()])
|
||||||
|
train_fn()
|
||||||
|
model_save_path = root.save(file_prefix=checkpoint_prefix)
|
||||||
|
self.evaluate(optimizer.variables()[0].assign(42.))
|
||||||
|
optimizer_save_path = optimizer_checkpoint.save(optimizer_only_prefix)
|
||||||
|
|
||||||
|
# Restore into a graph with the optimizer
|
||||||
|
with test_util.device(use_gpu=True):
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model,
|
||||||
|
global_step=training_util.get_or_create_global_step())
|
||||||
|
status = root.restore(save_path=model_save_path)
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
train_fn = functools.partial(
|
||||||
|
optimizer.minimize,
|
||||||
|
functools.partial(model, input_value),
|
||||||
|
global_step=root.global_step)
|
||||||
|
if not context.executing_eagerly():
|
||||||
|
train_fn = functools.partial(self.evaluate, train_fn())
|
||||||
|
status.initialize_or_restore()
|
||||||
|
train_fn()
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
status.assert_consumed()
|
||||||
|
|
||||||
|
# Make sure initialization doesn't clobber later restores
|
||||||
|
with test_util.device(use_gpu=True):
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001, beta1=1.0)
|
||||||
|
root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model,
|
||||||
|
global_step=training_util.get_or_create_global_step())
|
||||||
|
opt_root = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer)
|
||||||
|
status = root.restore(save_path=model_save_path)
|
||||||
|
init_only_optimizer_status = opt_root.restore(save_path=None)
|
||||||
|
optimizer_status = opt_root.restore(save_path=optimizer_save_path)
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
train_fn = functools.partial(
|
||||||
|
optimizer.minimize,
|
||||||
|
functools.partial(model, input_value),
|
||||||
|
global_step=root.global_step)
|
||||||
|
if not context.executing_eagerly():
|
||||||
|
train_fn = functools.partial(self.evaluate, train_fn())
|
||||||
|
optimizer_status.run_restore_ops()
|
||||||
|
status.initialize_or_restore()
|
||||||
|
init_only_optimizer_status.initialize_or_restore()
|
||||||
|
train_fn()
|
||||||
|
self.assertEqual(42., self.evaluate(optimizer.variables()[0]))
|
||||||
|
|
||||||
|
|
||||||
|
class CheckpointCompatibilityTests(test.TestCase):
|
||||||
|
|
||||||
|
def _initialized_model(self):
|
||||||
|
input_value = constant_op.constant([[3.]])
|
||||||
|
model = MyModel()
|
||||||
|
optimizer = adam.AdamOptimizer(0.001)
|
||||||
|
optimizer_step = training_util.get_or_create_global_step()
|
||||||
|
root_trackable = trackable_utils.Checkpoint(
|
||||||
|
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
||||||
|
train_op = optimizer.minimize(
|
||||||
|
functools.partial(model, input_value),
|
||||||
|
global_step=optimizer_step)
|
||||||
|
self.evaluate(trackable_utils.gather_initializers(
|
||||||
|
root_trackable))
|
||||||
|
self.evaluate(train_op)
|
||||||
|
# A regular variable, a slot variable, and a non-slot Optimizer variable
|
||||||
|
# with known values to check when loading.
|
||||||
|
self.evaluate(model._named_dense.bias.assign([1.]))
|
||||||
|
self.evaluate(optimizer.get_slot(
|
||||||
|
var=model._named_dense.bias, name="m").assign([2.]))
|
||||||
|
beta1_power, _ = optimizer._get_beta_accumulators()
|
||||||
|
self.evaluate(beta1_power.assign(3.))
|
||||||
|
return root_trackable
|
||||||
|
|
||||||
|
def _set_sentinels(self, root_trackable):
|
||||||
|
self.evaluate(root_trackable.model._named_dense.bias.assign([101.]))
|
||||||
|
self.evaluate(
|
||||||
|
root_trackable.optimizer.get_slot(
|
||||||
|
var=root_trackable.model._named_dense.bias, name="m")
|
||||||
|
.assign([102.]))
|
||||||
|
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
||||||
|
self.evaluate(beta1_power.assign(103.))
|
||||||
|
|
||||||
|
def _check_sentinels(self, root_trackable):
|
||||||
|
self.assertAllEqual(
|
||||||
|
[1.], self.evaluate(root_trackable.model._named_dense.bias))
|
||||||
|
self.assertAllEqual([2.], self.evaluate(
|
||||||
|
root_trackable.optimizer.get_slot(
|
||||||
|
var=root_trackable.model._named_dense.bias, name="m")))
|
||||||
|
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
||||||
|
self.assertAllEqual(3., self.evaluate(beta1_power))
|
||||||
|
|
||||||
|
def _write_name_based_checkpoint(self):
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
with context.graph_mode():
|
||||||
|
save_graph = ops.Graph()
|
||||||
|
with save_graph.as_default(), self.session(
|
||||||
|
graph=save_graph) as session:
|
||||||
|
root = self._initialized_model()
|
||||||
|
name_saver = saver_lib.Saver()
|
||||||
|
return name_saver.save(
|
||||||
|
sess=session, save_path=checkpoint_prefix,
|
||||||
|
global_step=root.optimizer_step)
|
||||||
|
|
||||||
|
@test_util.run_in_graph_and_eager_modes
|
||||||
|
def testLoadFromNameBasedSaver(self):
|
||||||
|
"""Save a name-based checkpoint, load it using the object-based API."""
|
||||||
|
with test_util.device(use_gpu=True):
|
||||||
|
save_path = self._write_name_based_checkpoint()
|
||||||
|
root = self._initialized_model()
|
||||||
|
self._set_sentinels(root)
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
self._check_sentinels(root)
|
||||||
|
object_saver = trackable_utils.TrackableSaver(
|
||||||
|
graph_view.ObjectGraphView(root))
|
||||||
|
self._set_sentinels(root)
|
||||||
|
status = object_saver.restore(save_path)
|
||||||
|
if context.executing_eagerly():
|
||||||
|
self._check_sentinels(root)
|
||||||
|
if context.executing_eagerly():
|
||||||
|
status.assert_consumed()
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
status.assert_nontrivial_match()
|
||||||
|
else:
|
||||||
|
# When graph building, we haven't read any keys, so we don't know
|
||||||
|
# whether the restore will be complete.
|
||||||
|
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||||
|
status.assert_consumed()
|
||||||
|
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||||
|
status.assert_nontrivial_match()
|
||||||
|
status.run_restore_ops()
|
||||||
|
self._check_sentinels(root)
|
||||||
|
self._set_sentinels(root)
|
||||||
|
status = object_saver.restore(save_path)
|
||||||
|
status.initialize_or_restore()
|
||||||
|
self._check_sentinels(root)
|
||||||
|
# Check that there is no error when keys are missing from the name-based
|
||||||
|
# checkpoint.
|
||||||
|
root.not_in_name_checkpoint = resource_variable_ops.ResourceVariable([1.])
|
||||||
|
status = object_saver.restore(save_path)
|
||||||
|
with self.assertRaises(AssertionError):
|
||||||
|
status.assert_existing_objects_matched()
|
||||||
|
|
||||||
|
def testSaveGraphLoadEager(self):
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
with context.graph_mode():
|
||||||
|
save_graph = ops.Graph()
|
||||||
|
with save_graph.as_default(), self.session(
|
||||||
|
graph=save_graph):
|
||||||
|
root = self._initialized_model()
|
||||||
|
save_path = root.save(file_prefix=checkpoint_prefix)
|
||||||
|
with context.eager_mode():
|
||||||
|
root = self._initialized_model()
|
||||||
|
self._set_sentinels(root)
|
||||||
|
root.restore(save_path).assert_consumed()
|
||||||
|
self._check_sentinels(root)
|
||||||
|
|
||||||
|
def testSaveEagerLoadGraph(self):
|
||||||
|
checkpoint_directory = self.get_temp_dir()
|
||||||
|
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||||
|
with context.eager_mode():
|
||||||
|
root = self._initialized_model()
|
||||||
|
save_path = root.save(file_prefix=checkpoint_prefix)
|
||||||
|
with context.graph_mode():
|
||||||
|
save_graph = ops.Graph()
|
||||||
|
with save_graph.as_default(), self.session(
|
||||||
|
graph=save_graph):
|
||||||
|
root = self._initialized_model()
|
||||||
|
self._set_sentinels(root)
|
||||||
|
root.restore(save_path).assert_consumed().run_restore_ops()
|
||||||
|
self._check_sentinels(root)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
test.main()
|
@ -223,34 +223,19 @@ tf_py_test(
|
|||||||
"notsan", # b/74395663
|
"notsan", # b/74395663
|
||||||
],
|
],
|
||||||
deps = [
|
deps = [
|
||||||
":base",
|
|
||||||
":graph_view",
|
|
||||||
":tracking",
|
":tracking",
|
||||||
":util",
|
":util",
|
||||||
"//tensorflow/python:checkpoint_management",
|
|
||||||
"//tensorflow/python:constant_op",
|
|
||||||
"//tensorflow/python:control_flow_ops",
|
|
||||||
"//tensorflow/python:dtypes",
|
|
||||||
"//tensorflow/python:framework_ops",
|
"//tensorflow/python:framework_ops",
|
||||||
"//tensorflow/python:framework_test_lib",
|
"//tensorflow/python:framework_test_lib",
|
||||||
"//tensorflow/python:init_ops",
|
"//tensorflow/python:init_ops",
|
||||||
"//tensorflow/python:pywrap_tensorflow",
|
|
||||||
"//tensorflow/python:resource_variable_ops",
|
"//tensorflow/python:resource_variable_ops",
|
||||||
"//tensorflow/python:saver",
|
|
||||||
"//tensorflow/python:session",
|
"//tensorflow/python:session",
|
||||||
"//tensorflow/python:state_ops",
|
"//tensorflow/python:state_ops",
|
||||||
"//tensorflow/python:template",
|
"//tensorflow/python:template",
|
||||||
"//tensorflow/python:training",
|
"//tensorflow/python:training",
|
||||||
"//tensorflow/python:training_util",
|
|
||||||
"//tensorflow/python:variable_scope",
|
"//tensorflow/python:variable_scope",
|
||||||
"//tensorflow/python/distribute:mirrored_strategy",
|
|
||||||
"//tensorflow/python/eager:backprop",
|
|
||||||
"//tensorflow/python/eager:context",
|
"//tensorflow/python/eager:context",
|
||||||
"//tensorflow/python/eager:def_function",
|
|
||||||
"//tensorflow/python/eager:test",
|
"//tensorflow/python/eager:test",
|
||||||
"//tensorflow/python/keras:engine",
|
|
||||||
"//tensorflow/python/keras/layers",
|
|
||||||
"@absl_py//absl/testing:parameterized",
|
|
||||||
"@six_archive//:six",
|
"@six_archive//:six",
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
@ -17,486 +17,27 @@ from __future__ import absolute_import
|
|||||||
from __future__ import division
|
from __future__ import division
|
||||||
from __future__ import print_function
|
from __future__ import print_function
|
||||||
|
|
||||||
import functools
|
|
||||||
import os
|
import os
|
||||||
|
|
||||||
import six
|
import six
|
||||||
|
|
||||||
from tensorflow.python.client import session as session_lib
|
from tensorflow.python.client import session as session_lib
|
||||||
from tensorflow.python.distribute import mirrored_strategy
|
|
||||||
from tensorflow.python.eager import backprop
|
|
||||||
from tensorflow.python.eager import context
|
from tensorflow.python.eager import context
|
||||||
from tensorflow.python.eager import def_function
|
|
||||||
from tensorflow.python.eager import test
|
from tensorflow.python.eager import test
|
||||||
from tensorflow.python.framework import constant_op
|
|
||||||
from tensorflow.python.framework import dtypes
|
|
||||||
from tensorflow.python.framework import ops
|
from tensorflow.python.framework import ops
|
||||||
from tensorflow.python.framework import test_util
|
from tensorflow.python.framework import test_util
|
||||||
from tensorflow.python.keras.engine import training
|
|
||||||
from tensorflow.python.keras.layers import core
|
|
||||||
from tensorflow.python.ops import init_ops
|
from tensorflow.python.ops import init_ops
|
||||||
from tensorflow.python.ops import resource_variable_ops
|
from tensorflow.python.ops import resource_variable_ops
|
||||||
from tensorflow.python.ops import state_ops
|
from tensorflow.python.ops import state_ops
|
||||||
from tensorflow.python.ops import template
|
from tensorflow.python.ops import template
|
||||||
from tensorflow.python.ops import variable_scope
|
from tensorflow.python.ops import variable_scope
|
||||||
from tensorflow.python.training import adam
|
from tensorflow.python.training import adam
|
||||||
from tensorflow.python.training import checkpoint_management
|
|
||||||
from tensorflow.python.training import saver as saver_lib
|
|
||||||
from tensorflow.python.training import training_util
|
|
||||||
from tensorflow.python.training.tracking import graph_view
|
|
||||||
from tensorflow.python.training.tracking import tracking
|
from tensorflow.python.training.tracking import tracking
|
||||||
from tensorflow.python.training.tracking import util as trackable_utils
|
from tensorflow.python.training.tracking import util as trackable_utils
|
||||||
|
|
||||||
|
|
||||||
class NonLayerTrackable(tracking.AutoTrackable):
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
super(NonLayerTrackable, self).__init__()
|
|
||||||
self.a_variable = trackable_utils.add_variable(
|
|
||||||
self, name="a_variable", shape=[])
|
|
||||||
|
|
||||||
|
|
||||||
# pylint: disable=not-callable
|
|
||||||
class MyModel(training.Model):
|
|
||||||
"""A concrete Model for testing."""
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
super(MyModel, self).__init__()
|
|
||||||
self._named_dense = core.Dense(1, use_bias=True)
|
|
||||||
self._second = core.Dense(1, use_bias=False)
|
|
||||||
# We can still track Trackables which aren't Layers.
|
|
||||||
self._non_layer = NonLayerTrackable()
|
|
||||||
|
|
||||||
def call(self, values):
|
|
||||||
ret = self._second(self._named_dense(values))
|
|
||||||
return ret
|
|
||||||
|
|
||||||
|
|
||||||
class CheckpointingTests(test.TestCase):
|
class CheckpointingTests(test.TestCase):
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes(assert_no_eager_garbage=True)
|
|
||||||
def testNamingWithOptimizer(self):
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
model = MyModel()
|
|
||||||
# A nuisance Model using the same optimizer. Its slot variables should not
|
|
||||||
# go in the checkpoint, since it is never depended on.
|
|
||||||
other_model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
optimizer_step = training_util.get_or_create_global_step()
|
|
||||||
root_trackable = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
|
||||||
if context.executing_eagerly():
|
|
||||||
optimizer.minimize(
|
|
||||||
lambda: model(input_value),
|
|
||||||
global_step=optimizer_step)
|
|
||||||
optimizer.minimize(
|
|
||||||
lambda: other_model(input_value),
|
|
||||||
global_step=optimizer_step)
|
|
||||||
else:
|
|
||||||
train_op = optimizer.minimize(
|
|
||||||
model(input_value), global_step=optimizer_step)
|
|
||||||
optimizer.minimize(
|
|
||||||
other_model(input_value),
|
|
||||||
global_step=optimizer_step)
|
|
||||||
self.evaluate(trackable_utils.gather_initializers(
|
|
||||||
root_trackable))
|
|
||||||
self.evaluate(train_op)
|
|
||||||
named_variables, serialized_graph, _ = graph_view.ObjectGraphView(
|
|
||||||
root_trackable).serialize_object_graph()
|
|
||||||
expected_checkpoint_names = (
|
|
||||||
# Created in the root node, so no prefix.
|
|
||||||
"optimizer_step",
|
|
||||||
"model/_second/kernel",
|
|
||||||
"model/_named_dense/kernel",
|
|
||||||
"model/_named_dense/bias",
|
|
||||||
# non-Layer dependency of the model
|
|
||||||
"model/_non_layer/a_variable",
|
|
||||||
# The optimizer creates two non-slot variables
|
|
||||||
"optimizer/beta1_power",
|
|
||||||
"optimizer/beta2_power",
|
|
||||||
# Slot variables
|
|
||||||
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
|
||||||
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
|
||||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
|
||||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
|
||||||
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/m",
|
|
||||||
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/v",
|
|
||||||
)
|
|
||||||
suffix = "/.ATTRIBUTES/VARIABLE_VALUE"
|
|
||||||
expected_checkpoint_names = [
|
|
||||||
name + suffix for name in expected_checkpoint_names]
|
|
||||||
named_variables = {v.name: v for v in named_variables}
|
|
||||||
six.assertCountEqual(self, expected_checkpoint_names,
|
|
||||||
named_variables.keys())
|
|
||||||
# Check that we've mapped to the right variable objects (not exhaustive)
|
|
||||||
self.assertEqual(
|
|
||||||
"global_step",
|
|
||||||
named_variables["optimizer_step" + suffix].full_name)
|
|
||||||
self.assertEqual(
|
|
||||||
"my_model/dense_1/kernel",
|
|
||||||
named_variables["model/_second/kernel" + suffix].full_name)
|
|
||||||
self.assertEqual(
|
|
||||||
"my_model/dense/kernel",
|
|
||||||
named_variables["model/_named_dense/kernel" + suffix].full_name)
|
|
||||||
self.assertEqual(
|
|
||||||
"beta1_power",
|
|
||||||
named_variables["optimizer/beta1_power" + suffix].full_name)
|
|
||||||
self.assertEqual(
|
|
||||||
"beta2_power",
|
|
||||||
named_variables["optimizer/beta2_power" + suffix].full_name)
|
|
||||||
# Spot check the generated protocol buffers.
|
|
||||||
self.assertEqual("optimizer",
|
|
||||||
serialized_graph.nodes[0].children[1].local_name)
|
|
||||||
optimizer_node = serialized_graph.nodes[serialized_graph.nodes[0].children[
|
|
||||||
1].node_id]
|
|
||||||
self.assertEqual("beta1_power",
|
|
||||||
optimizer_node.children[0].local_name)
|
|
||||||
self.assertEqual("beta1_power",
|
|
||||||
serialized_graph.nodes[optimizer_node.children[0].node_id]
|
|
||||||
.attributes[0].full_name)
|
|
||||||
self.assertEqual(
|
|
||||||
"my_model/dense/kernel",
|
|
||||||
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
|
||||||
.original_variable_node_id]
|
|
||||||
.attributes[0].full_name)
|
|
||||||
# We strip off the :0 suffix, as variable.name-based saving does.
|
|
||||||
self.assertEqual(
|
|
||||||
"my_model/dense/kernel/Adam",
|
|
||||||
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
|
||||||
.slot_variable_node_id]
|
|
||||||
.attributes[0].full_name)
|
|
||||||
self.assertEqual(
|
|
||||||
"my_model/dense/kernel/Adam:0",
|
|
||||||
optimizer.get_slot(
|
|
||||||
var=model._named_dense.kernel,
|
|
||||||
name="m").name)
|
|
||||||
self.assertEqual(
|
|
||||||
"model/_named_dense/kernel" + suffix,
|
|
||||||
serialized_graph.nodes[
|
|
||||||
optimizer_node.slot_variables[0]
|
|
||||||
.original_variable_node_id].attributes[0].checkpoint_key)
|
|
||||||
self.assertEqual("m", optimizer_node.slot_variables[0].slot_name)
|
|
||||||
self.assertEqual(
|
|
||||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m" + suffix,
|
|
||||||
serialized_graph.nodes[
|
|
||||||
optimizer_node.slot_variables[0]
|
|
||||||
.slot_variable_node_id].attributes[0].checkpoint_key)
|
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
|
||||||
def testSaveRestore(self):
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root_trackable = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model)
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
if context.executing_eagerly():
|
|
||||||
optimizer.minimize(
|
|
||||||
lambda: model(input_value))
|
|
||||||
else:
|
|
||||||
train_op = optimizer.minimize(model(input_value))
|
|
||||||
# TODO(allenl): Make initialization more pleasant when graph building.
|
|
||||||
root_trackable.save_counter # pylint: disable=pointless-statement
|
|
||||||
self.evaluate(trackable_utils.gather_initializers(
|
|
||||||
root_trackable))
|
|
||||||
self.evaluate(train_op)
|
|
||||||
prefix = os.path.join(self.get_temp_dir(), "ckpt")
|
|
||||||
self.evaluate(state_ops.assign(model._named_dense.variables[1], [42.]))
|
|
||||||
m_bias_slot = optimizer.get_slot(model._named_dense.variables[1], "m")
|
|
||||||
self.evaluate(state_ops.assign(m_bias_slot, [1.5]))
|
|
||||||
save_path = root_trackable.save(file_prefix=prefix)
|
|
||||||
self.evaluate(state_ops.assign(model._named_dense.variables[1], [43.]))
|
|
||||||
self.evaluate(state_ops.assign(root_trackable.save_counter, 3))
|
|
||||||
optimizer_variables = self.evaluate(optimizer.variables())
|
|
||||||
self.evaluate(state_ops.assign(m_bias_slot, [-2.]))
|
|
||||||
# Immediate restoration
|
|
||||||
status = root_trackable.restore(save_path=save_path).assert_consumed()
|
|
||||||
status.run_restore_ops()
|
|
||||||
self.assertAllEqual([42.], self.evaluate(model._named_dense.variables[1]))
|
|
||||||
self.assertAllEqual(1, self.evaluate(root_trackable.save_counter))
|
|
||||||
self.assertAllEqual([1.5], self.evaluate(m_bias_slot))
|
|
||||||
if not context.executing_eagerly():
|
|
||||||
return # Restore-on-create is only supported when executing eagerly
|
|
||||||
on_create_model = MyModel()
|
|
||||||
on_create_optimizer = adam.AdamOptimizer(
|
|
||||||
0.001,
|
|
||||||
# Preserve beta1_power and beta2_power when applying gradients so we can
|
|
||||||
# test that they've been restored correctly.
|
|
||||||
beta1=1.0,
|
|
||||||
beta2=1.0)
|
|
||||||
on_create_root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=on_create_optimizer, model=on_create_model)
|
|
||||||
# Deferred restoration
|
|
||||||
status = on_create_root.restore(save_path=save_path)
|
|
||||||
status.assert_nontrivial_match()
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
status.assert_consumed()
|
|
||||||
on_create_model(constant_op.constant([[3.]])) # create variables
|
|
||||||
self.assertAllEqual(1, self.evaluate(on_create_root.save_counter))
|
|
||||||
self.assertAllEqual([42.],
|
|
||||||
self.evaluate(
|
|
||||||
on_create_model._named_dense.variables[1]))
|
|
||||||
on_create_m_bias_slot = on_create_optimizer.get_slot(
|
|
||||||
on_create_model._named_dense.variables[1], "m")
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
status.assert_consumed()
|
|
||||||
# Optimizer slot variables are created when the original variable is
|
|
||||||
# restored.
|
|
||||||
self.assertAllEqual([1.5], self.evaluate(on_create_m_bias_slot))
|
|
||||||
self.assertAllEqual(optimizer_variables[2:],
|
|
||||||
self.evaluate(on_create_optimizer.variables()))
|
|
||||||
dummy_var = resource_variable_ops.ResourceVariable([1.])
|
|
||||||
on_create_optimizer.minimize(loss=dummy_var.read_value)
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
status.assert_consumed()
|
|
||||||
beta1_power, beta2_power = on_create_optimizer._get_beta_accumulators()
|
|
||||||
self.assertAllEqual(optimizer_variables[0], self.evaluate(beta1_power))
|
|
||||||
self.assertAllEqual(optimizer_variables[1], self.evaluate(beta2_power))
|
|
||||||
|
|
||||||
# TODO(allenl): Debug garbage created by this test in python3.
|
|
||||||
def testDeferredRestorationUsageEager(self):
|
|
||||||
"""An idiomatic eager execution example."""
|
|
||||||
num_training_steps = 10
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
for training_continuation in range(3):
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model,
|
|
||||||
optimizer_step=training_util.get_or_create_global_step())
|
|
||||||
root.restore(checkpoint_management.latest_checkpoint(
|
|
||||||
checkpoint_directory))
|
|
||||||
for _ in range(num_training_steps):
|
|
||||||
# TODO(allenl): Use a Dataset and serialize/checkpoint it.
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
optimizer.minimize(
|
|
||||||
lambda: model(input_value), # pylint: disable=cell-var-from-loop
|
|
||||||
global_step=root.optimizer_step)
|
|
||||||
root.save(file_prefix=checkpoint_prefix)
|
|
||||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
|
||||||
root.optimizer_step.numpy())
|
|
||||||
|
|
||||||
def testEagerDistributionStrategy(self):
|
|
||||||
num_training_steps = 10
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
|
|
||||||
def _train_fn(optimizer, model):
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
optimizer.minimize(
|
|
||||||
functools.partial(model, input_value),
|
|
||||||
global_step=root.optimizer_step)
|
|
||||||
|
|
||||||
strategy = mirrored_strategy.MirroredStrategy()
|
|
||||||
with strategy.scope():
|
|
||||||
for training_continuation in range(3):
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer,
|
|
||||||
model=model,
|
|
||||||
optimizer_step=training_util.get_or_create_global_step())
|
|
||||||
root.restore(
|
|
||||||
checkpoint_management.latest_checkpoint(checkpoint_directory))
|
|
||||||
|
|
||||||
for _ in range(num_training_steps):
|
|
||||||
strategy.extended.call_for_each_replica(
|
|
||||||
functools.partial(_train_fn, optimizer, model))
|
|
||||||
root.save(file_prefix=checkpoint_prefix)
|
|
||||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
|
||||||
root.optimizer_step.numpy())
|
|
||||||
|
|
||||||
def testGraphDistributionStrategy(self):
|
|
||||||
self.skipTest("b/121381184")
|
|
||||||
num_training_steps = 10
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
|
|
||||||
def _train_fn(optimizer, model):
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
return optimizer.minimize(
|
|
||||||
functools.partial(model, input_value),
|
|
||||||
global_step=root.optimizer_step)
|
|
||||||
|
|
||||||
for training_continuation in range(3):
|
|
||||||
with ops.Graph().as_default():
|
|
||||||
strategy = mirrored_strategy.MirroredStrategy()
|
|
||||||
with strategy.scope():
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model,
|
|
||||||
optimizer_step=training_util.get_or_create_global_step())
|
|
||||||
status = root.restore(checkpoint_management.latest_checkpoint(
|
|
||||||
checkpoint_directory))
|
|
||||||
train_op = strategy.extended.call_for_each_replica(
|
|
||||||
functools.partial(_train_fn, optimizer, model))
|
|
||||||
with self.session() as session:
|
|
||||||
if training_continuation > 0:
|
|
||||||
status.assert_consumed()
|
|
||||||
status.initialize_or_restore()
|
|
||||||
for _ in range(num_training_steps):
|
|
||||||
session.run(train_op)
|
|
||||||
root.save(file_prefix=checkpoint_prefix)
|
|
||||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
|
||||||
root.optimizer_step.numpy())
|
|
||||||
|
|
||||||
def testUsageGraph(self):
|
|
||||||
"""Expected usage when graph building."""
|
|
||||||
with context.graph_mode():
|
|
||||||
num_training_steps = 10
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
for training_continuation in range(3):
|
|
||||||
with ops.Graph().as_default():
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root = trackable_utils.CheckpointV1(
|
|
||||||
optimizer=optimizer, model=model,
|
|
||||||
global_step=training_util.get_or_create_global_step())
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
train_op = optimizer.minimize(
|
|
||||||
model(input_value),
|
|
||||||
global_step=root.global_step)
|
|
||||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
|
||||||
checkpoint_directory)
|
|
||||||
with self.session(graph=ops.get_default_graph()) as session:
|
|
||||||
status = root.restore(save_path=checkpoint_path)
|
|
||||||
status.initialize_or_restore(session=session)
|
|
||||||
if checkpoint_path is None:
|
|
||||||
self.assertEqual(0, training_continuation)
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
status.assert_consumed()
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
else:
|
|
||||||
status.assert_consumed()
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
for _ in range(num_training_steps):
|
|
||||||
session.run(train_op)
|
|
||||||
root.save(file_prefix=checkpoint_prefix, session=session)
|
|
||||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
|
||||||
session.run(root.global_step))
|
|
||||||
self.assertEqual(training_continuation + 1,
|
|
||||||
session.run(root.save_counter))
|
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
|
||||||
def testAgnosticUsage(self):
|
|
||||||
"""Graph/eager agnostic usage."""
|
|
||||||
# Does create garbage when executing eagerly due to ops.Graph() creation.
|
|
||||||
num_training_steps = 10
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
for training_continuation in range(3):
|
|
||||||
with test_util.device(use_gpu=True):
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model,
|
|
||||||
global_step=training_util.get_or_create_global_step())
|
|
||||||
manager = checkpoint_management.CheckpointManager(
|
|
||||||
root, checkpoint_directory, max_to_keep=1)
|
|
||||||
status = root.restore(save_path=manager.latest_checkpoint)
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
train_fn = functools.partial(
|
|
||||||
optimizer.minimize,
|
|
||||||
functools.partial(model, input_value),
|
|
||||||
global_step=root.global_step)
|
|
||||||
if not context.executing_eagerly():
|
|
||||||
train_fn = functools.partial(self.evaluate, train_fn())
|
|
||||||
status.initialize_or_restore()
|
|
||||||
for _ in range(num_training_steps):
|
|
||||||
train_fn()
|
|
||||||
manager.save()
|
|
||||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
|
||||||
self.evaluate(root.global_step))
|
|
||||||
self.assertEqual(training_continuation + 1,
|
|
||||||
self.evaluate(root.save_counter))
|
|
||||||
|
|
||||||
# pylint: disable=cell-var-from-loop
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
|
||||||
def testWithDefun(self):
|
|
||||||
num_training_steps = 2
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
for training_continuation in range(3):
|
|
||||||
with test_util.device(use_gpu=True):
|
|
||||||
model = MyModel()
|
|
||||||
# Don't actually train so we can test variable values
|
|
||||||
optimizer = adam.AdamOptimizer(0.)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model,
|
|
||||||
global_step=training_util.get_or_create_global_step())
|
|
||||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
|
||||||
checkpoint_directory)
|
|
||||||
status = root.restore(save_path=checkpoint_path)
|
|
||||||
def train_fn():
|
|
||||||
@def_function.function
|
|
||||||
def _call_model(x):
|
|
||||||
return model(x)
|
|
||||||
with backprop.GradientTape() as tape:
|
|
||||||
loss = _call_model(constant_op.constant([[3.]]))
|
|
||||||
gradients = tape.gradient(loss, model.variables)
|
|
||||||
return optimizer.apply_gradients(zip(gradients, model.variables),
|
|
||||||
global_step=root.global_step)
|
|
||||||
if not context.executing_eagerly():
|
|
||||||
train_fn = functools.partial(
|
|
||||||
self.evaluate, train_fn())
|
|
||||||
status.initialize_or_restore()
|
|
||||||
for _ in range(num_training_steps):
|
|
||||||
train_fn()
|
|
||||||
if training_continuation > 0:
|
|
||||||
status.assert_consumed()
|
|
||||||
self.assertAllClose([[42.]], self.evaluate(model.variables[0]))
|
|
||||||
else:
|
|
||||||
self.evaluate(model.variables[0].assign([[42.]]))
|
|
||||||
root.save(file_prefix=checkpoint_prefix)
|
|
||||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
|
||||||
self.evaluate(root.global_step))
|
|
||||||
self.assertEqual(training_continuation + 1,
|
|
||||||
self.evaluate(root.save_counter))
|
|
||||||
# pylint: enable=cell-var-from-loop
|
|
||||||
|
|
||||||
def _get_checkpoint_name(self, name):
|
|
||||||
root = tracking.AutoTrackable()
|
|
||||||
trackable_utils.add_variable(
|
|
||||||
root, name=name, shape=[1, 2], dtype=dtypes.float64)
|
|
||||||
(named_variable,), _, _ = trackable_utils._serialize_object_graph(
|
|
||||||
root, saveables_cache=None)
|
|
||||||
with ops.name_scope("root/" + named_variable.name):
|
|
||||||
pass # Make sure we can use this as an op name if we prefix it.
|
|
||||||
return named_variable.name
|
|
||||||
|
|
||||||
def testAnonymousVarsInInit(self):
|
|
||||||
|
|
||||||
class Model(training.Model):
|
|
||||||
|
|
||||||
def __init__(self):
|
|
||||||
super(Model, self).__init__()
|
|
||||||
self.w = resource_variable_ops.ResourceVariable(0.0)
|
|
||||||
self.b = resource_variable_ops.ResourceVariable(0.0)
|
|
||||||
self.vars = [self.w, self.b]
|
|
||||||
|
|
||||||
def call(self, x):
|
|
||||||
return x * self.w + self.b
|
|
||||||
|
|
||||||
with context.eager_mode():
|
|
||||||
model = Model()
|
|
||||||
optimizer = adam.AdamOptimizer(learning_rate=0.05)
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
checkpoint = trackable_utils.Checkpoint(
|
|
||||||
model=model, optimizer=optimizer)
|
|
||||||
for _ in range(2):
|
|
||||||
checkpoint.save(checkpoint_prefix)
|
|
||||||
with backprop.GradientTape() as tape:
|
|
||||||
loss = (constant_op.constant(1.)
|
|
||||||
- model(constant_op.constant(1.))) ** 2
|
|
||||||
grad = tape.gradient(loss, model.vars)
|
|
||||||
optimizer.apply_gradients(
|
|
||||||
[(g, v) for g, v in zip(grad, model.vars)])
|
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
@test_util.run_in_graph_and_eager_modes
|
||||||
def testDeferredSlotRestoration(self):
|
def testDeferredSlotRestoration(self):
|
||||||
checkpoint_directory = self.get_temp_dir()
|
checkpoint_directory = self.get_temp_dir()
|
||||||
@ -652,84 +193,6 @@ class CheckpointingTests(test.TestCase):
|
|||||||
beta1_power, _ = optimizer._get_beta_accumulators()
|
beta1_power, _ = optimizer._get_beta_accumulators()
|
||||||
self.assertAllEqual(3., self.evaluate(beta1_power))
|
self.assertAllEqual(3., self.evaluate(beta1_power))
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
|
||||||
def test_initialize_if_not_restoring(self):
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
optimizer_only_prefix = os.path.join(checkpoint_directory, "opt")
|
|
||||||
with test_util.device(use_gpu=True):
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
model=model, # Do not save the optimizer with the checkpoint.
|
|
||||||
global_step=training_util.get_or_create_global_step())
|
|
||||||
optimizer_checkpoint = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer)
|
|
||||||
|
|
||||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
|
||||||
checkpoint_directory)
|
|
||||||
status = root.restore(save_path=checkpoint_path)
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
train_fn = functools.partial(
|
|
||||||
optimizer.minimize,
|
|
||||||
functools.partial(model, input_value),
|
|
||||||
global_step=root.global_step)
|
|
||||||
if not context.executing_eagerly():
|
|
||||||
train_fn = functools.partial(self.evaluate, train_fn())
|
|
||||||
status.initialize_or_restore()
|
|
||||||
self.evaluate([v.initializer for v in optimizer.variables()])
|
|
||||||
train_fn()
|
|
||||||
model_save_path = root.save(file_prefix=checkpoint_prefix)
|
|
||||||
self.evaluate(optimizer.variables()[0].assign(42.))
|
|
||||||
optimizer_save_path = optimizer_checkpoint.save(optimizer_only_prefix)
|
|
||||||
|
|
||||||
# Restore into a graph with the optimizer
|
|
||||||
with test_util.device(use_gpu=True):
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model,
|
|
||||||
global_step=training_util.get_or_create_global_step())
|
|
||||||
status = root.restore(save_path=model_save_path)
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
train_fn = functools.partial(
|
|
||||||
optimizer.minimize,
|
|
||||||
functools.partial(model, input_value),
|
|
||||||
global_step=root.global_step)
|
|
||||||
if not context.executing_eagerly():
|
|
||||||
train_fn = functools.partial(self.evaluate, train_fn())
|
|
||||||
status.initialize_or_restore()
|
|
||||||
train_fn()
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
status.assert_consumed()
|
|
||||||
|
|
||||||
# Make sure initialization doesn't clobber later restores
|
|
||||||
with test_util.device(use_gpu=True):
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001, beta1=1.0)
|
|
||||||
root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model,
|
|
||||||
global_step=training_util.get_or_create_global_step())
|
|
||||||
opt_root = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer)
|
|
||||||
status = root.restore(save_path=model_save_path)
|
|
||||||
init_only_optimizer_status = opt_root.restore(save_path=None)
|
|
||||||
optimizer_status = opt_root.restore(save_path=optimizer_save_path)
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
train_fn = functools.partial(
|
|
||||||
optimizer.minimize,
|
|
||||||
functools.partial(model, input_value),
|
|
||||||
global_step=root.global_step)
|
|
||||||
if not context.executing_eagerly():
|
|
||||||
train_fn = functools.partial(self.evaluate, train_fn())
|
|
||||||
optimizer_status.run_restore_ops()
|
|
||||||
status.initialize_or_restore()
|
|
||||||
init_only_optimizer_status.initialize_or_restore()
|
|
||||||
train_fn()
|
|
||||||
self.assertEqual(42., self.evaluate(optimizer.variables()[0]))
|
|
||||||
|
|
||||||
|
|
||||||
class _ManualScope(tracking.AutoTrackable):
|
class _ManualScope(tracking.AutoTrackable):
|
||||||
|
|
||||||
@ -797,132 +260,5 @@ class TemplateTests(test.TestCase):
|
|||||||
self.assertAllEqual([14.], self.evaluate(var2))
|
self.assertAllEqual([14.], self.evaluate(var2))
|
||||||
|
|
||||||
|
|
||||||
class CheckpointCompatibilityTests(test.TestCase):
|
|
||||||
|
|
||||||
def _initialized_model(self):
|
|
||||||
input_value = constant_op.constant([[3.]])
|
|
||||||
model = MyModel()
|
|
||||||
optimizer = adam.AdamOptimizer(0.001)
|
|
||||||
optimizer_step = training_util.get_or_create_global_step()
|
|
||||||
root_trackable = trackable_utils.Checkpoint(
|
|
||||||
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
|
||||||
train_op = optimizer.minimize(
|
|
||||||
functools.partial(model, input_value),
|
|
||||||
global_step=optimizer_step)
|
|
||||||
self.evaluate(trackable_utils.gather_initializers(
|
|
||||||
root_trackable))
|
|
||||||
self.evaluate(train_op)
|
|
||||||
# A regular variable, a slot variable, and a non-slot Optimizer variable
|
|
||||||
# with known values to check when loading.
|
|
||||||
self.evaluate(model._named_dense.bias.assign([1.]))
|
|
||||||
self.evaluate(optimizer.get_slot(
|
|
||||||
var=model._named_dense.bias, name="m").assign([2.]))
|
|
||||||
beta1_power, _ = optimizer._get_beta_accumulators()
|
|
||||||
self.evaluate(beta1_power.assign(3.))
|
|
||||||
return root_trackable
|
|
||||||
|
|
||||||
def _set_sentinels(self, root_trackable):
|
|
||||||
self.evaluate(root_trackable.model._named_dense.bias.assign([101.]))
|
|
||||||
self.evaluate(
|
|
||||||
root_trackable.optimizer.get_slot(
|
|
||||||
var=root_trackable.model._named_dense.bias, name="m")
|
|
||||||
.assign([102.]))
|
|
||||||
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
|
||||||
self.evaluate(beta1_power.assign(103.))
|
|
||||||
|
|
||||||
def _check_sentinels(self, root_trackable):
|
|
||||||
self.assertAllEqual(
|
|
||||||
[1.], self.evaluate(root_trackable.model._named_dense.bias))
|
|
||||||
self.assertAllEqual([2.], self.evaluate(
|
|
||||||
root_trackable.optimizer.get_slot(
|
|
||||||
var=root_trackable.model._named_dense.bias, name="m")))
|
|
||||||
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
|
||||||
self.assertAllEqual(3., self.evaluate(beta1_power))
|
|
||||||
|
|
||||||
def _write_name_based_checkpoint(self):
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
with context.graph_mode():
|
|
||||||
save_graph = ops.Graph()
|
|
||||||
with save_graph.as_default(), self.session(
|
|
||||||
graph=save_graph) as session:
|
|
||||||
root = self._initialized_model()
|
|
||||||
name_saver = saver_lib.Saver()
|
|
||||||
return name_saver.save(
|
|
||||||
sess=session, save_path=checkpoint_prefix,
|
|
||||||
global_step=root.optimizer_step)
|
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
|
||||||
def testLoadFromNameBasedSaver(self):
|
|
||||||
"""Save a name-based checkpoint, load it using the object-based API."""
|
|
||||||
with test_util.device(use_gpu=True):
|
|
||||||
save_path = self._write_name_based_checkpoint()
|
|
||||||
root = self._initialized_model()
|
|
||||||
self._set_sentinels(root)
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
self._check_sentinels(root)
|
|
||||||
object_saver = trackable_utils.TrackableSaver(
|
|
||||||
graph_view.ObjectGraphView(root))
|
|
||||||
self._set_sentinels(root)
|
|
||||||
status = object_saver.restore(save_path)
|
|
||||||
if context.executing_eagerly():
|
|
||||||
self._check_sentinels(root)
|
|
||||||
if context.executing_eagerly():
|
|
||||||
status.assert_consumed()
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
status.assert_nontrivial_match()
|
|
||||||
else:
|
|
||||||
# When graph building, we haven't read any keys, so we don't know
|
|
||||||
# whether the restore will be complete.
|
|
||||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
|
||||||
status.assert_consumed()
|
|
||||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
|
||||||
status.assert_nontrivial_match()
|
|
||||||
status.run_restore_ops()
|
|
||||||
self._check_sentinels(root)
|
|
||||||
self._set_sentinels(root)
|
|
||||||
status = object_saver.restore(save_path)
|
|
||||||
status.initialize_or_restore()
|
|
||||||
self._check_sentinels(root)
|
|
||||||
# Check that there is no error when keys are missing from the name-based
|
|
||||||
# checkpoint.
|
|
||||||
root.not_in_name_checkpoint = resource_variable_ops.ResourceVariable([1.])
|
|
||||||
status = object_saver.restore(save_path)
|
|
||||||
with self.assertRaises(AssertionError):
|
|
||||||
status.assert_existing_objects_matched()
|
|
||||||
|
|
||||||
def testSaveGraphLoadEager(self):
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
with context.graph_mode():
|
|
||||||
save_graph = ops.Graph()
|
|
||||||
with save_graph.as_default(), self.session(
|
|
||||||
graph=save_graph):
|
|
||||||
root = self._initialized_model()
|
|
||||||
save_path = root.save(file_prefix=checkpoint_prefix)
|
|
||||||
with context.eager_mode():
|
|
||||||
root = self._initialized_model()
|
|
||||||
self._set_sentinels(root)
|
|
||||||
root.restore(save_path).assert_consumed()
|
|
||||||
self._check_sentinels(root)
|
|
||||||
|
|
||||||
def testSaveEagerLoadGraph(self):
|
|
||||||
checkpoint_directory = self.get_temp_dir()
|
|
||||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
|
||||||
with context.eager_mode():
|
|
||||||
root = self._initialized_model()
|
|
||||||
save_path = root.save(file_prefix=checkpoint_prefix)
|
|
||||||
with context.graph_mode():
|
|
||||||
save_graph = ops.Graph()
|
|
||||||
with save_graph.as_default(), self.session(
|
|
||||||
graph=save_graph):
|
|
||||||
root = self._initialized_model()
|
|
||||||
self._set_sentinels(root)
|
|
||||||
root.restore(save_path).assert_consumed().run_restore_ops()
|
|
||||||
self._check_sentinels(root)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
test.main()
|
test.main()
|
||||||
|
Loading…
x
Reference in New Issue
Block a user