Fork tracking/util_with_v1_optimizers_test to keras/tests.
PiperOrigin-RevId: 316496459 Change-Id: I9563ba20c0d3cd72bc0651e49ab9a48e0b355d34
This commit is contained in:
parent
e3b8a8e720
commit
a2dc78458b
@ -435,6 +435,37 @@ tf_py_test(
|
||||
],
|
||||
)
|
||||
|
||||
tf_py_test(
|
||||
name = "tracking_util_with_v1_optimizers_test",
|
||||
srcs = ["tracking_util_with_v1_optimizers_test.py"],
|
||||
tags = [
|
||||
"notsan", # b/74395663
|
||||
],
|
||||
deps = [
|
||||
"//tensorflow/python:checkpoint_management",
|
||||
"//tensorflow/python:constant_op",
|
||||
"//tensorflow/python:dtypes",
|
||||
"//tensorflow/python:extra_py_tests_deps",
|
||||
"//tensorflow/python:framework_ops",
|
||||
"//tensorflow/python:framework_test_lib",
|
||||
"//tensorflow/python:resource_variable_ops",
|
||||
"//tensorflow/python:saver",
|
||||
"//tensorflow/python:state_ops",
|
||||
"//tensorflow/python:training_lib",
|
||||
"//tensorflow/python:training_util",
|
||||
"//tensorflow/python/distribute:mirrored_strategy",
|
||||
"//tensorflow/python/eager:backprop",
|
||||
"//tensorflow/python/eager:context",
|
||||
"//tensorflow/python/eager:def_function",
|
||||
"//tensorflow/python/eager:test",
|
||||
"//tensorflow/python/keras/engine",
|
||||
"//tensorflow/python/keras/layers:core",
|
||||
"//tensorflow/python/training/tracking",
|
||||
"//tensorflow/python/training/tracking:graph_view",
|
||||
"//tensorflow/python/training/tracking:util",
|
||||
],
|
||||
)
|
||||
|
||||
py_library(
|
||||
name = "get_config_samples",
|
||||
srcs = ["get_config_samples.py"],
|
||||
|
@ -0,0 +1,703 @@
|
||||
# Copyright 2017 The TensorFlow Authors. All Rights Reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
# ==============================================================================
|
||||
"""Tests for object-based saving which use tf.train.* optimizers."""
|
||||
from __future__ import absolute_import
|
||||
from __future__ import division
|
||||
from __future__ import print_function
|
||||
|
||||
import functools
|
||||
import os
|
||||
|
||||
import six
|
||||
|
||||
from tensorflow.python.distribute import mirrored_strategy
|
||||
from tensorflow.python.eager import backprop
|
||||
from tensorflow.python.eager import context
|
||||
from tensorflow.python.eager import def_function
|
||||
from tensorflow.python.eager import test
|
||||
from tensorflow.python.framework import constant_op
|
||||
from tensorflow.python.framework import dtypes
|
||||
from tensorflow.python.framework import ops
|
||||
from tensorflow.python.framework import test_util
|
||||
from tensorflow.python.keras.engine import training
|
||||
from tensorflow.python.keras.layers import core
|
||||
from tensorflow.python.ops import resource_variable_ops
|
||||
from tensorflow.python.ops import state_ops
|
||||
from tensorflow.python.training import adam
|
||||
from tensorflow.python.training import checkpoint_management
|
||||
from tensorflow.python.training import saver as saver_lib
|
||||
from tensorflow.python.training import training_util
|
||||
from tensorflow.python.training.tracking import graph_view
|
||||
from tensorflow.python.training.tracking import tracking
|
||||
from tensorflow.python.training.tracking import util as trackable_utils
|
||||
|
||||
|
||||
class NonLayerTrackable(tracking.AutoTrackable):
|
||||
|
||||
def __init__(self):
|
||||
super(NonLayerTrackable, self).__init__()
|
||||
self.a_variable = trackable_utils.add_variable(
|
||||
self, name="a_variable", shape=[])
|
||||
|
||||
|
||||
# pylint: disable=not-callable
|
||||
class MyModel(training.Model):
|
||||
"""A concrete Model for testing."""
|
||||
|
||||
def __init__(self):
|
||||
super(MyModel, self).__init__()
|
||||
self._named_dense = core.Dense(1, use_bias=True)
|
||||
self._second = core.Dense(1, use_bias=False)
|
||||
# We can still track Trackables which aren't Layers.
|
||||
self._non_layer = NonLayerTrackable()
|
||||
|
||||
def call(self, values):
|
||||
ret = self._second(self._named_dense(values))
|
||||
return ret
|
||||
|
||||
|
||||
class CheckpointingTests(test.TestCase):
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes(assert_no_eager_garbage=True)
|
||||
def testNamingWithOptimizer(self):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
model = MyModel()
|
||||
# A nuisance Model using the same optimizer. Its slot variables should not
|
||||
# go in the checkpoint, since it is never depended on.
|
||||
other_model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
optimizer_step = training_util.get_or_create_global_step()
|
||||
root_trackable = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
||||
if context.executing_eagerly():
|
||||
optimizer.minimize(
|
||||
lambda: model(input_value),
|
||||
global_step=optimizer_step)
|
||||
optimizer.minimize(
|
||||
lambda: other_model(input_value),
|
||||
global_step=optimizer_step)
|
||||
else:
|
||||
train_op = optimizer.minimize(
|
||||
model(input_value), global_step=optimizer_step)
|
||||
optimizer.minimize(
|
||||
other_model(input_value),
|
||||
global_step=optimizer_step)
|
||||
self.evaluate(trackable_utils.gather_initializers(
|
||||
root_trackable))
|
||||
self.evaluate(train_op)
|
||||
named_variables, serialized_graph, _ = graph_view.ObjectGraphView(
|
||||
root_trackable).serialize_object_graph()
|
||||
expected_checkpoint_names = (
|
||||
# Created in the root node, so no prefix.
|
||||
"optimizer_step",
|
||||
"model/_second/kernel",
|
||||
"model/_named_dense/kernel",
|
||||
"model/_named_dense/bias",
|
||||
# non-Layer dependency of the model
|
||||
"model/_non_layer/a_variable",
|
||||
# The optimizer creates two non-slot variables
|
||||
"optimizer/beta1_power",
|
||||
"optimizer/beta2_power",
|
||||
# Slot variables
|
||||
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
||||
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
||||
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/m",
|
||||
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/v",
|
||||
)
|
||||
suffix = "/.ATTRIBUTES/VARIABLE_VALUE"
|
||||
expected_checkpoint_names = [
|
||||
name + suffix for name in expected_checkpoint_names]
|
||||
named_variables = {v.name: v for v in named_variables}
|
||||
six.assertCountEqual(self, expected_checkpoint_names,
|
||||
named_variables.keys())
|
||||
# Check that we've mapped to the right variable objects (not exhaustive)
|
||||
self.assertEqual(
|
||||
"global_step",
|
||||
named_variables["optimizer_step" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense_1/kernel",
|
||||
named_variables["model/_second/kernel" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel",
|
||||
named_variables["model/_named_dense/kernel" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"beta1_power",
|
||||
named_variables["optimizer/beta1_power" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"beta2_power",
|
||||
named_variables["optimizer/beta2_power" + suffix].full_name)
|
||||
# Spot check the generated protocol buffers.
|
||||
self.assertEqual("optimizer",
|
||||
serialized_graph.nodes[0].children[1].local_name)
|
||||
optimizer_node = serialized_graph.nodes[serialized_graph.nodes[0].children[
|
||||
1].node_id]
|
||||
self.assertEqual("beta1_power",
|
||||
optimizer_node.children[0].local_name)
|
||||
self.assertEqual("beta1_power",
|
||||
serialized_graph.nodes[optimizer_node.children[0].node_id]
|
||||
.attributes[0].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel",
|
||||
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
||||
.original_variable_node_id]
|
||||
.attributes[0].full_name)
|
||||
# We strip off the :0 suffix, as variable.name-based saving does.
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel/Adam",
|
||||
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
||||
.slot_variable_node_id]
|
||||
.attributes[0].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel/Adam:0",
|
||||
optimizer.get_slot(
|
||||
var=model._named_dense.kernel,
|
||||
name="m").name)
|
||||
self.assertEqual(
|
||||
"model/_named_dense/kernel" + suffix,
|
||||
serialized_graph.nodes[
|
||||
optimizer_node.slot_variables[0]
|
||||
.original_variable_node_id].attributes[0].checkpoint_key)
|
||||
self.assertEqual("m", optimizer_node.slot_variables[0].slot_name)
|
||||
self.assertEqual(
|
||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m" + suffix,
|
||||
serialized_graph.nodes[
|
||||
optimizer_node.slot_variables[0]
|
||||
.slot_variable_node_id].attributes[0].checkpoint_key)
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testSaveRestore(self):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root_trackable = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
if context.executing_eagerly():
|
||||
optimizer.minimize(
|
||||
lambda: model(input_value))
|
||||
else:
|
||||
train_op = optimizer.minimize(model(input_value))
|
||||
# TODO(allenl): Make initialization more pleasant when graph building.
|
||||
root_trackable.save_counter # pylint: disable=pointless-statement
|
||||
self.evaluate(trackable_utils.gather_initializers(
|
||||
root_trackable))
|
||||
self.evaluate(train_op)
|
||||
prefix = os.path.join(self.get_temp_dir(), "ckpt")
|
||||
self.evaluate(state_ops.assign(model._named_dense.variables[1], [42.]))
|
||||
m_bias_slot = optimizer.get_slot(model._named_dense.variables[1], "m")
|
||||
self.evaluate(state_ops.assign(m_bias_slot, [1.5]))
|
||||
save_path = root_trackable.save(file_prefix=prefix)
|
||||
self.evaluate(state_ops.assign(model._named_dense.variables[1], [43.]))
|
||||
self.evaluate(state_ops.assign(root_trackable.save_counter, 3))
|
||||
optimizer_variables = self.evaluate(optimizer.variables())
|
||||
self.evaluate(state_ops.assign(m_bias_slot, [-2.]))
|
||||
# Immediate restoration
|
||||
status = root_trackable.restore(save_path=save_path).assert_consumed()
|
||||
status.run_restore_ops()
|
||||
self.assertAllEqual([42.], self.evaluate(model._named_dense.variables[1]))
|
||||
self.assertAllEqual(1, self.evaluate(root_trackable.save_counter))
|
||||
self.assertAllEqual([1.5], self.evaluate(m_bias_slot))
|
||||
if not context.executing_eagerly():
|
||||
return # Restore-on-create is only supported when executing eagerly
|
||||
on_create_model = MyModel()
|
||||
on_create_optimizer = adam.AdamOptimizer(
|
||||
0.001,
|
||||
# Preserve beta1_power and beta2_power when applying gradients so we can
|
||||
# test that they've been restored correctly.
|
||||
beta1=1.0,
|
||||
beta2=1.0)
|
||||
on_create_root = trackable_utils.Checkpoint(
|
||||
optimizer=on_create_optimizer, model=on_create_model)
|
||||
# Deferred restoration
|
||||
status = on_create_root.restore(save_path=save_path)
|
||||
status.assert_nontrivial_match()
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
on_create_model(constant_op.constant([[3.]])) # create variables
|
||||
self.assertAllEqual(1, self.evaluate(on_create_root.save_counter))
|
||||
self.assertAllEqual([42.],
|
||||
self.evaluate(
|
||||
on_create_model._named_dense.variables[1]))
|
||||
on_create_m_bias_slot = on_create_optimizer.get_slot(
|
||||
on_create_model._named_dense.variables[1], "m")
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
# Optimizer slot variables are created when the original variable is
|
||||
# restored.
|
||||
self.assertAllEqual([1.5], self.evaluate(on_create_m_bias_slot))
|
||||
self.assertAllEqual(optimizer_variables[2:],
|
||||
self.evaluate(on_create_optimizer.variables()))
|
||||
dummy_var = resource_variable_ops.ResourceVariable([1.])
|
||||
on_create_optimizer.minimize(loss=dummy_var.read_value)
|
||||
status.assert_existing_objects_matched()
|
||||
status.assert_consumed()
|
||||
beta1_power, beta2_power = on_create_optimizer._get_beta_accumulators()
|
||||
self.assertAllEqual(optimizer_variables[0], self.evaluate(beta1_power))
|
||||
self.assertAllEqual(optimizer_variables[1], self.evaluate(beta2_power))
|
||||
|
||||
# TODO(allenl): Debug garbage created by this test in python3.
|
||||
def testDeferredRestorationUsageEager(self):
|
||||
"""An idiomatic eager execution example."""
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
for training_continuation in range(3):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
optimizer_step=training_util.get_or_create_global_step())
|
||||
root.restore(checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory))
|
||||
for _ in range(num_training_steps):
|
||||
# TODO(allenl): Use a Dataset and serialize/checkpoint it.
|
||||
input_value = constant_op.constant([[3.]])
|
||||
optimizer.minimize(
|
||||
lambda: model(input_value), # pylint: disable=cell-var-from-loop
|
||||
global_step=root.optimizer_step)
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
root.optimizer_step.numpy())
|
||||
|
||||
def testEagerDistributionStrategy(self):
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
|
||||
def _train_fn(optimizer, model):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
optimizer.minimize(
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.optimizer_step)
|
||||
|
||||
strategy = mirrored_strategy.MirroredStrategy()
|
||||
with strategy.scope():
|
||||
for training_continuation in range(3):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer,
|
||||
model=model,
|
||||
optimizer_step=training_util.get_or_create_global_step())
|
||||
root.restore(
|
||||
checkpoint_management.latest_checkpoint(checkpoint_directory))
|
||||
|
||||
for _ in range(num_training_steps):
|
||||
strategy.extended.call_for_each_replica(
|
||||
functools.partial(_train_fn, optimizer, model))
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
root.optimizer_step.numpy())
|
||||
|
||||
def testGraphDistributionStrategy(self):
|
||||
self.skipTest("b/121381184")
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
|
||||
def _train_fn(optimizer, model):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
return optimizer.minimize(
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.optimizer_step)
|
||||
|
||||
for training_continuation in range(3):
|
||||
with ops.Graph().as_default():
|
||||
strategy = mirrored_strategy.MirroredStrategy()
|
||||
with strategy.scope():
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
optimizer_step=training_util.get_or_create_global_step())
|
||||
status = root.restore(checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory))
|
||||
train_op = strategy.extended.call_for_each_replica(
|
||||
functools.partial(_train_fn, optimizer, model))
|
||||
with self.session() as session:
|
||||
if training_continuation > 0:
|
||||
status.assert_consumed()
|
||||
status.initialize_or_restore()
|
||||
for _ in range(num_training_steps):
|
||||
session.run(train_op)
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
root.optimizer_step.numpy())
|
||||
|
||||
def testUsageGraph(self):
|
||||
"""Expected usage when graph building."""
|
||||
with context.graph_mode():
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
for training_continuation in range(3):
|
||||
with ops.Graph().as_default():
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.CheckpointV1(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_op = optimizer.minimize(
|
||||
model(input_value),
|
||||
global_step=root.global_step)
|
||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory)
|
||||
with self.session(graph=ops.get_default_graph()) as session:
|
||||
status = root.restore(save_path=checkpoint_path)
|
||||
status.initialize_or_restore(session=session)
|
||||
if checkpoint_path is None:
|
||||
self.assertEqual(0, training_continuation)
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_existing_objects_matched()
|
||||
else:
|
||||
status.assert_consumed()
|
||||
status.assert_existing_objects_matched()
|
||||
for _ in range(num_training_steps):
|
||||
session.run(train_op)
|
||||
root.save(file_prefix=checkpoint_prefix, session=session)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
session.run(root.global_step))
|
||||
self.assertEqual(training_continuation + 1,
|
||||
session.run(root.save_counter))
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testAgnosticUsage(self):
|
||||
"""Graph/eager agnostic usage."""
|
||||
# Does create garbage when executing eagerly due to ops.Graph() creation.
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
for training_continuation in range(3):
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
manager = checkpoint_management.CheckpointManager(
|
||||
root, checkpoint_directory, max_to_keep=1)
|
||||
status = root.restore(save_path=manager.latest_checkpoint)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
for _ in range(num_training_steps):
|
||||
train_fn()
|
||||
manager.save()
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
self.evaluate(root.global_step))
|
||||
self.assertEqual(training_continuation + 1,
|
||||
self.evaluate(root.save_counter))
|
||||
|
||||
# pylint: disable=cell-var-from-loop
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testWithDefun(self):
|
||||
num_training_steps = 2
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
for training_continuation in range(3):
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
# Don't actually train so we can test variable values
|
||||
optimizer = adam.AdamOptimizer(0.)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory)
|
||||
status = root.restore(save_path=checkpoint_path)
|
||||
def train_fn():
|
||||
@def_function.function
|
||||
def _call_model(x):
|
||||
return model(x)
|
||||
with backprop.GradientTape() as tape:
|
||||
loss = _call_model(constant_op.constant([[3.]]))
|
||||
gradients = tape.gradient(loss, model.variables)
|
||||
return optimizer.apply_gradients(zip(gradients, model.variables),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(
|
||||
self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
for _ in range(num_training_steps):
|
||||
train_fn()
|
||||
if training_continuation > 0:
|
||||
status.assert_consumed()
|
||||
self.assertAllClose([[42.]], self.evaluate(model.variables[0]))
|
||||
else:
|
||||
self.evaluate(model.variables[0].assign([[42.]]))
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
self.evaluate(root.global_step))
|
||||
self.assertEqual(training_continuation + 1,
|
||||
self.evaluate(root.save_counter))
|
||||
# pylint: enable=cell-var-from-loop
|
||||
|
||||
def _get_checkpoint_name(self, name):
|
||||
root = tracking.AutoTrackable()
|
||||
trackable_utils.add_variable(
|
||||
root, name=name, shape=[1, 2], dtype=dtypes.float64)
|
||||
(named_variable,), _, _ = trackable_utils._serialize_object_graph(
|
||||
root, saveables_cache=None)
|
||||
with ops.name_scope("root/" + named_variable.name):
|
||||
pass # Make sure we can use this as an op name if we prefix it.
|
||||
return named_variable.name
|
||||
|
||||
def testAnonymousVarsInInit(self):
|
||||
|
||||
class Model(training.Model):
|
||||
|
||||
def __init__(self):
|
||||
super(Model, self).__init__()
|
||||
self.w = resource_variable_ops.ResourceVariable(0.0)
|
||||
self.b = resource_variable_ops.ResourceVariable(0.0)
|
||||
self.vars = [self.w, self.b]
|
||||
|
||||
def call(self, x):
|
||||
return x * self.w + self.b
|
||||
|
||||
with context.eager_mode():
|
||||
model = Model()
|
||||
optimizer = adam.AdamOptimizer(learning_rate=0.05)
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
checkpoint = trackable_utils.Checkpoint(
|
||||
model=model, optimizer=optimizer)
|
||||
for _ in range(2):
|
||||
checkpoint.save(checkpoint_prefix)
|
||||
with backprop.GradientTape() as tape:
|
||||
loss = (constant_op.constant(1.)
|
||||
- model(constant_op.constant(1.))) ** 2
|
||||
grad = tape.gradient(loss, model.vars)
|
||||
optimizer.apply_gradients(
|
||||
[(g, v) for g, v in zip(grad, model.vars)])
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def test_initialize_if_not_restoring(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
optimizer_only_prefix = os.path.join(checkpoint_directory, "opt")
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
model=model, # Do not save the optimizer with the checkpoint.
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
optimizer_checkpoint = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer)
|
||||
|
||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory)
|
||||
status = root.restore(save_path=checkpoint_path)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
self.evaluate([v.initializer for v in optimizer.variables()])
|
||||
train_fn()
|
||||
model_save_path = root.save(file_prefix=checkpoint_prefix)
|
||||
self.evaluate(optimizer.variables()[0].assign(42.))
|
||||
optimizer_save_path = optimizer_checkpoint.save(optimizer_only_prefix)
|
||||
|
||||
# Restore into a graph with the optimizer
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
status = root.restore(save_path=model_save_path)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
train_fn()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
|
||||
# Make sure initialization doesn't clobber later restores
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001, beta1=1.0)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
opt_root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer)
|
||||
status = root.restore(save_path=model_save_path)
|
||||
init_only_optimizer_status = opt_root.restore(save_path=None)
|
||||
optimizer_status = opt_root.restore(save_path=optimizer_save_path)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
optimizer_status.run_restore_ops()
|
||||
status.initialize_or_restore()
|
||||
init_only_optimizer_status.initialize_or_restore()
|
||||
train_fn()
|
||||
self.assertEqual(42., self.evaluate(optimizer.variables()[0]))
|
||||
|
||||
|
||||
class CheckpointCompatibilityTests(test.TestCase):
|
||||
|
||||
def _initialized_model(self):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
optimizer_step = training_util.get_or_create_global_step()
|
||||
root_trackable = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
||||
train_op = optimizer.minimize(
|
||||
functools.partial(model, input_value),
|
||||
global_step=optimizer_step)
|
||||
self.evaluate(trackable_utils.gather_initializers(
|
||||
root_trackable))
|
||||
self.evaluate(train_op)
|
||||
# A regular variable, a slot variable, and a non-slot Optimizer variable
|
||||
# with known values to check when loading.
|
||||
self.evaluate(model._named_dense.bias.assign([1.]))
|
||||
self.evaluate(optimizer.get_slot(
|
||||
var=model._named_dense.bias, name="m").assign([2.]))
|
||||
beta1_power, _ = optimizer._get_beta_accumulators()
|
||||
self.evaluate(beta1_power.assign(3.))
|
||||
return root_trackable
|
||||
|
||||
def _set_sentinels(self, root_trackable):
|
||||
self.evaluate(root_trackable.model._named_dense.bias.assign([101.]))
|
||||
self.evaluate(
|
||||
root_trackable.optimizer.get_slot(
|
||||
var=root_trackable.model._named_dense.bias, name="m")
|
||||
.assign([102.]))
|
||||
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
||||
self.evaluate(beta1_power.assign(103.))
|
||||
|
||||
def _check_sentinels(self, root_trackable):
|
||||
self.assertAllEqual(
|
||||
[1.], self.evaluate(root_trackable.model._named_dense.bias))
|
||||
self.assertAllEqual([2.], self.evaluate(
|
||||
root_trackable.optimizer.get_slot(
|
||||
var=root_trackable.model._named_dense.bias, name="m")))
|
||||
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
||||
self.assertAllEqual(3., self.evaluate(beta1_power))
|
||||
|
||||
def _write_name_based_checkpoint(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
with context.graph_mode():
|
||||
save_graph = ops.Graph()
|
||||
with save_graph.as_default(), self.session(
|
||||
graph=save_graph) as session:
|
||||
root = self._initialized_model()
|
||||
name_saver = saver_lib.Saver()
|
||||
return name_saver.save(
|
||||
sess=session, save_path=checkpoint_prefix,
|
||||
global_step=root.optimizer_step)
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testLoadFromNameBasedSaver(self):
|
||||
"""Save a name-based checkpoint, load it using the object-based API."""
|
||||
with test_util.device(use_gpu=True):
|
||||
save_path = self._write_name_based_checkpoint()
|
||||
root = self._initialized_model()
|
||||
self._set_sentinels(root)
|
||||
with self.assertRaises(AssertionError):
|
||||
self._check_sentinels(root)
|
||||
object_saver = trackable_utils.TrackableSaver(
|
||||
graph_view.ObjectGraphView(root))
|
||||
self._set_sentinels(root)
|
||||
status = object_saver.restore(save_path)
|
||||
if context.executing_eagerly():
|
||||
self._check_sentinels(root)
|
||||
if context.executing_eagerly():
|
||||
status.assert_consumed()
|
||||
status.assert_existing_objects_matched()
|
||||
status.assert_nontrivial_match()
|
||||
else:
|
||||
# When graph building, we haven't read any keys, so we don't know
|
||||
# whether the restore will be complete.
|
||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||
status.assert_consumed()
|
||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||
status.assert_nontrivial_match()
|
||||
status.run_restore_ops()
|
||||
self._check_sentinels(root)
|
||||
self._set_sentinels(root)
|
||||
status = object_saver.restore(save_path)
|
||||
status.initialize_or_restore()
|
||||
self._check_sentinels(root)
|
||||
# Check that there is no error when keys are missing from the name-based
|
||||
# checkpoint.
|
||||
root.not_in_name_checkpoint = resource_variable_ops.ResourceVariable([1.])
|
||||
status = object_saver.restore(save_path)
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_existing_objects_matched()
|
||||
|
||||
def testSaveGraphLoadEager(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
with context.graph_mode():
|
||||
save_graph = ops.Graph()
|
||||
with save_graph.as_default(), self.session(
|
||||
graph=save_graph):
|
||||
root = self._initialized_model()
|
||||
save_path = root.save(file_prefix=checkpoint_prefix)
|
||||
with context.eager_mode():
|
||||
root = self._initialized_model()
|
||||
self._set_sentinels(root)
|
||||
root.restore(save_path).assert_consumed()
|
||||
self._check_sentinels(root)
|
||||
|
||||
def testSaveEagerLoadGraph(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
with context.eager_mode():
|
||||
root = self._initialized_model()
|
||||
save_path = root.save(file_prefix=checkpoint_prefix)
|
||||
with context.graph_mode():
|
||||
save_graph = ops.Graph()
|
||||
with save_graph.as_default(), self.session(
|
||||
graph=save_graph):
|
||||
root = self._initialized_model()
|
||||
self._set_sentinels(root)
|
||||
root.restore(save_path).assert_consumed().run_restore_ops()
|
||||
self._check_sentinels(root)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
test.main()
|
@ -223,34 +223,19 @@ tf_py_test(
|
||||
"notsan", # b/74395663
|
||||
],
|
||||
deps = [
|
||||
":base",
|
||||
":graph_view",
|
||||
":tracking",
|
||||
":util",
|
||||
"//tensorflow/python:checkpoint_management",
|
||||
"//tensorflow/python:constant_op",
|
||||
"//tensorflow/python:control_flow_ops",
|
||||
"//tensorflow/python:dtypes",
|
||||
"//tensorflow/python:framework_ops",
|
||||
"//tensorflow/python:framework_test_lib",
|
||||
"//tensorflow/python:init_ops",
|
||||
"//tensorflow/python:pywrap_tensorflow",
|
||||
"//tensorflow/python:resource_variable_ops",
|
||||
"//tensorflow/python:saver",
|
||||
"//tensorflow/python:session",
|
||||
"//tensorflow/python:state_ops",
|
||||
"//tensorflow/python:template",
|
||||
"//tensorflow/python:training",
|
||||
"//tensorflow/python:training_util",
|
||||
"//tensorflow/python:variable_scope",
|
||||
"//tensorflow/python/distribute:mirrored_strategy",
|
||||
"//tensorflow/python/eager:backprop",
|
||||
"//tensorflow/python/eager:context",
|
||||
"//tensorflow/python/eager:def_function",
|
||||
"//tensorflow/python/eager:test",
|
||||
"//tensorflow/python/keras:engine",
|
||||
"//tensorflow/python/keras/layers",
|
||||
"@absl_py//absl/testing:parameterized",
|
||||
"@six_archive//:six",
|
||||
],
|
||||
)
|
||||
|
@ -17,486 +17,27 @@ from __future__ import absolute_import
|
||||
from __future__ import division
|
||||
from __future__ import print_function
|
||||
|
||||
import functools
|
||||
import os
|
||||
|
||||
import six
|
||||
|
||||
from tensorflow.python.client import session as session_lib
|
||||
from tensorflow.python.distribute import mirrored_strategy
|
||||
from tensorflow.python.eager import backprop
|
||||
from tensorflow.python.eager import context
|
||||
from tensorflow.python.eager import def_function
|
||||
from tensorflow.python.eager import test
|
||||
from tensorflow.python.framework import constant_op
|
||||
from tensorflow.python.framework import dtypes
|
||||
from tensorflow.python.framework import ops
|
||||
from tensorflow.python.framework import test_util
|
||||
from tensorflow.python.keras.engine import training
|
||||
from tensorflow.python.keras.layers import core
|
||||
from tensorflow.python.ops import init_ops
|
||||
from tensorflow.python.ops import resource_variable_ops
|
||||
from tensorflow.python.ops import state_ops
|
||||
from tensorflow.python.ops import template
|
||||
from tensorflow.python.ops import variable_scope
|
||||
from tensorflow.python.training import adam
|
||||
from tensorflow.python.training import checkpoint_management
|
||||
from tensorflow.python.training import saver as saver_lib
|
||||
from tensorflow.python.training import training_util
|
||||
from tensorflow.python.training.tracking import graph_view
|
||||
from tensorflow.python.training.tracking import tracking
|
||||
from tensorflow.python.training.tracking import util as trackable_utils
|
||||
|
||||
|
||||
class NonLayerTrackable(tracking.AutoTrackable):
|
||||
|
||||
def __init__(self):
|
||||
super(NonLayerTrackable, self).__init__()
|
||||
self.a_variable = trackable_utils.add_variable(
|
||||
self, name="a_variable", shape=[])
|
||||
|
||||
|
||||
# pylint: disable=not-callable
|
||||
class MyModel(training.Model):
|
||||
"""A concrete Model for testing."""
|
||||
|
||||
def __init__(self):
|
||||
super(MyModel, self).__init__()
|
||||
self._named_dense = core.Dense(1, use_bias=True)
|
||||
self._second = core.Dense(1, use_bias=False)
|
||||
# We can still track Trackables which aren't Layers.
|
||||
self._non_layer = NonLayerTrackable()
|
||||
|
||||
def call(self, values):
|
||||
ret = self._second(self._named_dense(values))
|
||||
return ret
|
||||
|
||||
|
||||
class CheckpointingTests(test.TestCase):
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes(assert_no_eager_garbage=True)
|
||||
def testNamingWithOptimizer(self):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
model = MyModel()
|
||||
# A nuisance Model using the same optimizer. Its slot variables should not
|
||||
# go in the checkpoint, since it is never depended on.
|
||||
other_model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
optimizer_step = training_util.get_or_create_global_step()
|
||||
root_trackable = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
||||
if context.executing_eagerly():
|
||||
optimizer.minimize(
|
||||
lambda: model(input_value),
|
||||
global_step=optimizer_step)
|
||||
optimizer.minimize(
|
||||
lambda: other_model(input_value),
|
||||
global_step=optimizer_step)
|
||||
else:
|
||||
train_op = optimizer.minimize(
|
||||
model(input_value), global_step=optimizer_step)
|
||||
optimizer.minimize(
|
||||
other_model(input_value),
|
||||
global_step=optimizer_step)
|
||||
self.evaluate(trackable_utils.gather_initializers(
|
||||
root_trackable))
|
||||
self.evaluate(train_op)
|
||||
named_variables, serialized_graph, _ = graph_view.ObjectGraphView(
|
||||
root_trackable).serialize_object_graph()
|
||||
expected_checkpoint_names = (
|
||||
# Created in the root node, so no prefix.
|
||||
"optimizer_step",
|
||||
"model/_second/kernel",
|
||||
"model/_named_dense/kernel",
|
||||
"model/_named_dense/bias",
|
||||
# non-Layer dependency of the model
|
||||
"model/_non_layer/a_variable",
|
||||
# The optimizer creates two non-slot variables
|
||||
"optimizer/beta1_power",
|
||||
"optimizer/beta2_power",
|
||||
# Slot variables
|
||||
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
||||
"model/_second/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m",
|
||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/v",
|
||||
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/m",
|
||||
"model/_named_dense/bias/.OPTIMIZER_SLOT/optimizer/v",
|
||||
)
|
||||
suffix = "/.ATTRIBUTES/VARIABLE_VALUE"
|
||||
expected_checkpoint_names = [
|
||||
name + suffix for name in expected_checkpoint_names]
|
||||
named_variables = {v.name: v for v in named_variables}
|
||||
six.assertCountEqual(self, expected_checkpoint_names,
|
||||
named_variables.keys())
|
||||
# Check that we've mapped to the right variable objects (not exhaustive)
|
||||
self.assertEqual(
|
||||
"global_step",
|
||||
named_variables["optimizer_step" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense_1/kernel",
|
||||
named_variables["model/_second/kernel" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel",
|
||||
named_variables["model/_named_dense/kernel" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"beta1_power",
|
||||
named_variables["optimizer/beta1_power" + suffix].full_name)
|
||||
self.assertEqual(
|
||||
"beta2_power",
|
||||
named_variables["optimizer/beta2_power" + suffix].full_name)
|
||||
# Spot check the generated protocol buffers.
|
||||
self.assertEqual("optimizer",
|
||||
serialized_graph.nodes[0].children[1].local_name)
|
||||
optimizer_node = serialized_graph.nodes[serialized_graph.nodes[0].children[
|
||||
1].node_id]
|
||||
self.assertEqual("beta1_power",
|
||||
optimizer_node.children[0].local_name)
|
||||
self.assertEqual("beta1_power",
|
||||
serialized_graph.nodes[optimizer_node.children[0].node_id]
|
||||
.attributes[0].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel",
|
||||
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
||||
.original_variable_node_id]
|
||||
.attributes[0].full_name)
|
||||
# We strip off the :0 suffix, as variable.name-based saving does.
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel/Adam",
|
||||
serialized_graph.nodes[optimizer_node.slot_variables[0]
|
||||
.slot_variable_node_id]
|
||||
.attributes[0].full_name)
|
||||
self.assertEqual(
|
||||
"my_model/dense/kernel/Adam:0",
|
||||
optimizer.get_slot(
|
||||
var=model._named_dense.kernel,
|
||||
name="m").name)
|
||||
self.assertEqual(
|
||||
"model/_named_dense/kernel" + suffix,
|
||||
serialized_graph.nodes[
|
||||
optimizer_node.slot_variables[0]
|
||||
.original_variable_node_id].attributes[0].checkpoint_key)
|
||||
self.assertEqual("m", optimizer_node.slot_variables[0].slot_name)
|
||||
self.assertEqual(
|
||||
"model/_named_dense/kernel/.OPTIMIZER_SLOT/optimizer/m" + suffix,
|
||||
serialized_graph.nodes[
|
||||
optimizer_node.slot_variables[0]
|
||||
.slot_variable_node_id].attributes[0].checkpoint_key)
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testSaveRestore(self):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root_trackable = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
if context.executing_eagerly():
|
||||
optimizer.minimize(
|
||||
lambda: model(input_value))
|
||||
else:
|
||||
train_op = optimizer.minimize(model(input_value))
|
||||
# TODO(allenl): Make initialization more pleasant when graph building.
|
||||
root_trackable.save_counter # pylint: disable=pointless-statement
|
||||
self.evaluate(trackable_utils.gather_initializers(
|
||||
root_trackable))
|
||||
self.evaluate(train_op)
|
||||
prefix = os.path.join(self.get_temp_dir(), "ckpt")
|
||||
self.evaluate(state_ops.assign(model._named_dense.variables[1], [42.]))
|
||||
m_bias_slot = optimizer.get_slot(model._named_dense.variables[1], "m")
|
||||
self.evaluate(state_ops.assign(m_bias_slot, [1.5]))
|
||||
save_path = root_trackable.save(file_prefix=prefix)
|
||||
self.evaluate(state_ops.assign(model._named_dense.variables[1], [43.]))
|
||||
self.evaluate(state_ops.assign(root_trackable.save_counter, 3))
|
||||
optimizer_variables = self.evaluate(optimizer.variables())
|
||||
self.evaluate(state_ops.assign(m_bias_slot, [-2.]))
|
||||
# Immediate restoration
|
||||
status = root_trackable.restore(save_path=save_path).assert_consumed()
|
||||
status.run_restore_ops()
|
||||
self.assertAllEqual([42.], self.evaluate(model._named_dense.variables[1]))
|
||||
self.assertAllEqual(1, self.evaluate(root_trackable.save_counter))
|
||||
self.assertAllEqual([1.5], self.evaluate(m_bias_slot))
|
||||
if not context.executing_eagerly():
|
||||
return # Restore-on-create is only supported when executing eagerly
|
||||
on_create_model = MyModel()
|
||||
on_create_optimizer = adam.AdamOptimizer(
|
||||
0.001,
|
||||
# Preserve beta1_power and beta2_power when applying gradients so we can
|
||||
# test that they've been restored correctly.
|
||||
beta1=1.0,
|
||||
beta2=1.0)
|
||||
on_create_root = trackable_utils.Checkpoint(
|
||||
optimizer=on_create_optimizer, model=on_create_model)
|
||||
# Deferred restoration
|
||||
status = on_create_root.restore(save_path=save_path)
|
||||
status.assert_nontrivial_match()
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
on_create_model(constant_op.constant([[3.]])) # create variables
|
||||
self.assertAllEqual(1, self.evaluate(on_create_root.save_counter))
|
||||
self.assertAllEqual([42.],
|
||||
self.evaluate(
|
||||
on_create_model._named_dense.variables[1]))
|
||||
on_create_m_bias_slot = on_create_optimizer.get_slot(
|
||||
on_create_model._named_dense.variables[1], "m")
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
# Optimizer slot variables are created when the original variable is
|
||||
# restored.
|
||||
self.assertAllEqual([1.5], self.evaluate(on_create_m_bias_slot))
|
||||
self.assertAllEqual(optimizer_variables[2:],
|
||||
self.evaluate(on_create_optimizer.variables()))
|
||||
dummy_var = resource_variable_ops.ResourceVariable([1.])
|
||||
on_create_optimizer.minimize(loss=dummy_var.read_value)
|
||||
status.assert_existing_objects_matched()
|
||||
status.assert_consumed()
|
||||
beta1_power, beta2_power = on_create_optimizer._get_beta_accumulators()
|
||||
self.assertAllEqual(optimizer_variables[0], self.evaluate(beta1_power))
|
||||
self.assertAllEqual(optimizer_variables[1], self.evaluate(beta2_power))
|
||||
|
||||
# TODO(allenl): Debug garbage created by this test in python3.
|
||||
def testDeferredRestorationUsageEager(self):
|
||||
"""An idiomatic eager execution example."""
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
for training_continuation in range(3):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
optimizer_step=training_util.get_or_create_global_step())
|
||||
root.restore(checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory))
|
||||
for _ in range(num_training_steps):
|
||||
# TODO(allenl): Use a Dataset and serialize/checkpoint it.
|
||||
input_value = constant_op.constant([[3.]])
|
||||
optimizer.minimize(
|
||||
lambda: model(input_value), # pylint: disable=cell-var-from-loop
|
||||
global_step=root.optimizer_step)
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
root.optimizer_step.numpy())
|
||||
|
||||
def testEagerDistributionStrategy(self):
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
|
||||
def _train_fn(optimizer, model):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
optimizer.minimize(
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.optimizer_step)
|
||||
|
||||
strategy = mirrored_strategy.MirroredStrategy()
|
||||
with strategy.scope():
|
||||
for training_continuation in range(3):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer,
|
||||
model=model,
|
||||
optimizer_step=training_util.get_or_create_global_step())
|
||||
root.restore(
|
||||
checkpoint_management.latest_checkpoint(checkpoint_directory))
|
||||
|
||||
for _ in range(num_training_steps):
|
||||
strategy.extended.call_for_each_replica(
|
||||
functools.partial(_train_fn, optimizer, model))
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
root.optimizer_step.numpy())
|
||||
|
||||
def testGraphDistributionStrategy(self):
|
||||
self.skipTest("b/121381184")
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
|
||||
def _train_fn(optimizer, model):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
return optimizer.minimize(
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.optimizer_step)
|
||||
|
||||
for training_continuation in range(3):
|
||||
with ops.Graph().as_default():
|
||||
strategy = mirrored_strategy.MirroredStrategy()
|
||||
with strategy.scope():
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
optimizer_step=training_util.get_or_create_global_step())
|
||||
status = root.restore(checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory))
|
||||
train_op = strategy.extended.call_for_each_replica(
|
||||
functools.partial(_train_fn, optimizer, model))
|
||||
with self.session() as session:
|
||||
if training_continuation > 0:
|
||||
status.assert_consumed()
|
||||
status.initialize_or_restore()
|
||||
for _ in range(num_training_steps):
|
||||
session.run(train_op)
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
root.optimizer_step.numpy())
|
||||
|
||||
def testUsageGraph(self):
|
||||
"""Expected usage when graph building."""
|
||||
with context.graph_mode():
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
for training_continuation in range(3):
|
||||
with ops.Graph().as_default():
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.CheckpointV1(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_op = optimizer.minimize(
|
||||
model(input_value),
|
||||
global_step=root.global_step)
|
||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory)
|
||||
with self.session(graph=ops.get_default_graph()) as session:
|
||||
status = root.restore(save_path=checkpoint_path)
|
||||
status.initialize_or_restore(session=session)
|
||||
if checkpoint_path is None:
|
||||
self.assertEqual(0, training_continuation)
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_existing_objects_matched()
|
||||
else:
|
||||
status.assert_consumed()
|
||||
status.assert_existing_objects_matched()
|
||||
for _ in range(num_training_steps):
|
||||
session.run(train_op)
|
||||
root.save(file_prefix=checkpoint_prefix, session=session)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
session.run(root.global_step))
|
||||
self.assertEqual(training_continuation + 1,
|
||||
session.run(root.save_counter))
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testAgnosticUsage(self):
|
||||
"""Graph/eager agnostic usage."""
|
||||
# Does create garbage when executing eagerly due to ops.Graph() creation.
|
||||
num_training_steps = 10
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
for training_continuation in range(3):
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
manager = checkpoint_management.CheckpointManager(
|
||||
root, checkpoint_directory, max_to_keep=1)
|
||||
status = root.restore(save_path=manager.latest_checkpoint)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
for _ in range(num_training_steps):
|
||||
train_fn()
|
||||
manager.save()
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
self.evaluate(root.global_step))
|
||||
self.assertEqual(training_continuation + 1,
|
||||
self.evaluate(root.save_counter))
|
||||
|
||||
# pylint: disable=cell-var-from-loop
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testWithDefun(self):
|
||||
num_training_steps = 2
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
for training_continuation in range(3):
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
# Don't actually train so we can test variable values
|
||||
optimizer = adam.AdamOptimizer(0.)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory)
|
||||
status = root.restore(save_path=checkpoint_path)
|
||||
def train_fn():
|
||||
@def_function.function
|
||||
def _call_model(x):
|
||||
return model(x)
|
||||
with backprop.GradientTape() as tape:
|
||||
loss = _call_model(constant_op.constant([[3.]]))
|
||||
gradients = tape.gradient(loss, model.variables)
|
||||
return optimizer.apply_gradients(zip(gradients, model.variables),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(
|
||||
self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
for _ in range(num_training_steps):
|
||||
train_fn()
|
||||
if training_continuation > 0:
|
||||
status.assert_consumed()
|
||||
self.assertAllClose([[42.]], self.evaluate(model.variables[0]))
|
||||
else:
|
||||
self.evaluate(model.variables[0].assign([[42.]]))
|
||||
root.save(file_prefix=checkpoint_prefix)
|
||||
self.assertEqual((training_continuation + 1) * num_training_steps,
|
||||
self.evaluate(root.global_step))
|
||||
self.assertEqual(training_continuation + 1,
|
||||
self.evaluate(root.save_counter))
|
||||
# pylint: enable=cell-var-from-loop
|
||||
|
||||
def _get_checkpoint_name(self, name):
|
||||
root = tracking.AutoTrackable()
|
||||
trackable_utils.add_variable(
|
||||
root, name=name, shape=[1, 2], dtype=dtypes.float64)
|
||||
(named_variable,), _, _ = trackable_utils._serialize_object_graph(
|
||||
root, saveables_cache=None)
|
||||
with ops.name_scope("root/" + named_variable.name):
|
||||
pass # Make sure we can use this as an op name if we prefix it.
|
||||
return named_variable.name
|
||||
|
||||
def testAnonymousVarsInInit(self):
|
||||
|
||||
class Model(training.Model):
|
||||
|
||||
def __init__(self):
|
||||
super(Model, self).__init__()
|
||||
self.w = resource_variable_ops.ResourceVariable(0.0)
|
||||
self.b = resource_variable_ops.ResourceVariable(0.0)
|
||||
self.vars = [self.w, self.b]
|
||||
|
||||
def call(self, x):
|
||||
return x * self.w + self.b
|
||||
|
||||
with context.eager_mode():
|
||||
model = Model()
|
||||
optimizer = adam.AdamOptimizer(learning_rate=0.05)
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
checkpoint = trackable_utils.Checkpoint(
|
||||
model=model, optimizer=optimizer)
|
||||
for _ in range(2):
|
||||
checkpoint.save(checkpoint_prefix)
|
||||
with backprop.GradientTape() as tape:
|
||||
loss = (constant_op.constant(1.)
|
||||
- model(constant_op.constant(1.))) ** 2
|
||||
grad = tape.gradient(loss, model.vars)
|
||||
optimizer.apply_gradients(
|
||||
[(g, v) for g, v in zip(grad, model.vars)])
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testDeferredSlotRestoration(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
@ -652,84 +193,6 @@ class CheckpointingTests(test.TestCase):
|
||||
beta1_power, _ = optimizer._get_beta_accumulators()
|
||||
self.assertAllEqual(3., self.evaluate(beta1_power))
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def test_initialize_if_not_restoring(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
optimizer_only_prefix = os.path.join(checkpoint_directory, "opt")
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
model=model, # Do not save the optimizer with the checkpoint.
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
optimizer_checkpoint = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer)
|
||||
|
||||
checkpoint_path = checkpoint_management.latest_checkpoint(
|
||||
checkpoint_directory)
|
||||
status = root.restore(save_path=checkpoint_path)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
self.evaluate([v.initializer for v in optimizer.variables()])
|
||||
train_fn()
|
||||
model_save_path = root.save(file_prefix=checkpoint_prefix)
|
||||
self.evaluate(optimizer.variables()[0].assign(42.))
|
||||
optimizer_save_path = optimizer_checkpoint.save(optimizer_only_prefix)
|
||||
|
||||
# Restore into a graph with the optimizer
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
status = root.restore(save_path=model_save_path)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
status.initialize_or_restore()
|
||||
train_fn()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_consumed()
|
||||
|
||||
# Make sure initialization doesn't clobber later restores
|
||||
with test_util.device(use_gpu=True):
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001, beta1=1.0)
|
||||
root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model,
|
||||
global_step=training_util.get_or_create_global_step())
|
||||
opt_root = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer)
|
||||
status = root.restore(save_path=model_save_path)
|
||||
init_only_optimizer_status = opt_root.restore(save_path=None)
|
||||
optimizer_status = opt_root.restore(save_path=optimizer_save_path)
|
||||
input_value = constant_op.constant([[3.]])
|
||||
train_fn = functools.partial(
|
||||
optimizer.minimize,
|
||||
functools.partial(model, input_value),
|
||||
global_step=root.global_step)
|
||||
if not context.executing_eagerly():
|
||||
train_fn = functools.partial(self.evaluate, train_fn())
|
||||
optimizer_status.run_restore_ops()
|
||||
status.initialize_or_restore()
|
||||
init_only_optimizer_status.initialize_or_restore()
|
||||
train_fn()
|
||||
self.assertEqual(42., self.evaluate(optimizer.variables()[0]))
|
||||
|
||||
|
||||
class _ManualScope(tracking.AutoTrackable):
|
||||
|
||||
@ -797,132 +260,5 @@ class TemplateTests(test.TestCase):
|
||||
self.assertAllEqual([14.], self.evaluate(var2))
|
||||
|
||||
|
||||
class CheckpointCompatibilityTests(test.TestCase):
|
||||
|
||||
def _initialized_model(self):
|
||||
input_value = constant_op.constant([[3.]])
|
||||
model = MyModel()
|
||||
optimizer = adam.AdamOptimizer(0.001)
|
||||
optimizer_step = training_util.get_or_create_global_step()
|
||||
root_trackable = trackable_utils.Checkpoint(
|
||||
optimizer=optimizer, model=model, optimizer_step=optimizer_step)
|
||||
train_op = optimizer.minimize(
|
||||
functools.partial(model, input_value),
|
||||
global_step=optimizer_step)
|
||||
self.evaluate(trackable_utils.gather_initializers(
|
||||
root_trackable))
|
||||
self.evaluate(train_op)
|
||||
# A regular variable, a slot variable, and a non-slot Optimizer variable
|
||||
# with known values to check when loading.
|
||||
self.evaluate(model._named_dense.bias.assign([1.]))
|
||||
self.evaluate(optimizer.get_slot(
|
||||
var=model._named_dense.bias, name="m").assign([2.]))
|
||||
beta1_power, _ = optimizer._get_beta_accumulators()
|
||||
self.evaluate(beta1_power.assign(3.))
|
||||
return root_trackable
|
||||
|
||||
def _set_sentinels(self, root_trackable):
|
||||
self.evaluate(root_trackable.model._named_dense.bias.assign([101.]))
|
||||
self.evaluate(
|
||||
root_trackable.optimizer.get_slot(
|
||||
var=root_trackable.model._named_dense.bias, name="m")
|
||||
.assign([102.]))
|
||||
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
||||
self.evaluate(beta1_power.assign(103.))
|
||||
|
||||
def _check_sentinels(self, root_trackable):
|
||||
self.assertAllEqual(
|
||||
[1.], self.evaluate(root_trackable.model._named_dense.bias))
|
||||
self.assertAllEqual([2.], self.evaluate(
|
||||
root_trackable.optimizer.get_slot(
|
||||
var=root_trackable.model._named_dense.bias, name="m")))
|
||||
beta1_power, _ = root_trackable.optimizer._get_beta_accumulators()
|
||||
self.assertAllEqual(3., self.evaluate(beta1_power))
|
||||
|
||||
def _write_name_based_checkpoint(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
with context.graph_mode():
|
||||
save_graph = ops.Graph()
|
||||
with save_graph.as_default(), self.session(
|
||||
graph=save_graph) as session:
|
||||
root = self._initialized_model()
|
||||
name_saver = saver_lib.Saver()
|
||||
return name_saver.save(
|
||||
sess=session, save_path=checkpoint_prefix,
|
||||
global_step=root.optimizer_step)
|
||||
|
||||
@test_util.run_in_graph_and_eager_modes
|
||||
def testLoadFromNameBasedSaver(self):
|
||||
"""Save a name-based checkpoint, load it using the object-based API."""
|
||||
with test_util.device(use_gpu=True):
|
||||
save_path = self._write_name_based_checkpoint()
|
||||
root = self._initialized_model()
|
||||
self._set_sentinels(root)
|
||||
with self.assertRaises(AssertionError):
|
||||
self._check_sentinels(root)
|
||||
object_saver = trackable_utils.TrackableSaver(
|
||||
graph_view.ObjectGraphView(root))
|
||||
self._set_sentinels(root)
|
||||
status = object_saver.restore(save_path)
|
||||
if context.executing_eagerly():
|
||||
self._check_sentinels(root)
|
||||
if context.executing_eagerly():
|
||||
status.assert_consumed()
|
||||
status.assert_existing_objects_matched()
|
||||
status.assert_nontrivial_match()
|
||||
else:
|
||||
# When graph building, we haven't read any keys, so we don't know
|
||||
# whether the restore will be complete.
|
||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||
status.assert_consumed()
|
||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||
status.assert_existing_objects_matched()
|
||||
with self.assertRaisesRegexp(AssertionError, "not restored"):
|
||||
status.assert_nontrivial_match()
|
||||
status.run_restore_ops()
|
||||
self._check_sentinels(root)
|
||||
self._set_sentinels(root)
|
||||
status = object_saver.restore(save_path)
|
||||
status.initialize_or_restore()
|
||||
self._check_sentinels(root)
|
||||
# Check that there is no error when keys are missing from the name-based
|
||||
# checkpoint.
|
||||
root.not_in_name_checkpoint = resource_variable_ops.ResourceVariable([1.])
|
||||
status = object_saver.restore(save_path)
|
||||
with self.assertRaises(AssertionError):
|
||||
status.assert_existing_objects_matched()
|
||||
|
||||
def testSaveGraphLoadEager(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
with context.graph_mode():
|
||||
save_graph = ops.Graph()
|
||||
with save_graph.as_default(), self.session(
|
||||
graph=save_graph):
|
||||
root = self._initialized_model()
|
||||
save_path = root.save(file_prefix=checkpoint_prefix)
|
||||
with context.eager_mode():
|
||||
root = self._initialized_model()
|
||||
self._set_sentinels(root)
|
||||
root.restore(save_path).assert_consumed()
|
||||
self._check_sentinels(root)
|
||||
|
||||
def testSaveEagerLoadGraph(self):
|
||||
checkpoint_directory = self.get_temp_dir()
|
||||
checkpoint_prefix = os.path.join(checkpoint_directory, "ckpt")
|
||||
with context.eager_mode():
|
||||
root = self._initialized_model()
|
||||
save_path = root.save(file_prefix=checkpoint_prefix)
|
||||
with context.graph_mode():
|
||||
save_graph = ops.Graph()
|
||||
with save_graph.as_default(), self.session(
|
||||
graph=save_graph):
|
||||
root = self._initialized_model()
|
||||
self._set_sentinels(root)
|
||||
root.restore(save_path).assert_consumed().run_restore_ops()
|
||||
self._check_sentinels(root)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
test.main()
|
||||
|
Loading…
Reference in New Issue
Block a user