From 94f361d057c969afa5dd83eac4aa03697dc7766d Mon Sep 17 00:00:00 2001
From: Ayush Dubey <ayushd@google.com>
Date: Thu, 25 Jun 2020 14:08:29 -0700
Subject: [PATCH] Remove `run_deprecated_v1` annotations from collective ops
 tests.

PiperOrigin-RevId: 318344535
Change-Id: Idce034f9c0eba341047a76e74a3a57d6220c786f
---
 .../python/ops/collective_ops_gpu_test.py     | 37 ++++++++++---------
 .../python/ops/collective_ops_xla_test.py     |  5 +--
 2 files changed, 21 insertions(+), 21 deletions(-)

diff --git a/tensorflow/python/ops/collective_ops_gpu_test.py b/tensorflow/python/ops/collective_ops_gpu_test.py
index dfa4d445b0d..872fb49834c 100644
--- a/tensorflow/python/ops/collective_ops_gpu_test.py
+++ b/tensorflow/python/ops/collective_ops_gpu_test.py
@@ -60,7 +60,6 @@ class CollectiveOpGPUTest(test.TestCase):
           len(gpus)))
     context.ensure_initialized()
 
-  @test_util.run_deprecated_v1
   def testBasicNcclAllReduce(self):
     inputs = [[0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1],
               [0.3, 1.3, 2.3, 3.3, 4.3, 5.3, 6.3, 7.3]]
@@ -69,7 +68,9 @@ class CollectiveOpGPUTest(test.TestCase):
     instance_key = 1
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(config=self._configure()) as sess:
+    # Tests that execute collectives need to be enclosed in graph or tf.function
+    with ops.Graph().as_default(), self.session(
+        config=self._configure()) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
       collectives = []
@@ -82,14 +83,15 @@ class CollectiveOpGPUTest(test.TestCase):
     for result in results:
       self.assertAllClose(result, expected, rtol=1e-5, atol=1e-5)
 
-  @test_util.run_deprecated_v1
   def testInt32Error(self):
     inputs = [[0, 1], [2, 3]]
     group_key = 1
     instance_key = 50
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(config=self._configure()) as sess:
+    # Tests that execute collectives need to be enclosed in graph or tf.function
+    with ops.Graph().as_default(), self.session(
+        config=self._configure()) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
       collectives = []
@@ -103,7 +105,6 @@ class CollectiveOpGPUTest(test.TestCase):
           'does not support datatype DT_INT32 on DEVICE_GPU'):
         sess.run(collectives)
 
-  @test_util.run_deprecated_v1
   def testFp16Reduce(self):
     inputs = [[0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1],
               [0.3, 1.3, 2.3, 3.3, 4.3, 5.3, 6.3, 7.3]]
@@ -112,7 +113,8 @@ class CollectiveOpGPUTest(test.TestCase):
     instance_key = 100
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(config=self._configure()) as sess:
+    with ops.Graph().as_default(), self.session(
+        config=self._configure()) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
       collectives = []
@@ -126,7 +128,6 @@ class CollectiveOpGPUTest(test.TestCase):
       logging.info('i {} result {} expected {}'.format(i, results[i], expected))
       self.assertAllClose(result, expected, rtol=1e-3, atol=1e-3)
 
-  @test_util.run_deprecated_v1
   def testNcclHintAllReduce(self):
     inputs = [[0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1],
               [0.3, 1.3, 2.3, 3.3, 4.3, 5.3, 6.3, 7.3]]
@@ -135,7 +136,7 @@ class CollectiveOpGPUTest(test.TestCase):
     instance_key = 1
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(
+    with ops.Graph().as_default(), self.session(
         config=self._configure(set_config_proto_nccl=False)) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
@@ -150,14 +151,14 @@ class CollectiveOpGPUTest(test.TestCase):
     for result in results:
       self.assertAllClose(result, expected, rtol=1e-5, atol=1e-5)
 
-  @test_util.run_deprecated_v1
   def testBasicNcclBroadcast(self):
     tensor_value = [0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1]
     group_key = 1
     instance_key = 1
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(config=self._configure()) as sess:
+    with ops.Graph().as_default(), self.session(
+        config=self._configure()) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
       collectives = []
@@ -173,14 +174,14 @@ class CollectiveOpGPUTest(test.TestCase):
     for result in results:
       self.assertAllClose(result, tensor_value, rtol=1e-5, atol=1e-5)
 
-  @test_util.run_deprecated_v1
   def testNcclBroadcastDoubleRecv(self):
     tensor_value = [0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1]
     group_key = 1
     instance_key = 1
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(config=self._configure()) as sess:
+    with ops.Graph().as_default(), self.session(
+        config=self._configure()) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
       collectives = []
@@ -192,14 +193,14 @@ class CollectiveOpGPUTest(test.TestCase):
       with self.assertRaisesRegexp(errors.InternalError, 'found no source'):
         sess.run(collectives)
 
-  @test_util.run_deprecated_v1
   def testNcclBroadcastDoubleSend(self):
     tensor_value = [0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1]
     group_key = 1
     instance_key = 1
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(config=self._configure()) as sess:
+    with ops.Graph().as_default(), self.session(
+        config=self._configure()) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
       collectives = []
@@ -211,7 +212,6 @@ class CollectiveOpGPUTest(test.TestCase):
       with self.assertRaisesRegexp(errors.InternalError, 'already has source'):
         sess.run(collectives)
 
-  @test_util.run_deprecated_v1
   def testBasicNcclAllGather(self):
     inputs = [[0.1, 1.1, 2.1, 3.1, 4.1, 5.1, 6.1, 7.1],
               [0.3, 1.3, 2.3, 3.3, 4.3, 5.3, 6.3, 7.3]]
@@ -221,7 +221,8 @@ class CollectiveOpGPUTest(test.TestCase):
     instance_key = 1
     devices = ['/GPU:{}'.format(i) for i in range(self._group_size)]
 
-    with self.session(config=self._configure()) as sess:
+    with ops.Graph().as_default(), self.session(
+        config=self._configure()) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
       collectives = []
@@ -234,13 +235,13 @@ class CollectiveOpGPUTest(test.TestCase):
     for result in results:
       self.assertAllClose(result, expected, rtol=1e-5, atol=1e-5)
 
-  @test_util.run_deprecated_v1
   def testCollectiveDeviceMismatch(self):
     group_key = 10
     instance_key = 20
     t0 = [1, 2, 3, 4]
     t1 = [5, 6, 7, 8]
-    with self.session(
+
+    with ops.Graph().as_default(), self.session(
         config=self._configure(set_config_proto_nccl=False)) as sess:
       if not test_util.is_gpu_available(cuda_only=True):
         self.skipTest('No GPU available')
diff --git a/tensorflow/python/ops/collective_ops_xla_test.py b/tensorflow/python/ops/collective_ops_xla_test.py
index 613dd2527f4..c7550c854e0 100644
--- a/tensorflow/python/ops/collective_ops_xla_test.py
+++ b/tensorflow/python/ops/collective_ops_xla_test.py
@@ -23,7 +23,6 @@ from tensorflow.core.protobuf import rewriter_config_pb2
 from tensorflow.python.eager import def_function
 from tensorflow.python.framework import constant_op
 from tensorflow.python.framework import ops
-from tensorflow.python.framework import test_util
 from tensorflow.python.ops import array_ops
 from tensorflow.python.ops import collective_ops
 from tensorflow.python.platform import test
@@ -31,7 +30,6 @@ from tensorflow.python.platform import test
 
 class CollectiveOpXlaTest(test.TestCase):
 
-  @test_util.run_deprecated_v1
   def testScopedAllocatorWithXla(self):
     group_size = 2
     group_key = 1
@@ -50,7 +48,8 @@ class CollectiveOpXlaTest(test.TestCase):
     del rewrite_options.scoped_allocator_opts.enable_op[:]
     rewrite_options.scoped_allocator_opts.enable_op.append('CollectiveReduce')
 
-    with self.session(config=cfg) as sess:
+    # Tests that execute collectives need to be enclosed in graph or tf.function
+    with ops.Graph().as_default(), self.session(config=cfg) as sess:
       run_ops = []
       for i in range(group_size):
         with ops.device('CPU:%d' % i):