[TF:XLA] Attach useful information to disabled XLA tests.
PiperOrigin-RevId: 232669090
This commit is contained in:
parent
4232ed9143
commit
9f6937e0d1
@ -284,7 +284,6 @@ class FunctionTest(test.TestCase):
|
|||||||
out, = sess.run(dlogits, {logits: x, labels: y})
|
out, = sess.run(dlogits, {logits: x, labels: y})
|
||||||
self.assertAllClose(out, np.exp(prob - y))
|
self.assertAllClose(out, np.exp(prob - y))
|
||||||
|
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
|
||||||
def testCustomGradientError(self):
|
def testCustomGradientError(self):
|
||||||
dtype = dtypes.float32
|
dtype = dtypes.float32
|
||||||
|
|
||||||
|
@ -1460,7 +1460,7 @@ def disable_all_xla(description):
|
|||||||
value = getattr(cls, name)
|
value = getattr(cls, name)
|
||||||
if callable(value) and name.startswith(
|
if callable(value) and name.startswith(
|
||||||
"test") and not name == "test_session":
|
"test") and not name == "test_session":
|
||||||
setattr(cls, name, base_decorator(value))
|
setattr(cls, name, base_decorator(description)(value))
|
||||||
return cls
|
return cls
|
||||||
|
|
||||||
return disable_all_impl
|
return disable_all_impl
|
||||||
|
@ -126,7 +126,7 @@ class BenchmarkTest(test.TestCase):
|
|||||||
self.assertFalse(_ran_somebenchmark_2[0])
|
self.assertFalse(_ran_somebenchmark_2[0])
|
||||||
self.assertFalse(_ran_somebenchmark_but_shouldnt[0])
|
self.assertFalse(_ran_somebenchmark_but_shouldnt[0])
|
||||||
|
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123744455") # GPU memory is incorrect
|
||||||
def testReportingBenchmark(self):
|
def testReportingBenchmark(self):
|
||||||
tempdir = test.get_temp_dir()
|
tempdir = test.get_temp_dir()
|
||||||
try:
|
try:
|
||||||
|
@ -889,8 +889,8 @@ class EnsureShapeTest(test.TestCase):
|
|||||||
|
|
||||||
# Dynamic shape check
|
# Dynamic shape check
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA"
|
@test_util.disable_xla(
|
||||||
) # Dynamic shapes not supported now with XLA
|
"b/123337890") # Dynamic shapes not supported now with XLA
|
||||||
def testEnsuresDynamicShape_RaisesError(self):
|
def testEnsuresDynamicShape_RaisesError(self):
|
||||||
placeholder = array_ops.placeholder(dtypes.int32)
|
placeholder = array_ops.placeholder(dtypes.int32)
|
||||||
derived = math_ops.divide(placeholder, 3, name="MyDivide")
|
derived = math_ops.divide(placeholder, 3, name="MyDivide")
|
||||||
@ -904,8 +904,8 @@ class EnsureShapeTest(test.TestCase):
|
|||||||
sess.run(derived, feed_dict={placeholder: feed_val})
|
sess.run(derived, feed_dict={placeholder: feed_val})
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA"
|
@test_util.disable_xla(
|
||||||
) # Dynamic shapes not supported now with XLA
|
"b/123337890") # Dynamic shapes not supported now with XLA
|
||||||
def testEnsuresDynamicShape_RaisesErrorDimUnknown(self):
|
def testEnsuresDynamicShape_RaisesErrorDimUnknown(self):
|
||||||
placeholder = array_ops.placeholder(dtypes.int32)
|
placeholder = array_ops.placeholder(dtypes.int32)
|
||||||
derived = placeholder / 3
|
derived = placeholder / 3
|
||||||
|
@ -163,7 +163,9 @@ class CholeskyOpTest(test.TestCase):
|
|||||||
with self.assertRaises(ValueError):
|
with self.assertRaises(ValueError):
|
||||||
linalg_ops.cholesky(tensor3)
|
linalg_ops.cholesky(tensor3)
|
||||||
|
|
||||||
@test_util.disable_xla("This test never passed for XLA") # all nan on XLA
|
# The below invalid Cholesky call returns an error with TF Classic and just
|
||||||
|
# returns NaNs with XLA.
|
||||||
|
@test_util.disable_xla("b/123337890")
|
||||||
def testNotInvertibleCPU(self):
|
def testNotInvertibleCPU(self):
|
||||||
# The input should be invertible.
|
# The input should be invertible.
|
||||||
with self.session(use_gpu=True):
|
with self.session(use_gpu=True):
|
||||||
|
@ -33,7 +33,6 @@ from tensorflow.python.ops import variables
|
|||||||
from tensorflow.python.platform import test
|
from tensorflow.python.platform import test
|
||||||
|
|
||||||
|
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class ConcatOpTest(test.TestCase):
|
class ConcatOpTest(test.TestCase):
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@ -642,7 +641,6 @@ class ConcatOpTest(test.TestCase):
|
|||||||
self.assertAllEqual([[1, 2, 3, 7, 8, 9], [4, 5, 6, 10, 11, 12]], output)
|
self.assertAllEqual([[1, 2, 3, 7, 8, 9], [4, 5, 6, 10, 11, 12]], output)
|
||||||
|
|
||||||
|
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class ConcatOffsetTest(test.TestCase):
|
class ConcatOffsetTest(test.TestCase):
|
||||||
|
|
||||||
def testBasic(self):
|
def testBasic(self):
|
||||||
@ -686,8 +684,7 @@ class ConcatOffsetTest(test.TestCase):
|
|||||||
self.evaluate(off)
|
self.evaluate(off)
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla(
|
@test_util.disable_xla("b/123337890") # Error messages differ
|
||||||
"This test never passed for XLA") # Different error message on XLA
|
|
||||||
def testSizeMismatch(self):
|
def testSizeMismatch(self):
|
||||||
cdim = constant_op.constant(1, dtypes.int32)
|
cdim = constant_op.constant(1, dtypes.int32)
|
||||||
s0 = constant_op.constant([2, 3, 5], dtypes.int32)
|
s0 = constant_op.constant([2, 3, 5], dtypes.int32)
|
||||||
|
@ -2769,8 +2769,6 @@ class ControlFlowTest(test.TestCase):
|
|||||||
self.assertEqual(i_val, 3)
|
self.assertEqual(i_val, 3)
|
||||||
self.assertAllClose(x_val, 1.0)
|
self.assertAllClose(x_val, 1.0)
|
||||||
|
|
||||||
@test_util.disable_xla("This test never passed for XLA"
|
|
||||||
) # Resource variable issue for ControlFlowV2
|
|
||||||
@test_util.run_gpu_only
|
@test_util.run_gpu_only
|
||||||
def testGpuResourceAccess(self):
|
def testGpuResourceAccess(self):
|
||||||
with ops.device(test.gpu_device_name()):
|
with ops.device(test.gpu_device_name()):
|
||||||
|
@ -295,7 +295,7 @@ class DepthToSpaceTest(test.TestCase):
|
|||||||
actual_vals, expected_vals = self.evaluate([actual, expected])
|
actual_vals, expected_vals = self.evaluate([actual, expected])
|
||||||
self.assertTrue(np.array_equal(actual_vals, expected_vals))
|
self.assertTrue(np.array_equal(actual_vals, expected_vals))
|
||||||
|
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123553551") # Unsupported data format
|
||||||
def testAgainstTranspose(self):
|
def testAgainstTranspose(self):
|
||||||
self.compareToTranspose(3, 2, 3, 1, 2, "NHWC", False)
|
self.compareToTranspose(3, 2, 3, 1, 2, "NHWC", False)
|
||||||
self.compareToTranspose(3, 2, 3, 2, 2, "NHWC", False)
|
self.compareToTranspose(3, 2, 3, 2, 2, "NHWC", False)
|
||||||
|
@ -65,7 +65,7 @@ class MatrixDiagTest(test.TestCase):
|
|||||||
array_ops.matrix_diag(0)
|
array_ops.matrix_diag(0)
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123337890") # Error messages differ
|
||||||
def testInvalidShapeAtEval(self):
|
def testInvalidShapeAtEval(self):
|
||||||
with self.session(use_gpu=True):
|
with self.session(use_gpu=True):
|
||||||
v = array_ops.placeholder(dtype=dtypes_lib.float32)
|
v = array_ops.placeholder(dtype=dtypes_lib.float32)
|
||||||
@ -270,7 +270,7 @@ class MatrixDiagPartTest(test.TestCase):
|
|||||||
array_ops.matrix_diag_part(0)
|
array_ops.matrix_diag_part(0)
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123337890") # Error messages differ
|
||||||
def testInvalidShapeAtEval(self):
|
def testInvalidShapeAtEval(self):
|
||||||
with self.session(use_gpu=True):
|
with self.session(use_gpu=True):
|
||||||
v = array_ops.placeholder(dtype=dtypes_lib.float32)
|
v = array_ops.placeholder(dtype=dtypes_lib.float32)
|
||||||
|
@ -56,7 +56,6 @@ def simple_scoped_fn(a, x):
|
|||||||
|
|
||||||
|
|
||||||
@test_util.with_control_flow_v2
|
@test_util.with_control_flow_v2
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class FunctionalOpsTest(test.TestCase):
|
class FunctionalOpsTest(test.TestCase):
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
@test_util.run_in_graph_and_eager_modes
|
||||||
@ -660,8 +659,7 @@ class FunctionalOpsTest(test.TestCase):
|
|||||||
self.assertAllEqual(Run(100., True), 5050.)
|
self.assertAllEqual(Run(100., True), 5050.)
|
||||||
|
|
||||||
@test_util.run_v1_only("b/120545219")
|
@test_util.run_v1_only("b/120545219")
|
||||||
@test_util.disable_xla(
|
@test_util.disable_xla("b/123337890") # Different error message
|
||||||
"This test never passed for XLA") # Different error message
|
|
||||||
def testWhileError(self):
|
def testWhileError(self):
|
||||||
for use_gpu in (True, False):
|
for use_gpu in (True, False):
|
||||||
with ops.Graph().as_default() as g:
|
with ops.Graph().as_default() as g:
|
||||||
@ -938,7 +936,6 @@ class FunctionalOpsTest(test.TestCase):
|
|||||||
|
|
||||||
# TODO(akshayka): Replace `function.Defun` with tf.contrib.eager.defun` in the
|
# TODO(akshayka): Replace `function.Defun` with tf.contrib.eager.defun` in the
|
||||||
# below test cases.
|
# below test cases.
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class PartitionedCallTest(test.TestCase):
|
class PartitionedCallTest(test.TestCase):
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
|
@ -34,7 +34,6 @@ from tensorflow.python.ops import variables
|
|||||||
from tensorflow.python.platform import test
|
from tensorflow.python.platform import test
|
||||||
|
|
||||||
|
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class GatherNdTest(test.TestCase):
|
class GatherNdTest(test.TestCase):
|
||||||
|
|
||||||
def _testSimpleDtype(self, dtype):
|
def _testSimpleDtype(self, dtype):
|
||||||
@ -57,7 +56,7 @@ class GatherNdTest(test.TestCase):
|
|||||||
self._testSimpleDtype("|S") # byte strings in python2 + 3
|
self._testSimpleDtype("|S") # byte strings in python2 + 3
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123337890") # Error messages differ
|
||||||
def testEmptyIndicesAndParamsOKButJustEmptyParamsFails(self):
|
def testEmptyIndicesAndParamsOKButJustEmptyParamsFails(self):
|
||||||
with self.session(use_gpu=True):
|
with self.session(use_gpu=True):
|
||||||
params = np.ones((3, 3), dtype=np.float32)
|
params = np.ones((3, 3), dtype=np.float32)
|
||||||
@ -360,7 +359,6 @@ class GatherNdTest(test.TestCase):
|
|||||||
self.assertAllEqual(expected_grads, ops.convert_to_tensor(grads).eval())
|
self.assertAllEqual(expected_grads, ops.convert_to_tensor(grads).eval())
|
||||||
|
|
||||||
|
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class GatherNdOpBenchmark(test.Benchmark):
|
class GatherNdOpBenchmark(test.Benchmark):
|
||||||
|
|
||||||
def benchmark_gather_nd_op(self):
|
def benchmark_gather_nd_op(self):
|
||||||
|
@ -49,7 +49,6 @@ def simple_scoped_fn(a, x):
|
|||||||
|
|
||||||
|
|
||||||
@test_util.with_control_flow_v2
|
@test_util.with_control_flow_v2
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class MapFnTest(test.TestCase):
|
class MapFnTest(test.TestCase):
|
||||||
|
|
||||||
@test_util.run_in_graph_and_eager_modes
|
@test_util.run_in_graph_and_eager_modes
|
||||||
|
@ -303,7 +303,6 @@ class PoolingTest(test.TestCase):
|
|||||||
self.assertLess(err, err_tolerance)
|
self.assertLess(err, err_tolerance)
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA") # Much larger error
|
|
||||||
def testGradient1D(self):
|
def testGradient1D(self):
|
||||||
with self.session(use_gpu=test.is_gpu_available()):
|
with self.session(use_gpu=test.is_gpu_available()):
|
||||||
for padding in ["SAME", "VALID"]:
|
for padding in ["SAME", "VALID"]:
|
||||||
|
@ -730,7 +730,7 @@ class PoolingTest(test.TestCase):
|
|||||||
t = nn_ops.max_pool(
|
t = nn_ops.max_pool(
|
||||||
t, ksize=ksize, strides=strides, padding="SAME").eval()
|
t, ksize=ksize, strides=strides, padding="SAME").eval()
|
||||||
|
|
||||||
@test_util.disable_xla("b/123338077")
|
@test_util.disable_xla("b/123338077") # Passes with XLA
|
||||||
def testDepthwiseMaxPoolInvalidConfigs(self):
|
def testDepthwiseMaxPoolInvalidConfigs(self):
|
||||||
self._testDepthwiseMaxPoolInvalidConfig(
|
self._testDepthwiseMaxPoolInvalidConfig(
|
||||||
[1, 2, 2, 4], [1, 2, 2, 2], [1, 1, 1, 2],
|
[1, 2, 2, 4], [1, 2, 2, 2], [1, 1, 1, 2],
|
||||||
@ -1821,7 +1821,7 @@ class PoolingTest(test.TestCase):
|
|||||||
padding="SAME")
|
padding="SAME")
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123337890") # Error messages differ
|
||||||
def testOpEdgeCases(self):
|
def testOpEdgeCases(self):
|
||||||
with self.session(use_gpu=test.is_gpu_available()) as sess:
|
with self.session(use_gpu=test.is_gpu_available()) as sess:
|
||||||
pool_funcs = [nn_ops.max_pool, nn_ops.avg_pool]
|
pool_funcs = [nn_ops.max_pool, nn_ops.avg_pool]
|
||||||
|
@ -21,16 +21,17 @@ from __future__ import print_function
|
|||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
from tensorflow.python.framework import ops
|
from tensorflow.python.framework import ops
|
||||||
|
from tensorflow.python.framework import test_util
|
||||||
from tensorflow.python.ops import math_ops
|
from tensorflow.python.ops import math_ops
|
||||||
from tensorflow.python.platform import test
|
from tensorflow.python.platform import test
|
||||||
|
|
||||||
|
|
||||||
class BaseReductionTest(test.TestCase):
|
class BaseReductionTest(test.TestCase):
|
||||||
|
|
||||||
def _tf_reduce(self, x, reduction_axes, keepdims):
|
def _tf_reduce(self, x, reduction_axes, keepdims):
|
||||||
raise NotImplementedError()
|
raise NotImplementedError()
|
||||||
|
|
||||||
|
|
||||||
|
@test_util.disable_all_xla("b/123864762") # Test times out
|
||||||
class BigReductionTest(BaseReductionTest):
|
class BigReductionTest(BaseReductionTest):
|
||||||
"""Test reductions for sum and boolean all over a wide range of shapes."""
|
"""Test reductions for sum and boolean all over a wide range of shapes."""
|
||||||
|
|
||||||
|
@ -86,7 +86,7 @@ class ReluTest(test.TestCase):
|
|||||||
self.assertAllClose(np_relu, tf_relu)
|
self.assertAllClose(np_relu, tf_relu)
|
||||||
self.assertShapeEqual(np_relu, tf_relu)
|
self.assertShapeEqual(np_relu, tf_relu)
|
||||||
|
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123338077") # Passes with XLA
|
||||||
def testReluInt8x4BadShape(self):
|
def testReluInt8x4BadShape(self):
|
||||||
if not test.is_gpu_available(cuda_only=True):
|
if not test.is_gpu_available(cuda_only=True):
|
||||||
self.skipTest("No GPU available")
|
self.skipTest("No GPU available")
|
||||||
|
@ -70,7 +70,6 @@ def handle_options(func, x, axis, exclusive, reverse):
|
|||||||
return x
|
return x
|
||||||
|
|
||||||
|
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class CumsumTest(test.TestCase):
|
class CumsumTest(test.TestCase):
|
||||||
|
|
||||||
valid_dtypes = [
|
valid_dtypes = [
|
||||||
@ -135,6 +134,7 @@ class CumsumTest(test.TestCase):
|
|||||||
self._compareAll(x, axis)
|
self._compareAll(x, axis)
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
|
@test_util.disable_xla("b/123860949") # The computation is constant folded
|
||||||
def testLarge(self):
|
def testLarge(self):
|
||||||
for dtype in self.valid_dtypes:
|
for dtype in self.valid_dtypes:
|
||||||
x = np.ones([1000000], dtype=dtype) / 1024
|
x = np.ones([1000000], dtype=dtype) / 1024
|
||||||
@ -194,7 +194,6 @@ class CumsumTest(test.TestCase):
|
|||||||
self._compareGradient([5, 10], axis, exclusive, reverse)
|
self._compareGradient([5, 10], axis, exclusive, reverse)
|
||||||
|
|
||||||
|
|
||||||
@test_util.disable_all_xla("This test never passed for XLA")
|
|
||||||
class CumprodTest(test.TestCase):
|
class CumprodTest(test.TestCase):
|
||||||
|
|
||||||
valid_dtypes = [
|
valid_dtypes = [
|
||||||
|
@ -296,7 +296,7 @@ class StatefulScatterNdTest(test.TestCase):
|
|||||||
updates).get_shape().as_list(), shape)
|
updates).get_shape().as_list(), shape)
|
||||||
|
|
||||||
@test_util.run_v1_only("b/120545219")
|
@test_util.run_v1_only("b/120545219")
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123337890") # Error messages differ
|
||||||
def testResVarInvalidOutputShape(self):
|
def testResVarInvalidOutputShape(self):
|
||||||
res = variables.Variable(
|
res = variables.Variable(
|
||||||
initial_value=lambda: array_ops.zeros(shape=[], dtype=dtypes.float32),
|
initial_value=lambda: array_ops.zeros(shape=[], dtype=dtypes.float32),
|
||||||
|
@ -285,7 +285,7 @@ class SpaceToDepthTest(test.TestCase):
|
|||||||
actual_vals, expected_vals = self.evaluate([actual, expected])
|
actual_vals, expected_vals = self.evaluate([actual, expected])
|
||||||
self.assertTrue(np.array_equal(actual_vals, expected_vals))
|
self.assertTrue(np.array_equal(actual_vals, expected_vals))
|
||||||
|
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123553551") # Unsupported data format
|
||||||
def testAgainstTranspose(self):
|
def testAgainstTranspose(self):
|
||||||
self.compareToTranspose(3, 2, 3, 1, 2, "NHWC", False)
|
self.compareToTranspose(3, 2, 3, 1, 2, "NHWC", False)
|
||||||
self.compareToTranspose(1, 2, 3, 2, 2, "NHWC", False)
|
self.compareToTranspose(1, 2, 3, 2, 2, "NHWC", False)
|
||||||
|
@ -373,7 +373,7 @@ class SplitOpTest(test.TestCase):
|
|||||||
assert s1.shape.as_list() == [1]
|
assert s1.shape.as_list() == [1]
|
||||||
|
|
||||||
@test_util.run_deprecated_v1
|
@test_util.run_deprecated_v1
|
||||||
@test_util.disable_xla("This test never passed for XLA")
|
@test_util.disable_xla("b/123337890") # Error messages differ
|
||||||
def testNonexistentDimTensor(self):
|
def testNonexistentDimTensor(self):
|
||||||
x = array_ops.placeholder(dtypes.int32)
|
x = array_ops.placeholder(dtypes.int32)
|
||||||
values = np.zeros([5, 30])
|
values = np.zeros([5, 30])
|
||||||
|
Loading…
Reference in New Issue
Block a user