From c49eeeee5463aff02b4bafbd1596288ba4b27739 Mon Sep 17 00:00:00 2001 From: Allen Lavoie Date: Thu, 5 Oct 2017 09:54:37 -0700 Subject: [PATCH] Add a Cython build dependency, start using some Cython tensor utilities PiperOrigin-RevId: 171166294 --- .../core/platform/default/build_config.bzl | 71 +++++++++++- tensorflow/python/BUILD | 11 +- .../python/framework/fast_tensor_util.pyx | 103 ++++++++++++++++++ tensorflow/python/framework/tensor_util.py | 3 +- tensorflow/workspace.bzl | 11 ++ third_party/cython.BUILD | 28 +++++ 6 files changed, 222 insertions(+), 5 deletions(-) create mode 100644 tensorflow/python/framework/fast_tensor_util.pyx create mode 100644 third_party/cython.BUILD diff --git a/tensorflow/core/platform/default/build_config.bzl b/tensorflow/core/platform/default/build_config.bzl index 8a67951b245..51d37291eed 100644 --- a/tensorflow/core/platform/default/build_config.bzl +++ b/tensorflow/core/platform/default/build_config.bzl @@ -28,6 +28,76 @@ def tf_deps(deps, suffix): return tf_deps +# Modified from @cython//:Tools/rules.bzl +def pyx_library( + name, + deps=[], + py_deps=[], + srcs=[], + **kwargs): + """Compiles a group of .pyx / .pxd / .py files. + + First runs Cython to create .cpp files for each input .pyx or .py + .pxd + pair. Then builds a shared object for each, passing "deps" to each cc_binary + rule (includes Python headers by default). Finally, creates a py_library rule + with the shared objects and any pure Python "srcs", with py_deps as its + dependencies; the shared objects can be imported like normal Python files. + + Args: + name: Name for the rule. + deps: C/C++ dependencies of the Cython (e.g. Numpy headers). + py_deps: Pure Python dependencies of the final library. + srcs: .py, .pyx, or .pxd files to either compile or pass through. + **kwargs: Extra keyword arguments passed to the py_library. + """ + # First filter out files that should be run compiled vs. passed through. + py_srcs = [] + pyx_srcs = [] + pxd_srcs = [] + for src in srcs: + if src.endswith(".pyx") or (src.endswith(".py") + and src[:-3] + ".pxd" in srcs): + pyx_srcs.append(src) + elif src.endswith(".py"): + py_srcs.append(src) + else: + pxd_srcs.append(src) + if src.endswith("__init__.py"): + pxd_srcs.append(src) + + # Invoke cython to produce the shared object libraries. + cpp_outs = [src.split(".")[0] + ".cpp" for src in pyx_srcs] + native.genrule( + name = name + "_cython_translation", + srcs = pyx_srcs, + outs = cpp_outs, + cmd = ("PYTHONHASHSEED=0 $(location @cython//:cython_binary) --cplus $(SRCS)" + # Rename outputs to expected location. + + """ && python -c 'import shutil, sys; n = len(sys.argv); [shutil.copyfile(src.split(".")[0] + ".cpp", dst) for src, dst in zip(sys.argv[1:], sys.argv[1+n//2:])]' $(SRCS) $(OUTS)"""), + tools = ["@cython//:cython_binary"] + pxd_srcs, + ) + + shared_objects = [] + for src in pyx_srcs: + stem = src.split(".")[0] + shared_object_name = stem + ".so" + native.cc_binary( + name=shared_object_name, + srcs=[stem + ".cpp"], + deps=deps + ["//util/python:python_headers"], + linkshared = 1, + ) + shared_objects.append(shared_object_name) + + # Now create a py_library with these shared objects as data. + native.py_library( + name=name, + srcs=py_srcs, + deps=py_deps, + srcs_version = "PY2AND3", + data=shared_objects, + **kwargs + ) def _proto_cc_hdrs(srcs, use_grpc_plugin=False): ret = [s[:-len(".proto")] + ".pb.h" for s in srcs] @@ -299,7 +369,6 @@ def tf_additional_proto_srcs(): def tf_additional_all_protos(): return ["//tensorflow/core:protos_all"] - def tf_protos_all_impl(): return ["//tensorflow/core:protos_all_cc_impl"] diff --git a/tensorflow/python/BUILD b/tensorflow/python/BUILD index 3e846cd18ac..407ff079c15 100644 --- a/tensorflow/python/BUILD +++ b/tensorflow/python/BUILD @@ -24,6 +24,7 @@ load("//tensorflow:tensorflow.bzl", "tf_py_wrap_cc") load("//tensorflow:tensorflow.bzl", "tf_cc_shared_object") load("//tensorflow:tensorflow.bzl", "cuda_py_test") load("//tensorflow:tensorflow.bzl", "cuda_py_tests") +load("//tensorflow/core:platform/default/build_config.bzl", "pyx_library") load("//tensorflow/core:platform/default/build_config.bzl", "tf_proto_library") load("//tensorflow/core:platform/default/build_config.bzl", "tf_proto_library_py") load("//tensorflow/core:platform/default/build_config.bzl", "tf_additional_lib_deps") @@ -503,6 +504,7 @@ py_library( ":common_shapes", ":cpp_shape_inference_proto_py", ":errors", + ":framework_fast_tensor_util", ":framework_for_generated_wrappers", ":function", ":graph_util", @@ -733,8 +735,6 @@ py_library( ], ) -# load("//third_party/py/cython:build_defs.bzl", "pyx_library") - py_library( name = "extra_py_tests_deps", srcs_version = "PY2AND3", @@ -4358,3 +4358,10 @@ py_test( "//third_party/py/numpy", ], ) + +pyx_library( + name = "framework_fast_tensor_util", + srcs = ["framework/fast_tensor_util.pyx"], + py_deps = ["//tensorflow/python:util"], + deps = ["//third_party/py/numpy:headers"], +) diff --git a/tensorflow/python/framework/fast_tensor_util.pyx b/tensorflow/python/framework/fast_tensor_util.pyx new file mode 100644 index 00000000000..b43ddb4ad3a --- /dev/null +++ b/tensorflow/python/framework/fast_tensor_util.pyx @@ -0,0 +1,103 @@ +#cython: boundscheck=False +#cython: wraparound=False +#cython: infer_types=True +import numpy as np +cimport numpy as np + +from tensorflow.python.util import compat + + +def AppendFloat32ArrayToTensorProto( + tensor_proto, np.ndarray[np.float32_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.float_val.append(nparray[i]) + + +def AppendFloat64ArrayToTensorProto( + tensor_proto, np.ndarray[np.float64_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.double_val.append(nparray[i]) + + +def AppendInt32ArrayToTensorProto( + tensor_proto, np.ndarray[np.int32_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.int_val.append(nparray[i]) + + +def AppendInt64ArrayToTensorProto( + tensor_proto, np.ndarray[np.int64_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.int64_val.append(nparray[i]) + + +def AppendUInt8ArrayToTensorProto( + tensor_proto, np.ndarray[np.uint8_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.int_val.append(nparray[i]) + + +def AppendUInt16ArrayToTensorProto( + tensor_proto, np.ndarray[np.uint16_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.int_val.append(nparray[i]) + + +def AppendInt16ArrayToTensorProto( + tensor_proto, np.ndarray[np.int16_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.int_val.append(nparray[i]) + + +def AppendInt8ArrayToTensorProto( + tensor_proto, np.ndarray[np.int8_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.int_val.append(nparray[i]) + + +def AppendComplex64ArrayToTensorProto( + tensor_proto, np.ndarray[np.complex64_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.scomplex_val.append(nparray[i].real) + tensor_proto.scomplex_val.append(nparray[i].imag) + + +def AppendComplex128ArrayToTensorProto( + tensor_proto, np.ndarray[np.complex128_t, ndim=1] nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.dcomplex_val.append(nparray[i].real) + tensor_proto.dcomplex_val.append(nparray[i].imag) + + +def AppendObjectArrayToTensorProto(tensor_proto, np.ndarray nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.string_val.append(compat.as_bytes(nparray[i])) + + +def AppendBoolArrayToTensorProto(tensor_proto, nparray): + cdef long i, n + n = nparray.size + for i in range(n): + tensor_proto.bool_val.append(np.asscalar(nparray[i])) diff --git a/tensorflow/python/framework/tensor_util.py b/tensorflow/python/framework/tensor_util.py index 335db92a73d..414c61e9306 100644 --- a/tensorflow/python/framework/tensor_util.py +++ b/tensorflow/python/framework/tensor_util.py @@ -27,8 +27,7 @@ from tensorflow.python.framework import ops from tensorflow.python.framework import tensor_shape from tensorflow.python.util import compat -# TODO(opensource): Add support for pyx_library in the open-source build. -# For now, we use the slow versions that fast_tensor_util replaces. +# Fallback in case fast_tensor_util is not properly compiled. # pylint: disable=g-import-not-at-top try: from tensorflow.python.framework import fast_tensor_util diff --git a/tensorflow/workspace.bzl b/tensorflow/workspace.bzl index f33a942dc92..b226184261a 100644 --- a/tensorflow/workspace.bzl +++ b/tensorflow/workspace.bzl @@ -713,6 +713,17 @@ def tf_workspace(path_prefix="", tf_repo_name=""): actual = "@cub_archive//:cub", ) + native.new_http_archive( + name = "cython", + sha256 = "6dcd30b5ceb887b2b965ee7ceb82ea3acb5f0642fe2206c7636b45acea4798e5", + urls = [ + "http://mirror.bazel.build/github.com/cython/cython/archive/3732784c45cfb040a5b0936951d196f83a12ea17.tar.gz", + "https://github.com/cython/cython/archive/3732784c45cfb040a5b0936951d196f83a12ea17.tar.gz", + ], + strip_prefix = "cython-3732784c45cfb040a5b0936951d196f83a12ea17", + build_file = str(Label("//third_party:cython.BUILD")), + ) + native.http_archive( name = "bazel_toolchains", urls = [ diff --git a/third_party/cython.BUILD b/third_party/cython.BUILD new file mode 100644 index 00000000000..a8e72a1e366 --- /dev/null +++ b/third_party/cython.BUILD @@ -0,0 +1,28 @@ +# Modified version of @cython//:BUILD.bazel + +py_library( + name = "cython_lib", + srcs = glob( + ["Cython/**/*.py"], + exclude = [ + "**/Tests/*.py", + ], + ) + ["cython.py"], + data = glob([ + "Cython/**/*.pyx", + "Cython/Utility/*.*", + "Cython/Includes/**/*.pxd", + ]), + srcs_version = "PY2AND3", + visibility = ["//visibility:public"], +) + +# May not be named "cython", since that conflicts with Cython/ on OSX +py_binary( + name = "cython_binary", + srcs = ["cython.py"], + main = "cython.py", + srcs_version = "PY2AND3", + visibility = ["//visibility:public"], + deps = ["cython_lib"], +)