diff --git a/tensorflow/c/c_api_experimental.cc b/tensorflow/c/c_api_experimental.cc index eb875bde85a..e9734427bb0 100644 --- a/tensorflow/c/c_api_experimental.cc +++ b/tensorflow/c/c_api_experimental.cc @@ -777,9 +777,7 @@ TF_Library* TF_LoadPluggableDeviceLibrary(const char* library_filename, } else { status->status = env->LoadDynamicLibrary(library_filename, &lib_handle->lib_handle); - if (status->status.ok()) { - // Init PluggableDevice Plugin - } else { + if (!status->status.ok()) { delete lib_handle; return nullptr; } diff --git a/tensorflow/core/framework/device_base.h b/tensorflow/core/framework/device_base.h index 13e79853188..451490497f2 100644 --- a/tensorflow/core/framework/device_base.h +++ b/tensorflow/core/framework/device_base.h @@ -170,7 +170,6 @@ class DeviceBase { return device_thread_pool_; } - // Does not take ownership. void set_eigen_cpu_device(Eigen::ThreadPoolDevice* d); @@ -204,7 +203,6 @@ class DeviceBase { virtual const Eigen::ThreadPoolDevice* eigen_cpu_device(); - // Caller owns the return value. The OpKernelContext calls this even // for devices that do not implement an eigen_gpu_device. Overridden // by GPU devices to return a derived type. diff --git a/tensorflow/python/framework/load_library.py b/tensorflow/python/framework/load_library.py index ac462eb668d..e4222bfa124 100644 --- a/tensorflow/python/framework/load_library.py +++ b/tensorflow/python/framework/load_library.py @@ -159,7 +159,7 @@ def load_library(library_location): library_location) def load_pluggable_device_library(library_location): - """Loads a Tensorflow PluggableDevice plugin. + """Loads a TensorFlow PluggableDevice plugin. "library_location" can be a path to a specific shared object, or a folder. If it is a folder, all shared objects will be loaded. when the library is loaded, devices/kernels registered in the library via StreamExecutor C API