Retrying to remove resource_loader, now with oss fix.
PiperOrigin-RevId: 221091607
This commit is contained in:
parent
499b8bd16d
commit
7e511f2db8
@ -0,0 +1,10 @@
|
|||||||
|
op {
|
||||||
|
graph_op_name: "SerializeTensor"
|
||||||
|
endpoint {
|
||||||
|
name: "io.serialize_tensor"
|
||||||
|
}
|
||||||
|
endpoint {
|
||||||
|
name: "serialize_tensor"
|
||||||
|
deprecation_version: 2
|
||||||
|
}
|
||||||
|
}
|
@ -24,7 +24,7 @@ from tensorflow.python.util import tf_inspect as _inspect
|
|||||||
from tensorflow.python.util.tf_export import tf_export
|
from tensorflow.python.util.tf_export import tf_export
|
||||||
|
|
||||||
|
|
||||||
@tf_export('resource_loader.load_resource')
|
@tf_export(v1=['resource_loader.load_resource'])
|
||||||
def load_resource(path):
|
def load_resource(path):
|
||||||
"""Load the resource at given path, where path is relative to tensorflow/.
|
"""Load the resource at given path, where path is relative to tensorflow/.
|
||||||
|
|
||||||
@ -46,7 +46,7 @@ def load_resource(path):
|
|||||||
|
|
||||||
|
|
||||||
# pylint: disable=protected-access
|
# pylint: disable=protected-access
|
||||||
@tf_export('resource_loader.get_data_files_path')
|
@tf_export(v1=['resource_loader.get_data_files_path'])
|
||||||
def get_data_files_path():
|
def get_data_files_path():
|
||||||
"""Get a direct path to the data files colocated with the script.
|
"""Get a direct path to the data files colocated with the script.
|
||||||
|
|
||||||
@ -57,7 +57,7 @@ def get_data_files_path():
|
|||||||
return _os.path.dirname(_inspect.getfile(_sys._getframe(1)))
|
return _os.path.dirname(_inspect.getfile(_sys._getframe(1)))
|
||||||
|
|
||||||
|
|
||||||
@tf_export('resource_loader.get_root_dir_with_all_resources')
|
@tf_export(v1=['resource_loader.get_root_dir_with_all_resources'])
|
||||||
def get_root_dir_with_all_resources():
|
def get_root_dir_with_all_resources():
|
||||||
"""Get a root directory containing all the data attributes in the build rule.
|
"""Get a root directory containing all the data attributes in the build rule.
|
||||||
|
|
||||||
@ -97,7 +97,7 @@ def get_root_dir_with_all_resources():
|
|||||||
return data_files_dir or script_dir
|
return data_files_dir or script_dir
|
||||||
|
|
||||||
|
|
||||||
@tf_export('resource_loader.get_path_to_datafile')
|
@tf_export(v1=['resource_loader.get_path_to_datafile'])
|
||||||
def get_path_to_datafile(path):
|
def get_path_to_datafile(path):
|
||||||
"""Get the path to the specified file in the data dependencies.
|
"""Get the path to the specified file in the data dependencies.
|
||||||
|
|
||||||
@ -117,7 +117,7 @@ def get_path_to_datafile(path):
|
|||||||
return _os.path.join(data_files_path, path)
|
return _os.path.join(data_files_path, path)
|
||||||
|
|
||||||
|
|
||||||
@tf_export('resource_loader.readahead_file_path')
|
@tf_export(v1=['resource_loader.readahead_file_path'])
|
||||||
def readahead_file_path(path, readahead='128M'): # pylint: disable=unused-argument
|
def readahead_file_path(path, readahead='128M'): # pylint: disable=unused-argument
|
||||||
"""Readahead files not implemented; simply returns given path."""
|
"""Readahead files not implemented; simply returns given path."""
|
||||||
return path
|
return path
|
||||||
|
@ -71,7 +71,6 @@ TENSORFLOW_API_INIT_FILES = [
|
|||||||
"profiler/__init__.py",
|
"profiler/__init__.py",
|
||||||
"quantization/__init__.py",
|
"quantization/__init__.py",
|
||||||
"random/__init__.py",
|
"random/__init__.py",
|
||||||
"resource_loader/__init__.py",
|
|
||||||
"strings/__init__.py",
|
"strings/__init__.py",
|
||||||
"saved_model/__init__.py",
|
"saved_model/__init__.py",
|
||||||
"sets/__init__.py",
|
"sets/__init__.py",
|
||||||
|
@ -112,6 +112,10 @@ tf_module {
|
|||||||
name: "serialize_sparse"
|
name: "serialize_sparse"
|
||||||
argspec: "args=[\'sp_input\', \'name\', \'out_type\'], varargs=None, keywords=None, defaults=[\'None\', \"<dtype: \'string\'>\"], "
|
argspec: "args=[\'sp_input\', \'name\', \'out_type\'], varargs=None, keywords=None, defaults=[\'None\', \"<dtype: \'string\'>\"], "
|
||||||
}
|
}
|
||||||
|
member_method {
|
||||||
|
name: "serialize_tensor"
|
||||||
|
argspec: "args=[\'tensor\', \'name\'], varargs=None, keywords=None, defaults=[\'None\'], "
|
||||||
|
}
|
||||||
member_method {
|
member_method {
|
||||||
name: "tf_record_iterator"
|
name: "tf_record_iterator"
|
||||||
argspec: "args=[\'path\', \'options\'], varargs=None, keywords=None, defaults=[\'None\'], "
|
argspec: "args=[\'path\', \'options\'], varargs=None, keywords=None, defaults=[\'None\'], "
|
||||||
|
@ -112,6 +112,10 @@ tf_module {
|
|||||||
name: "serialize_sparse"
|
name: "serialize_sparse"
|
||||||
argspec: "args=[\'sp_input\', \'name\', \'out_type\'], varargs=None, keywords=None, defaults=[\'None\', \"<dtype: \'string\'>\"], "
|
argspec: "args=[\'sp_input\', \'name\', \'out_type\'], varargs=None, keywords=None, defaults=[\'None\', \"<dtype: \'string\'>\"], "
|
||||||
}
|
}
|
||||||
|
member_method {
|
||||||
|
name: "serialize_tensor"
|
||||||
|
argspec: "args=[\'tensor\', \'name\'], varargs=None, keywords=None, defaults=[\'None\'], "
|
||||||
|
}
|
||||||
member_method {
|
member_method {
|
||||||
name: "tf_record_iterator"
|
name: "tf_record_iterator"
|
||||||
argspec: "args=[\'path\', \'options\'], varargs=None, keywords=None, defaults=[\'None\'], "
|
argspec: "args=[\'path\', \'options\'], varargs=None, keywords=None, defaults=[\'None\'], "
|
||||||
|
@ -376,10 +376,6 @@ tf_module {
|
|||||||
name: "resource"
|
name: "resource"
|
||||||
mtype: "<class \'tensorflow.python.framework.dtypes.DType\'>"
|
mtype: "<class \'tensorflow.python.framework.dtypes.DType\'>"
|
||||||
}
|
}
|
||||||
member {
|
|
||||||
name: "resource_loader"
|
|
||||||
mtype: "<type \'module\'>"
|
|
||||||
}
|
|
||||||
member {
|
member {
|
||||||
name: "saved_model"
|
name: "saved_model"
|
||||||
mtype: "<type \'module\'>"
|
mtype: "<type \'module\'>"
|
||||||
@ -1116,10 +1112,6 @@ tf_module {
|
|||||||
name: "sequence_mask"
|
name: "sequence_mask"
|
||||||
argspec: "args=[\'lengths\', \'maxlen\', \'dtype\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \"<dtype: \'bool\'>\", \'None\'], "
|
argspec: "args=[\'lengths\', \'maxlen\', \'dtype\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \"<dtype: \'bool\'>\", \'None\'], "
|
||||||
}
|
}
|
||||||
member_method {
|
|
||||||
name: "serialize_tensor"
|
|
||||||
argspec: "args=[\'tensor\', \'name\'], varargs=None, keywords=None, defaults=[\'None\'], "
|
|
||||||
}
|
|
||||||
member_method {
|
member_method {
|
||||||
name: "set_random_seed"
|
name: "set_random_seed"
|
||||||
argspec: "args=[\'seed\'], varargs=None, keywords=None, defaults=None"
|
argspec: "args=[\'seed\'], varargs=None, keywords=None, defaults=None"
|
||||||
|
@ -1,23 +0,0 @@
|
|||||||
path: "tensorflow.resource_loader"
|
|
||||||
tf_module {
|
|
||||||
member_method {
|
|
||||||
name: "get_data_files_path"
|
|
||||||
argspec: "args=[], varargs=None, keywords=None, defaults=None"
|
|
||||||
}
|
|
||||||
member_method {
|
|
||||||
name: "get_path_to_datafile"
|
|
||||||
argspec: "args=[\'path\'], varargs=None, keywords=None, defaults=None"
|
|
||||||
}
|
|
||||||
member_method {
|
|
||||||
name: "get_root_dir_with_all_resources"
|
|
||||||
argspec: "args=[], varargs=None, keywords=None, defaults=None"
|
|
||||||
}
|
|
||||||
member_method {
|
|
||||||
name: "load_resource"
|
|
||||||
argspec: "args=[\'path\'], varargs=None, keywords=None, defaults=None"
|
|
||||||
}
|
|
||||||
member_method {
|
|
||||||
name: "readahead_file_path"
|
|
||||||
argspec: "args=[\'path\', \'readahead\'], varargs=None, keywords=None, defaults=[\'128M\'], "
|
|
||||||
}
|
|
||||||
}
|
|
Loading…
Reference in New Issue
Block a user