Automated g4 rollback of changelist 311477582.

PiperOrigin-RevId: 311538137
Change-Id: Id9c4f986f0c5a6408ea60147917fb72977b83efe
This commit is contained in:
A. Unique TensorFlower 2020-05-14 08:37:48 -07:00 committed by TensorFlower Gardener
parent ed01ecd92d
commit 83b0c2a225
11 changed files with 441 additions and 278 deletions

View File

@ -4,61 +4,62 @@ op {
in_arg { in_arg {
name: "values" name: "values"
description: <<END description: <<END
int32 or int64; Tensor containing data to count. Tensor containing data to count.
END END
} }
in_arg { in_arg {
name: "weights" name: "weights"
description: <<END description: <<END
float32; Optional rank 1 Tensor (shape=[max_values]) with weights for each count value. A Tensor of the same shape as indices containing per-index weight values. May
also be the empty tensor if no weights are used.
END END
} }
out_arg { out_arg {
name: "output_indices" name: "output_indices"
description: <<END description: <<END
int64; indices tensor for the resulting sparse tensor object. Indices tensor for the resulting sparse tensor object.
END END
} }
out_arg { out_arg {
name: "output_values" name: "output_values"
description: <<END description: <<END
int64 or float32; values tensor for the resulting sparse tensor object. Values tensor for the resulting sparse tensor object.
END END
} }
out_arg { out_arg {
name: "output_dense_shape" name: "output_dense_shape"
description: <<END description: <<END
int64; shape tensor for the resulting sparse tensor object. Shape tensor for the resulting sparse tensor object.
END END
} }
attr { attr {
name: "T" name: "T"
description: <<END description: <<END
dtype; dtype of the input values tensor. Dtype of the input values tensor.
END END
} }
attr { attr {
name: "minlength" name: "minlength"
description: <<END description: <<END
int32; minimum value to count. Can be set to -1 for no minimum. Minimum value to count. Can be set to -1 for no minimum.
END END
} }
attr { attr {
name: "maxlength" name: "maxlength"
description: <<END description: <<END
int32; maximum value to count. Can be set to -1 for no maximum. Maximum value to count. Can be set to -1 for no maximum.
END END
} }
attr { attr {
name: "binary_count" name: "binary_output"
description: <<END description: <<END
bool; whether to output the number of occurrences of each value or 1. Whether to output the number of occurrences of each value or 1.
END END
} }
attr { attr {
name: "output_type" name: "output_type"
description: <<END description: <<END
dtype; dtype of the output values tensor. Dtype of the output values tensor.
END END
} }
summary: "Performs sparse-output bin counting for a tf.tensor input." summary: "Performs sparse-output bin counting for a tf.tensor input."

View File

@ -4,67 +4,68 @@ op {
in_arg { in_arg {
name: "splits" name: "splits"
description: <<END description: <<END
int64; Tensor containing the row splits of the ragged tensor to count. Tensor containing the row splits of the ragged tensor to count.
END END
} }
in_arg { in_arg {
name: "values" name: "values"
description: <<END description: <<END
int32 or int64; Tensor containing values of the sparse tensor to count. Tensor containing values of the sparse tensor to count.
END END
} }
in_arg { in_arg {
name: "weights" name: "weights"
description: <<END description: <<END
float32; Optional rank 1 Tensor (shape=[max_values]) with weights for each count value. A Tensor of the same shape as indices containing per-index weight values.
May also be the empty tensor if no weights are used.
END END
} }
out_arg { out_arg {
name: "output_indices" name: "output_indices"
description: <<END description: <<END
int64; indices tensor for the resulting sparse tensor object. Indices tensor for the resulting sparse tensor object.
END END
} }
out_arg { out_arg {
name: "output_values" name: "output_values"
description: <<END description: <<END
int64 or float32; values tensor for the resulting sparse tensor object. Values tensor for the resulting sparse tensor object.
END END
} }
out_arg { out_arg {
name: "output_dense_shape" name: "output_dense_shape"
description: <<END description: <<END
int64; shape tensor for the resulting sparse tensor object. Shape tensor for the resulting sparse tensor object.
END END
} }
attr { attr {
name: "T" name: "T"
description: <<END description: <<END
dtype; dtype of the input values tensor. Dtype of the input values tensor.
END END
} }
attr { attr {
name: "minlength" name: "minlength"
description: <<END description: <<END
int32; minimum value to count. Can be set to -1 for no minimum. Minimum value to count. Can be set to -1 for no minimum.
END END
} }
attr { attr {
name: "maxlength" name: "maxlength"
description: <<END description: <<END
int32; maximum value to count. Can be set to -1 for no maximum. Maximum value to count. Can be set to -1 for no maximum.
END END
} }
attr { attr {
name: "binary_count" name: "binary_output"
description: <<END description: <<END
bool; whether to output the number of occurrences of each value or 1. Whether to output the number of occurrences of each value or 1.
END END
} }
attr { attr {
name: "output_type" name: "output_type"
description: <<END description: <<END
dtype; dtype of the output values tensor. Dtype of the output values tensor.
END END
} }
summary: "Performs sparse-output bin counting for a ragged tensor input." summary: "Performs sparse-output bin counting for a ragged tensor input."

View File

@ -4,73 +4,74 @@ op {
in_arg { in_arg {
name: "indices" name: "indices"
description: <<END description: <<END
int64; Tensor containing the indices of the sparse tensor to count. Tensor containing the indices of the sparse tensor to count.
END END
} }
in_arg { in_arg {
name: "values" name: "values"
description: <<END description: <<END
int32 or int64; Tensor containing values of the sparse tensor to count. Tensor containing values of the sparse tensor to count.
END END
} }
in_arg { in_arg {
name: "dense_shape" name: "dense_shape"
description: <<END description: <<END
int64; Tensor containing the dense shape of the sparse tensor to count. Tensor containing the dense shape of the sparse tensor to count.
END END
} }
in_arg { in_arg {
name: "weights" name: "weights"
description: <<END description: <<END
float32; Optional rank 1 Tensor (shape=[max_values]) with weights for each count value. A Tensor of the same shape as indices containing per-index weight values.
May also be the empty tensor if no weights are used.
END END
} }
out_arg { out_arg {
name: "output_indices" name: "output_indices"
description: <<END description: <<END
int64; indices tensor for the resulting sparse tensor object. Indices tensor for the resulting sparse tensor object.
END END
} }
out_arg { out_arg {
name: "output_values" name: "output_values"
description: <<END description: <<END
int64 or float32; values tensor for the resulting sparse tensor object. Values tensor for the resulting sparse tensor object.
END END
} }
out_arg { out_arg {
name: "output_dense_shape" name: "output_dense_shape"
description: <<END description: <<END
int64; shape tensor for the resulting sparse tensor object. Shape tensor for the resulting sparse tensor object.
END END
} }
attr { attr {
name: "T" name: "T"
description: <<END description: <<END
dtype; dtype of the input values tensor. Dtype of the input values tensor.
END END
} }
attr { attr {
name: "minlength" name: "minlength"
description: <<END description: <<END
int32; minimum value to count. Can be set to -1 for no minimum. Minimum value to count. Can be set to -1 for no minimum.
END END
} }
attr { attr {
name: "maxlength" name: "maxlength"
description: <<END description: <<END
int32; maximum value to count. Can be set to -1 for no maximum. Maximum value to count. Can be set to -1 for no maximum.
END END
} }
attr { attr {
name: "binary_count" name: "binary_output"
description: <<END description: <<END
bool; whether to output the number of occurrences of each value or 1. Whether to output the number of occurrences of each value or 1.
END END
} }
attr { attr {
name: "output_type" name: "output_type"
description: <<END description: <<END
dtype; dtype of the output values tensor. Dtype of the output values tensor.
END END
} }
summary: "Performs sparse-output bin counting for a sparse tensor input." summary: "Performs sparse-output bin counting for a sparse tensor input."

View File

@ -16,17 +16,20 @@ limitations under the License.
#include "absl/container/flat_hash_map.h" #include "absl/container/flat_hash_map.h"
#include "tensorflow/core/framework/op_kernel.h" #include "tensorflow/core/framework/op_kernel.h"
#include "tensorflow/core/framework/op_requires.h" #include "tensorflow/core/framework/op_requires.h"
#include "tensorflow/core/framework/register_types.h"
#include "tensorflow/core/framework/tensor.h" #include "tensorflow/core/framework/tensor.h"
#include "tensorflow/core/platform/errors.h" #include "tensorflow/core/platform/errors.h"
#include "tensorflow/core/platform/types.h" #include "tensorflow/core/platform/types.h"
namespace tensorflow { namespace tensorflow {
using BatchedIntMap = std::vector<absl::flat_hash_map<int64, int64>>; template <class T>
using BatchedMap = std::vector<absl::flat_hash_map<int64, T>>;
namespace { namespace {
// TODO(momernick): Extend this function to work with outputs of rank > 2. // TODO(momernick): Extend this function to work with outputs of rank > 2.
Status OutputSparse(const BatchedIntMap& per_batch_counts, int num_values, template <class T>
Status OutputSparse(const BatchedMap<T>& per_batch_counts, int num_values,
bool is_1d, OpKernelContext* context) { bool is_1d, OpKernelContext* context) {
int total_values = 0; int total_values = 0;
int num_batches = per_batch_counts.size(); int num_batches = per_batch_counts.size();
@ -44,12 +47,12 @@ Status OutputSparse(const BatchedIntMap& per_batch_counts, int num_values,
context->allocate_output(1, TensorShape({total_values}), &values)); context->allocate_output(1, TensorShape({total_values}), &values));
auto output_indices = indices->matrix<int64>(); auto output_indices = indices->matrix<int64>();
auto output_values = values->flat<int64>(); auto output_values = values->flat<T>();
int64 value_loc = 0; int64 value_loc = 0;
for (int b = 0; b < num_batches; ++b) { for (int b = 0; b < num_batches; ++b) {
const auto& per_batch_count = per_batch_counts[b]; const auto& per_batch_count = per_batch_counts[b];
std::vector<std::pair<int, int>> pairs(per_batch_count.begin(), std::vector<std::pair<int, T>> pairs(per_batch_count.begin(),
per_batch_count.end()); per_batch_count.end());
std::sort(pairs.begin(), pairs.end()); std::sort(pairs.begin(), pairs.end());
for (const auto& x : pairs) { for (const auto& x : pairs) {
if (is_1d) { if (is_1d) {
@ -77,85 +80,19 @@ Status OutputSparse(const BatchedIntMap& per_batch_counts, int num_values,
return Status::OK(); return Status::OK();
} }
Status OutputWeightedSparse(const BatchedIntMap& per_batch_counts, int GetOutputSize(int max_seen, int max_length, int min_length) {
int num_values, const Tensor& weights, bool is_1d,
OpKernelContext* context) {
if (!TensorShapeUtils::IsVector(weights.shape())) {
return errors::InvalidArgument(
"Weights must be a 1-dimensional tensor. Got: ",
weights.shape().DebugString());
}
if (num_values > weights.dim_size(0)) {
return errors::InvalidArgument("The maximum array value was ", num_values,
", but the weight array has size ",
weights.shape().DebugString());
}
auto weight_values = weights.flat<float>();
int total_values = 0;
int num_batches = per_batch_counts.size();
for (const auto& per_batch_count : per_batch_counts) {
total_values += per_batch_count.size();
}
Tensor* indices;
int inner_dim = is_1d ? 1 : 2;
TF_RETURN_IF_ERROR(context->allocate_output(
0, TensorShape({total_values, inner_dim}), &indices));
Tensor* values;
TF_RETURN_IF_ERROR(
context->allocate_output(1, TensorShape({total_values}), &values));
auto output_indices = indices->matrix<int64>();
auto output_values = values->flat<float>();
int64 value_loc = 0;
for (int b = 0; b < num_batches; ++b) {
const auto& per_batch_count = per_batch_counts[b];
std::vector<std::pair<int, int>> pairs(per_batch_count.begin(),
per_batch_count.end());
std::sort(pairs.begin(), pairs.end());
for (const auto& x : pairs) {
if (is_1d) {
output_indices(value_loc, 0) = x.first;
} else {
output_indices(value_loc, 0) = b;
output_indices(value_loc, 1) = x.first;
}
output_values(value_loc) = x.second * weight_values(x.first);
++value_loc;
}
}
Tensor* dense_shape;
if (is_1d) {
TF_RETURN_IF_ERROR(
context->allocate_output(2, TensorShape({1}), &dense_shape));
dense_shape->flat<int64>().data()[0] = num_values;
} else {
TF_RETURN_IF_ERROR(
context->allocate_output(2, TensorShape({2}), &dense_shape));
dense_shape->flat<int64>().data()[0] = num_batches;
dense_shape->flat<int64>().data()[1] = num_values;
}
return Status::OK();
}
template <class T>
T GetOutputSize(T max_seen, T max_length, T min_length) {
return max_length > 0 ? max_length : std::max((max_seen + 1), min_length); return max_length > 0 ? max_length : std::max((max_seen + 1), min_length);
} }
} // namespace } // namespace
template <class T> template <class T, class W>
class DenseCount : public OpKernel { class DenseCount : public OpKernel {
public: public:
explicit DenseCount(OpKernelConstruction* context) : OpKernel(context) { explicit DenseCount(OpKernelConstruction* context) : OpKernel(context) {
OP_REQUIRES_OK(context, context->GetAttr("minlength", &minlength_)); OP_REQUIRES_OK(context, context->GetAttr("minlength", &minlength_));
OP_REQUIRES_OK(context, context->GetAttr("maxlength", &maxlength_)); OP_REQUIRES_OK(context, context->GetAttr("maxlength", &maxlength_));
OP_REQUIRES_OK(context, context->GetAttr("binary_count", &binary_count_)); OP_REQUIRES_OK(context, context->GetAttr("binary_output", &binary_output_));
} }
void Compute(OpKernelContext* context) override { void Compute(OpKernelContext* context) override {
@ -170,6 +107,15 @@ class DenseCount : public OpKernel {
"Input must be a 1 or 2-dimensional tensor. Got: ", "Input must be a 1 or 2-dimensional tensor. Got: ",
data.shape().DebugString())); data.shape().DebugString()));
if (use_weights) {
OP_REQUIRES(
context, weights.shape() == data.shape(),
errors::InvalidArgument(
"Weights and data must have the same shape. Weight shape: ",
weights.shape().DebugString(),
"; data shape: ", data.shape().DebugString()));
}
bool is_1d = TensorShapeUtils::IsVector(data.shape()); bool is_1d = TensorShapeUtils::IsVector(data.shape());
int negative_valued_axis = -1; int negative_valued_axis = -1;
int num_batch_dimensions = (data.shape().dims() + negative_valued_axis); int num_batch_dimensions = (data.shape().dims() + negative_valued_axis);
@ -179,19 +125,23 @@ class DenseCount : public OpKernel {
num_batch_elements *= data.shape().dim_size(i); num_batch_elements *= data.shape().dim_size(i);
} }
int num_value_elements = data.shape().num_elements() / num_batch_elements; int num_value_elements = data.shape().num_elements() / num_batch_elements;
auto per_batch_counts = BatchedIntMap(num_batch_elements); auto per_batch_counts = BatchedMap<W>(num_batch_elements);
T max_value = 0; T max_value = 0;
const auto data_values = data.flat<T>(); const auto data_values = data.flat<T>();
const auto weight_values = weights.flat<W>();
int i = 0; int i = 0;
for (int b = 0; b < num_batch_elements; ++b) { for (int b = 0; b < num_batch_elements; ++b) {
for (int v = 0; v < num_value_elements; ++v) { for (int v = 0; v < num_value_elements; ++v) {
const auto& value = data_values(i); const auto& value = data_values(i);
if (value >= 0 && (maxlength_ <= 0 || value < maxlength_)) { if (value >= 0 && (maxlength_ <= 0 || value < maxlength_)) {
if (binary_count_) { if (binary_output_) {
(per_batch_counts[b])[value] = 1; per_batch_counts[b][value] = 1;
} else if (use_weights) {
per_batch_counts[b][value] += weight_values(i);
} else { } else {
(per_batch_counts[b])[value]++; per_batch_counts[b][value]++;
} }
if (value > max_value) { if (value > max_value) {
max_value = value; max_value = value;
@ -201,30 +151,24 @@ class DenseCount : public OpKernel {
} }
} }
T num_output_values = GetOutputSize<T>(max_value, maxlength_, minlength_); int num_output_values = GetOutputSize(max_value, maxlength_, minlength_);
if (use_weights) { OP_REQUIRES_OK(context, OutputSparse<W>(per_batch_counts, num_output_values,
OP_REQUIRES_OK(context, is_1d, context));
OutputWeightedSparse(per_batch_counts, num_output_values,
weights, is_1d, context));
} else {
OP_REQUIRES_OK(context, OutputSparse(per_batch_counts, num_output_values,
is_1d, context));
}
} }
private: private:
T minlength_; int maxlength_;
T maxlength_; int minlength_;
bool binary_count_; bool binary_output_;
}; };
template <class T> template <class T, class W>
class SparseCount : public OpKernel { class SparseCount : public OpKernel {
public: public:
explicit SparseCount(OpKernelConstruction* context) : OpKernel(context) { explicit SparseCount(OpKernelConstruction* context) : OpKernel(context) {
OP_REQUIRES_OK(context, context->GetAttr("minlength", &minlength_)); OP_REQUIRES_OK(context, context->GetAttr("minlength", &minlength_));
OP_REQUIRES_OK(context, context->GetAttr("maxlength", &maxlength_)); OP_REQUIRES_OK(context, context->GetAttr("maxlength", &maxlength_));
OP_REQUIRES_OK(context, context->GetAttr("binary_count", &binary_count_)); OP_REQUIRES_OK(context, context->GetAttr("binary_output", &binary_output_));
} }
void Compute(OpKernelContext* context) override { void Compute(OpKernelContext* context) override {
@ -235,23 +179,27 @@ class SparseCount : public OpKernel {
bool use_weights = weights.NumElements() > 0; bool use_weights = weights.NumElements() > 0;
bool is_1d = shape.NumElements() == 1; bool is_1d = shape.NumElements() == 1;
const auto indices_values = indices.matrix<int64>();
const auto values_values = values.flat<T>();
int num_batches = is_1d ? 1 : shape.flat<int64>()(0); int num_batches = is_1d ? 1 : shape.flat<int64>()(0);
int num_values = values.NumElements(); int num_values = values.NumElements();
auto per_batch_counts = BatchedIntMap(num_batches); const auto indices_values = indices.matrix<int64>();
const auto values_values = values.flat<T>();
const auto weight_values = weights.flat<W>();
auto per_batch_counts = BatchedMap<W>(num_batches);
T max_value = 0; T max_value = 0;
for (int idx = 0; idx < num_values; ++idx) { for (int idx = 0; idx < num_values; ++idx) {
int batch = is_1d ? 0 : indices_values(idx, 0); int batch = is_1d ? 0 : indices_values(idx, 0);
const auto& value = values_values(idx); const auto& value = values_values(idx);
if (value >= 0 && (maxlength_ <= 0 || value < maxlength_)) { if (value >= 0 && (maxlength_ <= 0 || value < maxlength_)) {
if (binary_count_) { if (binary_output_) {
(per_batch_counts[batch])[value] = 1; per_batch_counts[batch][value] = 1;
} else if (use_weights) {
per_batch_counts[batch][value] += weight_values(idx);
} else { } else {
(per_batch_counts[batch])[value]++; per_batch_counts[batch][value]++;
} }
if (value > max_value) { if (value > max_value) {
max_value = value; max_value = value;
@ -259,30 +207,25 @@ class SparseCount : public OpKernel {
} }
} }
T num_output_values = GetOutputSize<T>(max_value, maxlength_, minlength_); int num_output_values = GetOutputSize(max_value, maxlength_, minlength_);
if (use_weights) { OP_REQUIRES_OK(context, OutputSparse<W>(per_batch_counts, num_output_values,
OP_REQUIRES_OK(context, is_1d, context));
OutputWeightedSparse(per_batch_counts, num_output_values,
weights, is_1d, context));
} else {
OP_REQUIRES_OK(context, OutputSparse(per_batch_counts, num_output_values,
is_1d, context));
}
} }
private: private:
T minlength_; int maxlength_;
T maxlength_; int minlength_;
bool binary_count_; bool binary_output_;
bool validate_;
}; };
template <class T> template <class T, class W>
class RaggedCount : public OpKernel { class RaggedCount : public OpKernel {
public: public:
explicit RaggedCount(OpKernelConstruction* context) : OpKernel(context) { explicit RaggedCount(OpKernelConstruction* context) : OpKernel(context) {
OP_REQUIRES_OK(context, context->GetAttr("minlength", &minlength_)); OP_REQUIRES_OK(context, context->GetAttr("minlength", &minlength_));
OP_REQUIRES_OK(context, context->GetAttr("maxlength", &maxlength_)); OP_REQUIRES_OK(context, context->GetAttr("maxlength", &maxlength_));
OP_REQUIRES_OK(context, context->GetAttr("binary_count", &binary_count_)); OP_REQUIRES_OK(context, context->GetAttr("binary_output", &binary_output_));
} }
void Compute(OpKernelContext* context) override { void Compute(OpKernelContext* context) override {
@ -290,13 +233,15 @@ class RaggedCount : public OpKernel {
const Tensor& values = context->input(1); const Tensor& values = context->input(1);
const Tensor& weights = context->input(2); const Tensor& weights = context->input(2);
bool use_weights = weights.NumElements() > 0; bool use_weights = weights.NumElements() > 0;
bool is_1d = false;
const auto splits_values = splits.flat<int64>(); const auto splits_values = splits.flat<int64>();
const auto values_values = values.flat<T>(); const auto values_values = values.flat<T>();
const auto weight_values = weights.flat<W>();
int num_batches = splits.NumElements() - 1; int num_batches = splits.NumElements() - 1;
int num_values = values.NumElements(); int num_values = values.NumElements();
auto per_batch_counts = BatchedIntMap(num_batches); auto per_batch_counts = BatchedMap<W>(num_batches);
T max_value = 0; T max_value = 0;
int batch_idx = 0; int batch_idx = 0;
@ -306,10 +251,12 @@ class RaggedCount : public OpKernel {
} }
const auto& value = values_values(idx); const auto& value = values_values(idx);
if (value >= 0 && (maxlength_ <= 0 || value < maxlength_)) { if (value >= 0 && (maxlength_ <= 0 || value < maxlength_)) {
if (binary_count_) { if (binary_output_) {
(per_batch_counts[batch_idx - 1])[value] = 1; per_batch_counts[batch_idx - 1][value] = 1;
} else if (use_weights) {
per_batch_counts[batch_idx - 1][value] += weight_values(idx);
} else { } else {
(per_batch_counts[batch_idx - 1])[value]++; per_batch_counts[batch_idx - 1][value]++;
} }
if (value > max_value) { if (value > max_value) {
max_value = value; max_value = value;
@ -317,42 +264,47 @@ class RaggedCount : public OpKernel {
} }
} }
T num_output_values = GetOutputSize<T>(max_value, maxlength_, minlength_); int num_output_values = GetOutputSize(max_value, maxlength_, minlength_);
if (use_weights) { OP_REQUIRES_OK(context, OutputSparse<W>(per_batch_counts, num_output_values,
OP_REQUIRES_OK(context, is_1d, context));
OutputWeightedSparse(per_batch_counts, num_output_values,
weights, false, context));
} else {
OP_REQUIRES_OK(context, OutputSparse(per_batch_counts, num_output_values,
false, context));
}
} }
private: private:
T minlength_; int maxlength_;
T maxlength_; int minlength_;
bool binary_count_; bool binary_output_;
bool validate_;
}; };
#define REGISTER(TYPE) \ #define REGISTER_W(W_TYPE) \
\ REGISTER(int32, W_TYPE) \
REGISTER_KERNEL_BUILDER(Name("DenseCountSparseOutput") \ REGISTER(int64, W_TYPE)
.TypeConstraint<TYPE>("T") \
.Device(DEVICE_CPU), \
DenseCount<TYPE>) \
\
REGISTER_KERNEL_BUILDER(Name("SparseCountSparseOutput") \
.TypeConstraint<TYPE>("T") \
.Device(DEVICE_CPU), \
SparseCount<TYPE>) \
\
REGISTER_KERNEL_BUILDER(Name("RaggedCountSparseOutput") \
.TypeConstraint<TYPE>("T") \
.Device(DEVICE_CPU), \
RaggedCount<TYPE>)
REGISTER(int32); #define REGISTER(I_TYPE, W_TYPE) \
REGISTER(int64); \
REGISTER_KERNEL_BUILDER(Name("DenseCountSparseOutput") \
.TypeConstraint<I_TYPE>("T") \
.TypeConstraint<W_TYPE>("output_type") \
.Device(DEVICE_CPU), \
DenseCount<I_TYPE, W_TYPE>) \
\
REGISTER_KERNEL_BUILDER(Name("SparseCountSparseOutput") \
.TypeConstraint<I_TYPE>("T") \
.TypeConstraint<W_TYPE>("output_type") \
.Device(DEVICE_CPU), \
SparseCount<I_TYPE, W_TYPE>) \
\
REGISTER_KERNEL_BUILDER(Name("RaggedCountSparseOutput") \
.TypeConstraint<I_TYPE>("T") \
.TypeConstraint<W_TYPE>("output_type") \
.Device(DEVICE_CPU), \
RaggedCount<I_TYPE, W_TYPE>)
TF_CALL_INTEGRAL_TYPES(REGISTER_W);
TF_CALL_float(REGISTER_W);
TF_CALL_double(REGISTER_W);
#undef REGISTER_W
#undef REGISTER #undef REGISTER
} // namespace tensorflow } // namespace tensorflow

View File

@ -19,12 +19,21 @@ limitations under the License.
namespace tensorflow { namespace tensorflow {
using shape_inference::DimensionHandle;
using shape_inference::InferenceContext; using shape_inference::InferenceContext;
using shape_inference::ShapeHandle;
Status DenseCountSparseOutputShapeFn(InferenceContext *c) { Status DenseCountSparseOutputShapeFn(InferenceContext *c) {
int32 rank = c->Rank(c->input(0)); auto values = c->input(0);
DimensionHandle nvals = c->UnknownDim(); auto weights = c->input(1);
ShapeHandle output;
auto num_weights = c->NumElements(weights);
if (c->ValueKnown(num_weights) && c->Value(num_weights) == 0) {
output = values;
} else {
TF_RETURN_IF_ERROR(c->Merge(weights, values, &output));
}
auto rank = c->Rank(output);
auto nvals = c->UnknownDim();
c->set_output(0, c->Matrix(nvals, rank)); // out.indices c->set_output(0, c->Matrix(nvals, rank)); // out.indices
c->set_output(1, c->Vector(nvals)); // out.values c->set_output(1, c->Vector(nvals)); // out.values
c->set_output(2, c->Vector(rank)); // out.dense_shape c->set_output(2, c->Vector(rank)); // out.dense_shape
@ -32,8 +41,8 @@ Status DenseCountSparseOutputShapeFn(InferenceContext *c) {
} }
Status SparseCountSparseOutputShapeFn(InferenceContext *c) { Status SparseCountSparseOutputShapeFn(InferenceContext *c) {
DimensionHandle rank = c->Dim(c->input(0), 1); auto rank = c->Dim(c->input(0), 1);
DimensionHandle nvals = c->UnknownDim(); auto nvals = c->UnknownDim();
c->set_output(0, c->Matrix(nvals, rank)); // out.indices c->set_output(0, c->Matrix(nvals, rank)); // out.indices
c->set_output(1, c->Vector(nvals)); // out.values c->set_output(1, c->Vector(nvals)); // out.values
c->set_output(2, c->Vector(rank)); // out.dense_shape c->set_output(2, c->Vector(rank)); // out.dense_shape
@ -45,7 +54,7 @@ Status RaggedCountSparseOutputShapeFn(InferenceContext *c) {
if (rank != c->kUnknownRank) { if (rank != c->kUnknownRank) {
++rank; // Add the ragged dimension ++rank; // Add the ragged dimension
} }
DimensionHandle nvals = c->UnknownDim(); auto nvals = c->UnknownDim();
c->set_output(0, c->Matrix(nvals, rank)); // out.indices c->set_output(0, c->Matrix(nvals, rank)); // out.indices
c->set_output(1, c->Vector(nvals)); // out.values c->set_output(1, c->Vector(nvals)); // out.values
c->set_output(2, c->Vector(rank)); // out.dense_shape c->set_output(2, c->Vector(rank)); // out.dense_shape
@ -54,12 +63,12 @@ Status RaggedCountSparseOutputShapeFn(InferenceContext *c) {
REGISTER_OP("DenseCountSparseOutput") REGISTER_OP("DenseCountSparseOutput")
.Input("values: T") .Input("values: T")
.Input("weights: float") .Input("weights: output_type")
.Attr("T: {int32, int64}") .Attr("T: {int32, int64}")
.Attr("minlength: int >= -1 = -1") .Attr("minlength: int >= -1 = -1")
.Attr("maxlength: int >= -1 = -1") .Attr("maxlength: int >= -1 = -1")
.Attr("binary_count: bool") .Attr("binary_output: bool")
.Attr("output_type: {int64, float}") .Attr("output_type: {int32, int64, float, double}")
.SetShapeFn(DenseCountSparseOutputShapeFn) .SetShapeFn(DenseCountSparseOutputShapeFn)
.Output("output_indices: int64") .Output("output_indices: int64")
.Output("output_values: output_type") .Output("output_values: output_type")
@ -69,12 +78,12 @@ REGISTER_OP("SparseCountSparseOutput")
.Input("indices: int64") .Input("indices: int64")
.Input("values: T") .Input("values: T")
.Input("dense_shape: int64") .Input("dense_shape: int64")
.Input("weights: float") .Input("weights: output_type")
.Attr("T: {int32, int64}") .Attr("T: {int32, int64}")
.Attr("minlength: int >= -1 = -1") .Attr("minlength: int >= -1 = -1")
.Attr("maxlength: int >= -1 = -1") .Attr("maxlength: int >= -1 = -1")
.Attr("binary_count: bool") .Attr("binary_output: bool")
.Attr("output_type: {int64, float}") .Attr("output_type: {int32, int64, float, double}")
.SetShapeFn(SparseCountSparseOutputShapeFn) .SetShapeFn(SparseCountSparseOutputShapeFn)
.Output("output_indices: int64") .Output("output_indices: int64")
.Output("output_values: output_type") .Output("output_values: output_type")
@ -83,12 +92,12 @@ REGISTER_OP("SparseCountSparseOutput")
REGISTER_OP("RaggedCountSparseOutput") REGISTER_OP("RaggedCountSparseOutput")
.Input("splits: int64") .Input("splits: int64")
.Input("values: T") .Input("values: T")
.Input("weights: float") .Input("weights: output_type")
.Attr("T: {int32, int64}") .Attr("T: {int32, int64}")
.Attr("minlength: int >= -1 = -1") .Attr("minlength: int >= -1 = -1")
.Attr("maxlength: int >= -1 = -1") .Attr("maxlength: int >= -1 = -1")
.Attr("binary_count: bool") .Attr("binary_output: bool")
.Attr("output_type: {int64, float}") .Attr("output_type: {int32, int64, float, double}")
.SetShapeFn(RaggedCountSparseOutputShapeFn) .SetShapeFn(RaggedCountSparseOutputShapeFn)
.Output("output_indices: int64") .Output("output_indices: int64")
.Output("output_values: output_type") .Output("output_values: output_type")

View File

@ -18,10 +18,10 @@ from __future__ import absolute_import
from __future__ import division from __future__ import division
from __future__ import print_function from __future__ import print_function
from tensorflow.python.framework import dtypes
from tensorflow.python.framework import ops from tensorflow.python.framework import ops
from tensorflow.python.framework import sparse_tensor from tensorflow.python.framework import sparse_tensor
from tensorflow.python.ops import array_ops from tensorflow.python.ops import array_ops
from tensorflow.python.ops import check_ops
from tensorflow.python.ops import gen_count_ops from tensorflow.python.ops import gen_count_ops
from tensorflow.python.ops.ragged import ragged_tensor from tensorflow.python.ops.ragged import ragged_tensor
from tensorflow.python.util.tf_export import tf_export from tensorflow.python.util.tf_export import tf_export
@ -33,7 +33,7 @@ def sparse_bincount(values,
axis=0, axis=0,
minlength=None, minlength=None,
maxlength=None, maxlength=None,
binary_count=False, binary_output=False,
name=None): name=None):
"""Count the number of times an integer value appears in a tensor. """Count the number of times an integer value appears in a tensor.
@ -58,8 +58,9 @@ def sparse_bincount(values,
maxlength: If given, skips `values` that are greater than or equal to maxlength: If given, skips `values` that are greater than or equal to
`maxlength`, and ensures that the output has a `dense_shape` of at most `maxlength`, and ensures that the output has a `dense_shape` of at most
`maxlength` in the inner dimension. `maxlength` in the inner dimension.
binary_count: Whether to do a binary count. When True, this op will return 1 binary_output: If True, this op will output 1 instead of the number of times
for any value that exists instead of counting the number of occurrences. a token appears (equivalent to one_hot + reduce_any instead of one_hot +
reduce_add). Defaults to False.
name: A name for this op. name: A name for this op.
Returns: Returns:
@ -78,7 +79,7 @@ def sparse_bincount(values,
SparseTensor) and returns a SparseTensor where the value of (i,j) is the SparseTensor) and returns a SparseTensor where the value of (i,j) is the
number of times value j appears in batch i. number of times value j appears in batch i.
>>> data = [[10, 20, 30, 20], [11, 101, 11, 10001]] >>> data = np.array([[10, 20, 30, 20], [11, 101, 11, 10001]], dtype=np.int64)
>>> output = tf.sparse.bincount(data, axis=-1) >>> output = tf.sparse.bincount(data, axis=-1)
>>> print(output) >>> print(output)
SparseTensor(indices=tf.Tensor( SparseTensor(indices=tf.Tensor(
@ -102,7 +103,7 @@ def sparse_bincount(values,
dense shape is [2, 500] instead of [2,10002] or [2, 102]. dense shape is [2, 500] instead of [2,10002] or [2, 102].
>>> minlength = maxlength = 500 >>> minlength = maxlength = 500
>>> data = [[10, 20, 30, 20], [11, 101, 11, 10001]] >>> data = np.array([[10, 20, 30, 20], [11, 101, 11, 10001]], dtype=np.int64)
>>> output = tf.sparse.bincount( >>> output = tf.sparse.bincount(
... data, axis=-1, minlength=minlength, maxlength=maxlength) ... data, axis=-1, minlength=minlength, maxlength=maxlength)
>>> print(output) >>> print(output)
@ -123,8 +124,8 @@ def sparse_bincount(values,
some values (like 20 in batch 1 and 11 in batch 2) appear more than once, some values (like 20 in batch 1 and 11 in batch 2) appear more than once,
the 'values' tensor is all 1s. the 'values' tensor is all 1s.
>>> dense = [[10, 20, 30, 20], [11, 101, 11, 10001]] >>> data = np.array([[10, 20, 30, 20], [11, 101, 11, 10001]], dtype=np.int64)
>>> output = tf.sparse.bincount(dense, binary_count=True, axis=-1) >>> output = tf.sparse.bincount(data, binary_output=True, axis=-1)
>>> print(output) >>> print(output)
SparseTensor(indices=tf.Tensor( SparseTensor(indices=tf.Tensor(
[[ 0 10] [[ 0 10]
@ -136,20 +137,42 @@ def sparse_bincount(values,
values=tf.Tensor([1 1 1 1 1 1], shape=(6,), dtype=int64), values=tf.Tensor([1 1 1 1 1 1], shape=(6,), dtype=int64),
dense_shape=tf.Tensor([ 2 10002], shape=(2,), dtype=int64)) dense_shape=tf.Tensor([ 2 10002], shape=(2,), dtype=int64))
**Weighted bin-counting**
This example takes two inputs - a values tensor and a weights tensor. These
tensors must be identically shaped, and have the same row splits or indices
in the case of RaggedTensors or SparseTensors. When performing a weighted
count, the op will output a SparseTensor where the value of (i, j) is the
sum of the values in the weight tensor's batch i in the locations where
the values tensor has the value j. In this case, the output dtype is the
same as the dtype of the weights tensor.
>>> data = np.array([[10, 20, 30, 20], [11, 101, 11, 10001]], dtype=np.int64)
>>> weights = [[2, 0.25, 15, 0.5], [2, 17, 3, 0.9]]
>>> output = tf.sparse.bincount(data, weights=weights, axis=-1)
>>> print(output)
SparseTensor(indices=tf.Tensor(
[[ 0 10]
[ 0 20]
[ 0 30]
[ 1 11]
[ 1 101]
[ 1 10001]], shape=(6, 2), dtype=int64),
values=tf.Tensor([2. 0.75 15. 5. 17. 0.9], shape=(6,), dtype=float32),
dense_shape=tf.Tensor([ 2 10002], shape=(2,), dtype=int64))
""" """
with ops.name_scope(name, "count", [values, weights]): with ops.name_scope(name, "count", [values, weights]):
if not isinstance(values, sparse_tensor.SparseTensor): if not isinstance(values, sparse_tensor.SparseTensor):
values = ragged_tensor.convert_to_tensor_or_ragged_tensor( values = ragged_tensor.convert_to_tensor_or_ragged_tensor(
values, name="values") values, name="values")
if weights is not None:
if not isinstance(weights, sparse_tensor.SparseTensor):
weights = ragged_tensor.convert_to_tensor_or_ragged_tensor(
weights, name="weights")
if weights is not None and binary_count: if weights is not None and binary_output:
raise ValueError("binary_count and weights are mutually exclusive.") raise ValueError("binary_output and weights are mutually exclusive.")
if weights is None:
weights = []
output_type = dtypes.int64
else:
output_type = dtypes.float32
if axis is None: if axis is None:
axis = 0 axis = 0
@ -162,38 +185,114 @@ def sparse_bincount(values,
maxlength_value = maxlength if maxlength is not None else -1 maxlength_value = maxlength if maxlength is not None else -1
if axis == 0: if axis == 0:
if isinstance(values, if isinstance(values, sparse_tensor.SparseTensor):
(sparse_tensor.SparseTensor, ragged_tensor.RaggedTensor)): if weights is not None:
weights = validate_sparse_weights(values, weights)
values = values.values
elif isinstance(values, ragged_tensor.RaggedTensor):
if weights is not None:
weights = validate_ragged_weights(values, weights)
values = values.values values = values.values
else: else:
if weights is not None:
weights = array_ops.reshape(weights, [-1])
values = array_ops.reshape(values, [-1]) values = array_ops.reshape(values, [-1])
if isinstance(values, sparse_tensor.SparseTensor): if isinstance(values, sparse_tensor.SparseTensor):
weights = validate_sparse_weights(values, weights)
c_ind, c_val, c_shape = gen_count_ops.sparse_count_sparse_output( c_ind, c_val, c_shape = gen_count_ops.sparse_count_sparse_output(
values.indices, values.indices,
values.values, values.values,
values.dense_shape, values.dense_shape,
weights=weights, weights,
minlength=minlength_value, minlength=minlength_value,
maxlength=maxlength_value, maxlength=maxlength_value,
binary_count=binary_count, binary_output=binary_output)
output_type=output_type)
elif isinstance(values, ragged_tensor.RaggedTensor): elif isinstance(values, ragged_tensor.RaggedTensor):
weights = validate_ragged_weights(values, weights)
c_ind, c_val, c_shape = gen_count_ops.ragged_count_sparse_output( c_ind, c_val, c_shape = gen_count_ops.ragged_count_sparse_output(
values.row_splits, values.row_splits,
values.values, values.values,
weights=weights, weights,
minlength=minlength_value, minlength=minlength_value,
maxlength=maxlength_value, maxlength=maxlength_value,
binary_count=binary_count, binary_output=binary_output)
output_type=output_type)
else: else:
weights = validate_dense_weights(values, weights)
c_ind, c_val, c_shape = gen_count_ops.dense_count_sparse_output( c_ind, c_val, c_shape = gen_count_ops.dense_count_sparse_output(
values, values,
weights=weights, weights=weights,
minlength=minlength_value, minlength=minlength_value,
maxlength=maxlength_value, maxlength=maxlength_value,
binary_count=binary_count, binary_output=binary_output)
output_type=output_type)
return sparse_tensor.SparseTensor(c_ind, c_val, c_shape) return sparse_tensor.SparseTensor(c_ind, c_val, c_shape)
def validate_dense_weights(values, weights):
"""Validates the passed weight tensor or creates an empty one."""
if weights is None:
return array_ops.constant([], dtype=values.dtype)
if not isinstance(weights, ops.Tensor):
raise ValueError(
"`weights` must be a tf.Tensor if `values` is a tf.Tensor.")
return weights
def validate_sparse_weights(values, weights):
"""Validates the passed weight tensor or creates an empty one."""
if weights is None:
return array_ops.constant([], dtype=values.values.dtype)
if not isinstance(weights, sparse_tensor.SparseTensor):
raise ValueError(
"`weights` must be a SparseTensor if `values` is a SparseTensor.")
checks = []
if weights.dense_shape is not values.dense_shape:
checks.append(
check_ops.assert_equal(
weights.dense_shape,
values.dense_shape,
message="'weights' and 'values' must have the same dense shape."))
if weights.indices is not values.indices:
checks.append(
check_ops.assert_equal(
weights.indices,
values.indices,
message="'weights' and 'values' must have the same indices.")
)
if checks:
with ops.control_dependencies(checks):
weights = array_ops.identity(weights.values)
else:
weights = weights.values
return weights
def validate_ragged_weights(values, weights):
"""Validates the passed weight tensor or creates an empty one."""
if weights is None:
return array_ops.constant([], dtype=values.values.dtype)
if not isinstance(weights, ragged_tensor.RaggedTensor):
raise ValueError(
"`weights` must be a RaggedTensor if `values` is a RaggedTensor.")
checks = []
if weights.row_splits is not values.row_splits:
checks.append(
check_ops.assert_equal(
weights.row_splits,
values.row_splits,
message="'weights' and 'values' must have the same row splits."))
if checks:
with ops.control_dependencies(checks):
weights = array_ops.identity(weights.values)
else:
weights = weights.values
return weights

View File

@ -21,6 +21,8 @@ from __future__ import print_function
from absl.testing import parameterized from absl.testing import parameterized
import numpy as np import numpy as np
from tensorflow.python.eager import context
from tensorflow.python.framework import errors
from tensorflow.python.ops import bincount from tensorflow.python.ops import bincount
from tensorflow.python.ops import sparse_ops from tensorflow.python.ops import sparse_ops
from tensorflow.python.ops.ragged import ragged_factory_ops from tensorflow.python.ops.ragged import ragged_factory_ops
@ -65,7 +67,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_indices": [[0, 1], [0, 2], [0, 3], [1, 4], [1, 5]], "expected_indices": [[0, 1], [0, 2], [0, 3], [1, 4], [1, 5]],
"expected_values": [1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1],
"expected_shape": [2, 6], "expected_shape": [2, 6],
"binary_count": True, "binary_output": True,
}, { }, {
"testcase_name": "_maxlength_binary", "testcase_name": "_maxlength_binary",
"x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32), "x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32),
@ -73,7 +75,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_indices": [[0, 1], [0, 2], [0, 3], [1, 0], [1, 4]], "expected_indices": [[0, 1], [0, 2], [0, 3], [1, 0], [1, 4]],
"expected_values": [1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1],
"expected_shape": [2, 7], "expected_shape": [2, 7],
"binary_count": True, "binary_output": True,
}, { }, {
"testcase_name": "_minlength_binary", "testcase_name": "_minlength_binary",
"x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32), "x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32),
@ -82,7 +84,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
[1, 7]], [1, 7]],
"expected_values": [1, 1, 1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1, 1, 1],
"expected_shape": [2, 9], "expected_shape": [2, 9],
"binary_count": True, "binary_output": True,
}, { }, {
"testcase_name": "_minlength_larger_values_binary", "testcase_name": "_minlength_larger_values_binary",
"x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32), "x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32),
@ -91,40 +93,40 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
[1, 7]], [1, 7]],
"expected_values": [1, 1, 1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1, 1, 1],
"expected_shape": [2, 8], "expected_shape": [2, 8],
"binary_count": True, "binary_output": True,
}, { }, {
"testcase_name": "_no_maxlength_weights", "testcase_name": "_no_maxlength_weights",
"x": np.array([[3, 2, 1], [5, 4, 4]], dtype=np.int32), "x": np.array([[3, 2, 1], [5, 4, 4]], dtype=np.int32),
"expected_indices": [[0, 1], [0, 2], [0, 3], [1, 4], [1, 5]], "expected_indices": [[0, 1], [0, 2], [0, 3], [1, 4], [1, 5]],
"expected_values": [1, 2, 3, 8, 5], "expected_values": [2, 1, 0.5, 9, 3],
"expected_shape": [2, 6], "expected_shape": [2, 6],
"weights": [0.5, 1, 2, 3, 4, 5] "weights": [[0.5, 1, 2], [3, 4, 5]]
}, { }, {
"testcase_name": "_maxlength_weights", "testcase_name": "_maxlength_weights",
"x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32), "x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32),
"maxlength": 7, "maxlength": 7,
"expected_indices": [[0, 1], [0, 2], [0, 3], [1, 0], [1, 4]], "expected_indices": [[0, 1], [0, 2], [0, 3], [1, 0], [1, 4]],
"expected_values": [1, 2, 3, 0.5, 8], "expected_values": [2, 1, 0.5, 3, 9],
"expected_shape": [2, 7], "expected_shape": [2, 7],
"weights": [0.5, 1, 2, 3, 4, 5, 6] "weights": [[0.5, 1, 2, 11], [7, 3, 4, 5]]
}, { }, {
"testcase_name": "_minlength_weights", "testcase_name": "_minlength_weights",
"x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32), "x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32),
"minlength": 9, "minlength": 9,
"expected_indices": [[0, 1], [0, 2], [0, 3], [0, 7], [1, 0], [1, 4], "expected_indices": [[0, 1], [0, 2], [0, 3], [0, 7], [1, 0], [1, 4],
[1, 7]], [1, 7]],
"expected_values": [1, 2, 3, 7, 0.5, 8, 7], "expected_values": [2, 1, 0.5, 3, 5, 13, 4],
"expected_shape": [2, 9], "expected_shape": [2, 9],
"weights": [0.5, 1, 2, 3, 4, 5, 6, 7, 8] "weights": [[0.5, 1, 2, 3], [4, 5, 6, 7]]
}, { }, {
"testcase_name": "_minlength_larger_values_weights", "testcase_name": "_minlength_larger_values_weights",
"x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32), "x": np.array([[3, 2, 1, 7], [7, 0, 4, 4]], dtype=np.int32),
"minlength": 3, "minlength": 3,
"expected_indices": [[0, 1], [0, 2], [0, 3], [0, 7], [1, 0], [1, 4], "expected_indices": [[0, 1], [0, 2], [0, 3], [0, 7], [1, 0], [1, 4],
[1, 7]], [1, 7]],
"expected_values": [1, 2, 3, 7, 0.5, 8, 7], "expected_values": [2, 1, 0.5, 3, 5, 13, 4],
"expected_shape": [2, 8], "expected_shape": [2, 8],
"weights": [0.5, 1, 2, 3, 4, 5, 6, 7, 8] "weights": [[0.5, 1, 2, 3], [4, 5, 6, 7]]
}, { }, {
"testcase_name": "_1d", "testcase_name": "_1d",
"x": np.array([3, 2, 1, 1], dtype=np.int32), "x": np.array([3, 2, 1, 1], dtype=np.int32),
@ -146,7 +148,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
expected_shape, expected_shape,
minlength=None, minlength=None,
maxlength=None, maxlength=None,
binary_count=False, binary_output=False,
weights=None, weights=None,
axis=-1): axis=-1):
y = bincount.sparse_bincount( y = bincount.sparse_bincount(
@ -154,7 +156,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
weights=weights, weights=weights,
minlength=minlength, minlength=minlength,
maxlength=maxlength, maxlength=maxlength,
binary_count=binary_count, binary_output=binary_output,
axis=axis) axis=axis)
self.assertAllEqual(expected_indices, y.indices) self.assertAllEqual(expected_indices, y.indices)
self.assertAllEqual(expected_values, y.values) self.assertAllEqual(expected_values, y.values)
@ -216,7 +218,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_indices": [[0, 1], [0, 3], [2, 4], [2, 5]], "expected_indices": [[0, 1], [0, 3], [2, 4], [2, 5]],
"expected_values": [1, 1, 1, 1], "expected_values": [1, 1, 1, 1],
"expected_shape": [3, 6], "expected_shape": [3, 6],
"binary_count": "binary_output":
True, True,
}, },
{ {
@ -230,7 +232,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_shape": [3, 7], "expected_shape": [3, 7],
"maxlength": "maxlength":
7, 7,
"binary_count": "binary_output":
True, True,
}, },
{ {
@ -244,7 +246,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_shape": [3, 9], "expected_shape": [3, 9],
"minlength": "minlength":
9, 9,
"binary_count": "binary_output":
True, True,
}, },
{ {
@ -258,7 +260,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_shape": [3, 8], "expected_shape": [3, 8],
"minlength": "minlength":
3, 3,
"binary_count": "binary_output":
True, True,
}, },
{ {
@ -268,9 +270,10 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]],
dtype=np.int32), dtype=np.int32),
"expected_indices": [[0, 1], [0, 3], [2, 4], [2, 5]], "expected_indices": [[0, 1], [0, 3], [2, 4], [2, 5]],
"expected_values": [1, 3, 8, 5], "expected_values": [2, 6, 7, 10],
"expected_shape": [3, 6], "expected_shape": [3, 6],
"weights": [0.5, 1, 2, 3, 4, 5] "weights":
np.array([[6, 0, 2, 0], [0, 0, 0, 0], [10, 0, 3.5, 3.5]]),
}, },
{ {
"testcase_name": "testcase_name":
@ -279,11 +282,12 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
np.array([[3, 0, 1, 0], [0, 0, 7, 0], [5, 0, 4, 4]], np.array([[3, 0, 1, 0], [0, 0, 7, 0], [5, 0, 4, 4]],
dtype=np.int32), dtype=np.int32),
"expected_indices": [[0, 1], [0, 3], [2, 4], [2, 5]], "expected_indices": [[0, 1], [0, 3], [2, 4], [2, 5]],
"expected_values": [1, 3, 8, 5], "expected_values": [2, 6, 7, 10],
"expected_shape": [3, 7], "expected_shape": [3, 7],
"maxlength": "maxlength":
7, 7,
"weights": [0.5, 1, 2, 3, 4, 5, 6] "weights":
np.array([[6, 0, 2, 0], [0, 0, 14, 0], [10, 0, 3.5, 3.5]]),
}, },
{ {
"testcase_name": "testcase_name":
@ -292,11 +296,12 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
np.array([[3, 0, 1, 0], [7, 0, 0, 0], [5, 0, 4, 4]], np.array([[3, 0, 1, 0], [7, 0, 0, 0], [5, 0, 4, 4]],
dtype=np.int32), dtype=np.int32),
"expected_indices": [[0, 1], [0, 3], [1, 7], [2, 4], [2, 5]], "expected_indices": [[0, 1], [0, 3], [1, 7], [2, 4], [2, 5]],
"expected_values": [1, 3, 7, 8, 5], "expected_values": [2, 6, 14, 6.5, 10],
"expected_shape": [3, 9], "expected_shape": [3, 9],
"minlength": "minlength":
9, 9,
"weights": [0.5, 1, 2, 3, 4, 5, 6, 7, 8] "weights":
np.array([[6, 0, 2, 0], [14, 0, 0, 0], [10, 0, 3, 3.5]]),
}, },
{ {
"testcase_name": "testcase_name":
@ -305,11 +310,12 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
np.array([[3, 0, 1, 0], [7, 0, 0, 0], [5, 0, 4, 4]], np.array([[3, 0, 1, 0], [7, 0, 0, 0], [5, 0, 4, 4]],
dtype=np.int32), dtype=np.int32),
"expected_indices": [[0, 1], [0, 3], [1, 7], [2, 4], [2, 5]], "expected_indices": [[0, 1], [0, 3], [1, 7], [2, 4], [2, 5]],
"expected_values": [1, 3, 7, 8, 5], "expected_values": [2, 6, 14, 6.5, 10],
"expected_shape": [3, 8], "expected_shape": [3, 8],
"minlength": "minlength":
3, 3,
"weights": [0.5, 1, 2, 3, 4, 5, 6, 7, 8] "weights":
np.array([[6, 0, 2, 0], [14, 0, 0, 0], [10, 0, 3, 3.5]]),
}, },
{ {
"testcase_name": "_1d", "testcase_name": "_1d",
@ -338,16 +344,17 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
expected_shape, expected_shape,
maxlength=None, maxlength=None,
minlength=None, minlength=None,
binary_count=False, binary_output=False,
weights=None, weights=None,
axis=-1): axis=-1):
x_sparse = sparse_ops.from_dense(x) x_sparse = sparse_ops.from_dense(x)
w_sparse = sparse_ops.from_dense(weights) if weights is not None else None
y = bincount.sparse_bincount( y = bincount.sparse_bincount(
x_sparse, x_sparse,
weights=weights, weights=w_sparse,
minlength=minlength, minlength=minlength,
maxlength=maxlength, maxlength=maxlength,
binary_count=binary_count, binary_output=binary_output,
axis=axis) axis=axis)
self.assertAllEqual(expected_indices, y.indices) self.assertAllEqual(expected_indices, y.indices)
self.assertAllEqual(expected_values, y.values) self.assertAllEqual(expected_values, y.values)
@ -393,7 +400,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]], "expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]],
"expected_values": [1, 1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1, 1],
"expected_shape": [5, 6], "expected_shape": [5, 6],
"binary_count": True, "binary_output": True,
}, },
{ {
"testcase_name": "_maxlength_binary", "testcase_name": "_maxlength_binary",
@ -402,7 +409,7 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]], "expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]],
"expected_values": [1, 1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1, 1],
"expected_shape": [5, 7], "expected_shape": [5, 7],
"binary_count": True, "binary_output": True,
}, },
{ {
"testcase_name": "_minlength_binary", "testcase_name": "_minlength_binary",
@ -412,13 +419,13 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
[4, 5]], [4, 5]],
"expected_values": [1, 1, 1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1, 1, 1],
"expected_shape": [5, 9], "expected_shape": [5, 9],
"binary_count": True, "binary_output": True,
}, },
{ {
"testcase_name": "_minlength_larger_values_binary", "testcase_name": "_minlength_larger_values_binary",
"x": [[], [], [3, 0, 1], [7], [5, 0, 4, 4]], "x": [[], [], [3, 0, 1], [7], [5, 0, 4, 4]],
"minlength": 3, "minlength": 3,
"binary_count": True, "binary_output": True,
"expected_indices": [[2, 0], [2, 1], [2, 3], [3, 7], [4, 0], [4, 4], "expected_indices": [[2, 0], [2, 1], [2, 3], [3, 7], [4, 0], [4, 4],
[4, 5]], [4, 5]],
"expected_values": [1, 1, 1, 1, 1, 1, 1], "expected_values": [1, 1, 1, 1, 1, 1, 1],
@ -428,18 +435,18 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"testcase_name": "_no_maxlength_weights", "testcase_name": "_no_maxlength_weights",
"x": [[], [], [3, 0, 1], [], [5, 0, 4, 4]], "x": [[], [], [3, 0, 1], [], [5, 0, 4, 4]],
"expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]], "expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]],
"expected_values": [0.5, 1, 3, 0.5, 8, 5], "expected_values": [0.5, 2, 6, 0.25, 8, 10],
"expected_shape": [5, 6], "expected_shape": [5, 6],
"weights": [0.5, 1, 2, 3, 4, 5] "weights": [[], [], [6, 0.5, 2], [], [10, 0.25, 5, 3]],
}, },
{ {
"testcase_name": "_maxlength_weights", "testcase_name": "_maxlength_weights",
"x": [[], [], [3, 0, 1], [7], [5, 0, 4, 4]], "x": [[], [], [3, 0, 1], [7], [5, 0, 4, 4]],
"maxlength": 7, "maxlength": 7,
"expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]], "expected_indices": [[2, 0], [2, 1], [2, 3], [4, 0], [4, 4], [4, 5]],
"expected_values": [0.5, 1, 3, 0.5, 8, 5], "expected_values": [0.5, 2, 6, 0.25, 8, 10],
"expected_shape": [5, 7], "expected_shape": [5, 7],
"weights": [0.5, 1, 2, 3, 4, 5, 6] "weights": [[], [], [6, 0.5, 2], [14], [10, 0.25, 5, 3]],
}, },
{ {
"testcase_name": "_minlength_weights", "testcase_name": "_minlength_weights",
@ -447,9 +454,9 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"minlength": 9, "minlength": 9,
"expected_indices": [[2, 0], [2, 1], [2, 3], [3, 7], [4, 0], [4, 4], "expected_indices": [[2, 0], [2, 1], [2, 3], [3, 7], [4, 0], [4, 4],
[4, 5]], [4, 5]],
"expected_values": [0.5, 1, 3, 7, 0.5, 8, 5], "expected_values": [0.5, 2, 6, 14, 0.25, 8, 10],
"expected_shape": [5, 9], "expected_shape": [5, 9],
"weights": [0.5, 1, 2, 3, 4, 5, 6, 7, 8] "weights": [[], [], [6, 0.5, 2], [14], [10, 0.25, 5, 3]],
}, },
{ {
"testcase_name": "_minlength_larger_values_weights", "testcase_name": "_minlength_larger_values_weights",
@ -457,9 +464,9 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
"minlength": 3, "minlength": 3,
"expected_indices": [[2, 0], [2, 1], [2, 3], [3, 7], [4, 0], [4, 4], "expected_indices": [[2, 0], [2, 1], [2, 3], [3, 7], [4, 0], [4, 4],
[4, 5]], [4, 5]],
"expected_values": [0.5, 1, 3, 7, 0.5, 8, 5], "expected_values": [0.5, 2, 6, 14, 0.25, 8, 10],
"expected_shape": [5, 8], "expected_shape": [5, 8],
"weights": [0.5, 1, 2, 3, 4, 5, 6, 7, 8] "weights": [[], [], [6, 0.5, 2], [14], [10, 0.25, 5, 3]],
}, },
{ {
"testcase_name": "_1d", "testcase_name": "_1d",
@ -484,21 +491,114 @@ class TestSparseCount(test.TestCase, parameterized.TestCase):
expected_shape, expected_shape,
maxlength=None, maxlength=None,
minlength=None, minlength=None,
binary_count=False, binary_output=False,
weights=None, weights=None,
axis=-1): axis=-1):
x_ragged = ragged_factory_ops.constant(x) x_ragged = ragged_factory_ops.constant(x)
w = ragged_factory_ops.constant(weights) if weights is not None else None
y = bincount.sparse_bincount( y = bincount.sparse_bincount(
x_ragged, x_ragged,
weights=weights, weights=w,
minlength=minlength, minlength=minlength,
maxlength=maxlength, maxlength=maxlength,
binary_count=binary_count, binary_output=binary_output,
axis=axis) axis=axis)
self.assertAllEqual(expected_indices, y.indices) self.assertAllEqual(expected_indices, y.indices)
self.assertAllEqual(expected_values, y.values) self.assertAllEqual(expected_values, y.values)
self.assertAllEqual(expected_shape, y.dense_shape) self.assertAllEqual(expected_shape, y.dense_shape)
class TestSparseCountFailureModes(test.TestCase):
def test_dense_input_sparse_weights_fails(self):
x = np.array([[3, 2, 1], [5, 4, 4]], dtype=np.int32)
weights = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
with self.assertRaisesRegexp(ValueError, "must be a tf.Tensor"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_dense_input_ragged_weights_fails(self):
x = np.array([[3, 2, 1], [5, 4, 4]], dtype=np.int32)
weights = ragged_factory_ops.constant([[6, 0.5, 2], [14], [10, 0.25, 5, 3]])
with self.assertRaisesRegexp(ValueError, "must be a tf.Tensor"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_dense_input_wrong_shape_fails(self):
x = np.array([[3, 2, 1], [5, 4, 4]], dtype=np.int32)
weights = np.array([[3, 2], [5, 4], [4, 3]])
# Note: Eager mode and graph mode throw different errors here. Graph mode
# will fail with a ValueError from the shape checking logic, while Eager
# will fail with an InvalidArgumentError from the kernel itself.
if context.executing_eagerly():
with self.assertRaisesRegexp(errors.InvalidArgumentError,
"must have the same shape"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
else:
with self.assertRaisesRegexp(ValueError, "both shapes must be equal"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_sparse_input_dense_weights_fails(self):
x = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
weights = np.array([[3, 2, 1], [5, 4, 4]], dtype=np.int32)
with self.assertRaisesRegexp(ValueError, "must be a SparseTensor"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_sparse_input_ragged_weights_fails(self):
x = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
weights = ragged_factory_ops.constant([[6, 0.5, 2], [14], [10, 0.25, 5, 3]])
with self.assertRaisesRegexp(ValueError, "must be a SparseTensor"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_sparse_input_wrong_indices_fails(self):
x = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
weights = sparse_ops.from_dense(
np.array([[3, 1, 0, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
with self.assertRaisesRegexp(errors.InvalidArgumentError,
"must have the same indices"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_sparse_input_too_many_indices_fails(self):
x = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
weights = sparse_ops.from_dense(
np.array([[3, 1, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
with self.assertRaisesRegexp(errors.InvalidArgumentError,
"Incompatible shapes"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_sparse_input_wrong_shape_fails(self):
x = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
weights = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4], [0, 0, 0, 0]],
dtype=np.int32))
with self.assertRaisesRegexp(errors.InvalidArgumentError,
"must have the same dense shape"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_ragged_input_dense_weights_fails(self):
x = ragged_factory_ops.constant([[6, 1, 2], [14], [10, 1, 5, 3]])
weights = np.array([[3, 2, 1], [5, 4, 4]], dtype=np.int32)
with self.assertRaisesRegexp(ValueError, "must be a RaggedTensor"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_ragged_input_sparse_weights_fails(self):
x = ragged_factory_ops.constant([[6, 1, 2], [14], [10, 1, 5, 3]])
weights = sparse_ops.from_dense(
np.array([[3, 0, 1, 0], [0, 0, 0, 0], [5, 0, 4, 4]], dtype=np.int32))
with self.assertRaisesRegexp(ValueError, "must be a RaggedTensor"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
def test_ragged_input_different_shape_fails(self):
x = ragged_factory_ops.constant([[6, 1, 2], [14], [10, 1, 5, 3]])
weights = ragged_factory_ops.constant([[6, 0.5, 2], [], [10, 0.25, 5, 3]])
with self.assertRaisesRegexp(errors.InvalidArgumentError,
"must have the same row splits"):
self.evaluate(bincount.sparse_bincount(x, weights=weights, axis=-1))
if __name__ == "__main__": if __name__ == "__main__":
test.main() test.main()

View File

@ -1078,7 +1078,7 @@ tf_module {
} }
member_method { member_method {
name: "DenseCountSparseOutput" name: "DenseCountSparseOutput"
argspec: "args=[\'values\', \'weights\', \'binary_count\', \'output_type\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], " argspec: "args=[\'values\', \'weights\', \'binary_output\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], "
} }
member_method { member_method {
name: "DenseToCSRSparseMatrix" name: "DenseToCSRSparseMatrix"
@ -3074,7 +3074,7 @@ tf_module {
} }
member_method { member_method {
name: "RaggedCountSparseOutput" name: "RaggedCountSparseOutput"
argspec: "args=[\'splits\', \'values\', \'weights\', \'binary_count\', \'output_type\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], " argspec: "args=[\'splits\', \'values\', \'weights\', \'binary_output\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], "
} }
member_method { member_method {
name: "RaggedCross" name: "RaggedCross"
@ -4094,7 +4094,7 @@ tf_module {
} }
member_method { member_method {
name: "SparseCountSparseOutput" name: "SparseCountSparseOutput"
argspec: "args=[\'indices\', \'values\', \'dense_shape\', \'weights\', \'binary_count\', \'output_type\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], " argspec: "args=[\'indices\', \'values\', \'dense_shape\', \'weights\', \'binary_output\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], "
} }
member_method { member_method {
name: "SparseCross" name: "SparseCross"

View File

@ -14,7 +14,7 @@ tf_module {
} }
member_method { member_method {
name: "bincount" name: "bincount"
argspec: "args=[\'values\', \'weights\', \'axis\', \'minlength\', \'maxlength\', \'binary_count\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \'0\', \'None\', \'None\', \'False\', \'None\'], " argspec: "args=[\'values\', \'weights\', \'axis\', \'minlength\', \'maxlength\', \'binary_output\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \'0\', \'None\', \'None\', \'False\', \'None\'], "
} }
member_method { member_method {
name: "concat" name: "concat"

View File

@ -1078,7 +1078,7 @@ tf_module {
} }
member_method { member_method {
name: "DenseCountSparseOutput" name: "DenseCountSparseOutput"
argspec: "args=[\'values\', \'weights\', \'binary_count\', \'output_type\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], " argspec: "args=[\'values\', \'weights\', \'binary_output\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], "
} }
member_method { member_method {
name: "DenseToCSRSparseMatrix" name: "DenseToCSRSparseMatrix"
@ -3074,7 +3074,7 @@ tf_module {
} }
member_method { member_method {
name: "RaggedCountSparseOutput" name: "RaggedCountSparseOutput"
argspec: "args=[\'splits\', \'values\', \'weights\', \'binary_count\', \'output_type\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], " argspec: "args=[\'splits\', \'values\', \'weights\', \'binary_output\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], "
} }
member_method { member_method {
name: "RaggedCross" name: "RaggedCross"
@ -4094,7 +4094,7 @@ tf_module {
} }
member_method { member_method {
name: "SparseCountSparseOutput" name: "SparseCountSparseOutput"
argspec: "args=[\'indices\', \'values\', \'dense_shape\', \'weights\', \'binary_count\', \'output_type\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], " argspec: "args=[\'indices\', \'values\', \'dense_shape\', \'weights\', \'binary_output\', \'minlength\', \'maxlength\', \'name\'], varargs=None, keywords=None, defaults=[\'-1\', \'-1\', \'None\'], "
} }
member_method { member_method {
name: "SparseCross" name: "SparseCross"

View File

@ -10,7 +10,7 @@ tf_module {
} }
member_method { member_method {
name: "bincount" name: "bincount"
argspec: "args=[\'values\', \'weights\', \'axis\', \'minlength\', \'maxlength\', \'binary_count\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \'0\', \'None\', \'None\', \'False\', \'None\'], " argspec: "args=[\'values\', \'weights\', \'axis\', \'minlength\', \'maxlength\', \'binary_output\', \'name\'], varargs=None, keywords=None, defaults=[\'None\', \'0\', \'None\', \'None\', \'False\', \'None\'], "
} }
member_method { member_method {
name: "concat" name: "concat"