This change moves //tensorflow/contrib/lite to //tensorflow/lite in preparation for TensorFlow 2.0's deprecation of contrib/. If you refer to TF Lite build targets or headers, you will need to update them manually. If you use TF Lite from the TensorFlow python package, "tf.contrib.lite" now points to "tf.lite". Please update your imports as soon as possible. For more details, see https://groups.google.com/a/tensorflow.org/forum/#!topic/tflite/iIIXOTOFvwQ @angersson and @aselle are conducting this migration. Please contact them if you have any further questions. PiperOrigin-RevId: 219536476
433 lines
16 KiB
C++
433 lines
16 KiB
C++
/* Copyright 2018 The TensorFlow Authors. All Rights Reserved.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
==============================================================================*/
|
|
|
|
// Copied from tensorflow/core/util/ctc/ctc_beam_search.h
|
|
// TODO(b/111524997): Remove this file.
|
|
#ifndef TENSORFLOW_LITE_EXPERIMENTAL_KERNELS_CTC_BEAM_SEARCH_H_
|
|
#define TENSORFLOW_LITE_EXPERIMENTAL_KERNELS_CTC_BEAM_SEARCH_H_
|
|
|
|
#include <algorithm>
|
|
#include <cmath>
|
|
#include <limits>
|
|
#include <memory>
|
|
#include <vector>
|
|
|
|
#include "third_party/eigen3/Eigen/Core"
|
|
#include "tensorflow/lite/experimental/kernels/ctc_beam_entry.h"
|
|
#include "tensorflow/lite/experimental/kernels/ctc_beam_scorer.h"
|
|
#include "tensorflow/lite/experimental/kernels/ctc_decoder.h"
|
|
#include "tensorflow/lite/experimental/kernels/ctc_loss_util.h"
|
|
#include "tensorflow/lite/experimental/kernels/top_n.h"
|
|
#include "tensorflow/lite/kernels/internal/compatibility.h"
|
|
|
|
namespace tflite {
|
|
namespace experimental {
|
|
namespace ctc {
|
|
|
|
template <typename CTCBeamState = ctc_beam_search::EmptyBeamState,
|
|
typename CTCBeamComparer =
|
|
ctc_beam_search::BeamComparer<CTCBeamState>>
|
|
class CTCBeamSearchDecoder : public CTCDecoder {
|
|
// Beam Search
|
|
//
|
|
// Example (GravesTh Fig. 7.5):
|
|
// a -
|
|
// P = [ 0.3 0.7 ] t = 0
|
|
// [ 0.4 0.6 ] t = 1
|
|
//
|
|
// Then P(l = -) = P(--) = 0.7 * 0.6 = 0.42
|
|
// P(l = a) = P(a-) + P(aa) + P(-a) = 0.3*0.4 + ... = 0.58
|
|
//
|
|
// In this case, Best Path decoding is suboptimal.
|
|
//
|
|
// For Beam Search, we use the following main recurrence relations:
|
|
//
|
|
// Relation 1:
|
|
// ---------------------------------------------------------- Eq. 1
|
|
// P(l=abcd @ t=7) = P(l=abc @ t=6) * P(d @ 7)
|
|
// + P(l=abcd @ t=6) * (P(d @ 7) + P(- @ 7))
|
|
// where P(l=? @ t=7), ? = a, ab, abc, abcd are all stored and
|
|
// updated recursively in the beam entry.
|
|
//
|
|
// Relation 2:
|
|
// ---------------------------------------------------------- Eq. 2
|
|
// P(l=abc? @ t=3) = P(l=abc @ t=2) * P(? @ 3)
|
|
// for ? in a, b, d, ..., (not including c or the blank index),
|
|
// and the recurrence starts from the beam entry for P(l=abc @ t=2).
|
|
//
|
|
// For this case, the length of the new sequence equals t+1 (t
|
|
// starts at 0). This special case can be calculated as:
|
|
// P(l=abc? @ t=3) = P(a @ 0)*P(b @ 1)*P(c @ 2)*P(? @ 3)
|
|
// but we calculate it recursively for speed purposes.
|
|
typedef ctc_beam_search::BeamEntry<CTCBeamState> BeamEntry;
|
|
typedef ctc_beam_search::BeamRoot<CTCBeamState> BeamRoot;
|
|
typedef ctc_beam_search::BeamProbability BeamProbability;
|
|
|
|
public:
|
|
typedef BaseBeamScorer<CTCBeamState> DefaultBeamScorer;
|
|
|
|
// The beam search decoder is constructed specifying the beam_width (number of
|
|
// candidates to keep at each decoding timestep) and a beam scorer (used for
|
|
// custom scoring, for example enabling the use of a language model).
|
|
// The ownership of the scorer remains with the caller. The default
|
|
// implementation, CTCBeamSearchDecoder<>::DefaultBeamScorer, generates the
|
|
// standard beam search.
|
|
CTCBeamSearchDecoder(int num_classes, int beam_width,
|
|
BaseBeamScorer<CTCBeamState>* scorer, int batch_size = 1,
|
|
bool merge_repeated = false)
|
|
: CTCDecoder(num_classes, batch_size, merge_repeated),
|
|
beam_width_(beam_width),
|
|
leaves_(beam_width),
|
|
beam_scorer_(scorer) {
|
|
Reset();
|
|
}
|
|
|
|
~CTCBeamSearchDecoder() override {}
|
|
|
|
// Run the hibernating beam search algorithm on the given input.
|
|
bool Decode(const CTCDecoder::SequenceLength& seq_len,
|
|
const std::vector<CTCDecoder::Input>& input,
|
|
std::vector<CTCDecoder::Output>* output,
|
|
CTCDecoder::ScoreOutput* scores) override;
|
|
|
|
// Calculate the next step of the beam search and update the internal state.
|
|
template <typename Vector>
|
|
void Step(const Vector& log_input_t);
|
|
|
|
template <typename Vector>
|
|
float GetTopK(const int K, const Vector& input,
|
|
std::vector<float>* top_k_logits,
|
|
std::vector<int>* top_k_indices);
|
|
|
|
// Retrieve the beam scorer instance used during decoding.
|
|
BaseBeamScorer<CTCBeamState>* GetBeamScorer() const { return beam_scorer_; }
|
|
|
|
// Set label selection parameters for faster decoding.
|
|
// See comments for label_selection_size_ and label_selection_margin_.
|
|
void SetLabelSelectionParameters(int label_selection_size,
|
|
float label_selection_margin) {
|
|
label_selection_size_ = label_selection_size;
|
|
label_selection_margin_ = label_selection_margin;
|
|
}
|
|
|
|
// Reset the beam search
|
|
void Reset();
|
|
|
|
// Extract the top n paths at current time step
|
|
bool TopPaths(int n, std::vector<std::vector<int>>* paths,
|
|
std::vector<float>* log_probs, bool merge_repeated) const;
|
|
|
|
private:
|
|
int beam_width_;
|
|
|
|
// Label selection is designed to avoid possibly very expensive scorer calls,
|
|
// by pruning the hypotheses based on the input alone.
|
|
// Label selection size controls how many items in each beam are passed
|
|
// through to the beam scorer. Only items with top N input scores are
|
|
// considered.
|
|
// Label selection margin controls the difference between minimal input score
|
|
// (versus the best scoring label) for an item to be passed to the beam
|
|
// scorer. This margin is expressed in terms of log-probability.
|
|
// Default is to do no label selection.
|
|
// For more detail: https://research.google.com/pubs/pub44823.html
|
|
int label_selection_size_ = 0; // zero means unlimited
|
|
float label_selection_margin_ = -1; // -1 means unlimited.
|
|
|
|
gtl::TopN<BeamEntry*, CTCBeamComparer> leaves_;
|
|
std::unique_ptr<BeamRoot> beam_root_;
|
|
BaseBeamScorer<CTCBeamState>* beam_scorer_;
|
|
|
|
CTCBeamSearchDecoder(const CTCBeamSearchDecoder&) = delete;
|
|
void operator=(const CTCBeamSearchDecoder&) = delete;
|
|
};
|
|
|
|
template <typename CTCBeamState, typename CTCBeamComparer>
|
|
bool CTCBeamSearchDecoder<CTCBeamState, CTCBeamComparer>::Decode(
|
|
const CTCDecoder::SequenceLength& seq_len,
|
|
const std::vector<CTCDecoder::Input>& input,
|
|
std::vector<CTCDecoder::Output>* output, ScoreOutput* scores) {
|
|
// Storage for top paths.
|
|
std::vector<std::vector<int>> beams;
|
|
std::vector<float> beam_log_probabilities;
|
|
int top_n = output->size();
|
|
if (std::any_of(output->begin(), output->end(),
|
|
[this](const CTCDecoder::Output& output) -> bool {
|
|
return output.size() < this->batch_size_;
|
|
})) {
|
|
return false;
|
|
}
|
|
if (scores->rows() < batch_size_ || scores->cols() < top_n) {
|
|
return false;
|
|
}
|
|
|
|
for (int b = 0; b < batch_size_; ++b) {
|
|
int seq_len_b = seq_len[b];
|
|
Reset();
|
|
|
|
for (int t = 0; t < seq_len_b; ++t) {
|
|
// Pass log-probabilities for this example + time.
|
|
Step(input[t].row(b));
|
|
} // for (int t...
|
|
|
|
// O(n * log(n))
|
|
std::unique_ptr<std::vector<BeamEntry*>> branches(leaves_.Extract());
|
|
leaves_.Reset();
|
|
for (int i = 0; i < branches->size(); ++i) {
|
|
BeamEntry* entry = (*branches)[i];
|
|
beam_scorer_->ExpandStateEnd(&entry->state);
|
|
entry->newp.total +=
|
|
beam_scorer_->GetStateEndExpansionScore(entry->state);
|
|
leaves_.push(entry);
|
|
}
|
|
|
|
bool status =
|
|
TopPaths(top_n, &beams, &beam_log_probabilities, merge_repeated_);
|
|
if (!status) {
|
|
return status;
|
|
}
|
|
|
|
TFLITE_DCHECK_EQ(top_n, beam_log_probabilities.size());
|
|
TFLITE_DCHECK_EQ(beams.size(), beam_log_probabilities.size());
|
|
|
|
for (int i = 0; i < top_n; ++i) {
|
|
// Copy output to the correct beam + batch
|
|
(*output)[i][b].swap(beams[i]);
|
|
(*scores)(b, i) = -beam_log_probabilities[i];
|
|
}
|
|
} // for (int b...
|
|
return true;
|
|
}
|
|
|
|
template <typename CTCBeamState, typename CTCBeamComparer>
|
|
template <typename Vector>
|
|
float CTCBeamSearchDecoder<CTCBeamState, CTCBeamComparer>::GetTopK(
|
|
const int K, const Vector& input, std::vector<float>* top_k_logits,
|
|
std::vector<int>* top_k_indices) {
|
|
// Find Top K choices, complexity nk in worst case. The array input is read
|
|
// just once.
|
|
TFLITE_DCHECK_EQ(num_classes_, input.size());
|
|
top_k_logits->clear();
|
|
top_k_indices->clear();
|
|
top_k_logits->resize(K, -INFINITY);
|
|
top_k_indices->resize(K, -1);
|
|
for (int j = 0; j < num_classes_ - 1; ++j) {
|
|
const float logit = input(j);
|
|
if (logit > (*top_k_logits)[K - 1]) {
|
|
int k = K - 1;
|
|
while (k > 0 && logit > (*top_k_logits)[k - 1]) {
|
|
(*top_k_logits)[k] = (*top_k_logits)[k - 1];
|
|
(*top_k_indices)[k] = (*top_k_indices)[k - 1];
|
|
k--;
|
|
}
|
|
(*top_k_logits)[k] = logit;
|
|
(*top_k_indices)[k] = j;
|
|
}
|
|
}
|
|
// Return max value which is in 0th index or blank character logit
|
|
return std::max((*top_k_logits)[0], input(num_classes_ - 1));
|
|
}
|
|
|
|
template <typename CTCBeamState, typename CTCBeamComparer>
|
|
template <typename Vector>
|
|
void CTCBeamSearchDecoder<CTCBeamState, CTCBeamComparer>::Step(
|
|
const Vector& raw_input) {
|
|
std::vector<float> top_k_logits;
|
|
std::vector<int> top_k_indices;
|
|
const bool top_k =
|
|
(label_selection_size_ > 0 && label_selection_size_ < raw_input.size());
|
|
// Number of character classes to consider in each step.
|
|
const int max_classes = top_k ? label_selection_size_ : (num_classes_ - 1);
|
|
// Get max coefficient and remove it from raw_input later.
|
|
float max_coeff;
|
|
if (top_k) {
|
|
max_coeff = GetTopK(label_selection_size_, raw_input, &top_k_logits,
|
|
&top_k_indices);
|
|
} else {
|
|
max_coeff = raw_input.maxCoeff();
|
|
}
|
|
|
|
// Get normalization term of softmax: log(sum(exp(logit[j]-max_coeff))).
|
|
float logsumexp = 0.0;
|
|
for (int j = 0; j < raw_input.size(); ++j) {
|
|
logsumexp += Eigen::numext::exp(raw_input(j) - max_coeff);
|
|
}
|
|
logsumexp = Eigen::numext::log(logsumexp);
|
|
// Final normalization offset to get correct log probabilities.
|
|
float norm_offset = max_coeff + logsumexp;
|
|
|
|
const float label_selection_input_min =
|
|
(label_selection_margin_ >= 0) ? (max_coeff - label_selection_margin_)
|
|
: -std::numeric_limits<float>::infinity();
|
|
|
|
// Extract the beams sorted in decreasing new probability
|
|
TFLITE_DCHECK_EQ(num_classes_, raw_input.size());
|
|
|
|
std::unique_ptr<std::vector<BeamEntry*>> branches(leaves_.Extract());
|
|
leaves_.Reset();
|
|
|
|
for (BeamEntry* b : *branches) {
|
|
// P(.. @ t) becomes the new P(.. @ t-1)
|
|
b->oldp = b->newp;
|
|
}
|
|
|
|
for (BeamEntry* b : *branches) {
|
|
if (b->parent != nullptr) { // if not the root
|
|
if (b->parent->Active()) {
|
|
// If last two sequence characters are identical:
|
|
// Plabel(l=acc @ t=6) = (Plabel(l=acc @ t=5)
|
|
// + Pblank(l=ac @ t=5))
|
|
// else:
|
|
// Plabel(l=abc @ t=6) = (Plabel(l=abc @ t=5)
|
|
// + P(l=ab @ t=5))
|
|
float previous = (b->label == b->parent->label) ? b->parent->oldp.blank
|
|
: b->parent->oldp.total;
|
|
b->newp.label =
|
|
LogSumExp(b->newp.label,
|
|
beam_scorer_->GetStateExpansionScore(b->state, previous));
|
|
}
|
|
// Plabel(l=abc @ t=6) *= P(c @ 6)
|
|
b->newp.label += raw_input(b->label) - norm_offset;
|
|
}
|
|
// Pblank(l=abc @ t=6) = P(l=abc @ t=5) * P(- @ 6)
|
|
b->newp.blank = b->oldp.total + raw_input(blank_index_) - norm_offset;
|
|
// P(l=abc @ t=6) = Plabel(l=abc @ t=6) + Pblank(l=abc @ t=6)
|
|
b->newp.total = LogSumExp(b->newp.blank, b->newp.label);
|
|
|
|
// Push the entry back to the top paths list.
|
|
// Note, this will always fill leaves back up in sorted order.
|
|
leaves_.push(b);
|
|
}
|
|
|
|
// we need to resort branches in descending oldp order.
|
|
|
|
// branches is in descending oldp order because it was
|
|
// originally in descending newp order and we copied newp to oldp.
|
|
|
|
// Grow new leaves
|
|
for (BeamEntry* b : *branches) {
|
|
// A new leaf (represented by its BeamProbability) is a candidate
|
|
// iff its total probability is nonzero and either the beam list
|
|
// isn't full, or the lowest probability entry in the beam has a
|
|
// lower probability than the leaf.
|
|
auto is_candidate = [this](const BeamProbability& prob) {
|
|
return (prob.total > kLogZero &&
|
|
(leaves_.size() < beam_width_ ||
|
|
prob.total > leaves_.peek_bottom()->newp.total));
|
|
};
|
|
|
|
if (!is_candidate(b->oldp)) {
|
|
continue;
|
|
}
|
|
|
|
for (int ind = 0; ind < max_classes; ind++) {
|
|
const int label = top_k ? top_k_indices[ind] : ind;
|
|
const float logit = top_k ? top_k_logits[ind] : raw_input(ind);
|
|
// Perform label selection: if input for this label looks very
|
|
// unpromising, never evaluate it with a scorer.
|
|
// We may compare logits instead of log probabilities,
|
|
// since the difference is the same in both cases.
|
|
if (logit < label_selection_input_min) {
|
|
continue;
|
|
}
|
|
BeamEntry& c = b->GetChild(label);
|
|
if (!c.Active()) {
|
|
// Pblank(l=abcd @ t=6) = 0
|
|
c.newp.blank = kLogZero;
|
|
// If new child label is identical to beam label:
|
|
// Plabel(l=abcc @ t=6) = Pblank(l=abc @ t=5) * P(c @ 6)
|
|
// Otherwise:
|
|
// Plabel(l=abcd @ t=6) = P(l=abc @ t=5) * P(d @ 6)
|
|
beam_scorer_->ExpandState(b->state, b->label, &c.state, c.label);
|
|
float previous = (c.label == b->label) ? b->oldp.blank : b->oldp.total;
|
|
c.newp.label = logit - norm_offset +
|
|
beam_scorer_->GetStateExpansionScore(c.state, previous);
|
|
// P(l=abcd @ t=6) = Plabel(l=abcd @ t=6)
|
|
c.newp.total = c.newp.label;
|
|
|
|
if (is_candidate(c.newp)) {
|
|
// Before adding the new node to the beam, check if the beam
|
|
// is already at maximum width.
|
|
if (leaves_.size() == beam_width_) {
|
|
// Bottom is no longer in the beam search. Reset
|
|
// its probability; signal it's no longer in the beam search.
|
|
BeamEntry* bottom = leaves_.peek_bottom();
|
|
bottom->newp.Reset();
|
|
}
|
|
leaves_.push(&c);
|
|
} else {
|
|
// Deactivate child.
|
|
c.oldp.Reset();
|
|
c.newp.Reset();
|
|
}
|
|
}
|
|
}
|
|
} // for (BeamEntry* b...
|
|
}
|
|
|
|
template <typename CTCBeamState, typename CTCBeamComparer>
|
|
void CTCBeamSearchDecoder<CTCBeamState, CTCBeamComparer>::Reset() {
|
|
leaves_.Reset();
|
|
|
|
// This beam root, and all of its children, will be in memory until
|
|
// the next reset.
|
|
beam_root_.reset(new BeamRoot(nullptr, -1));
|
|
beam_root_->RootEntry()->newp.total = 0.0; // ln(1)
|
|
beam_root_->RootEntry()->newp.blank = 0.0; // ln(1)
|
|
|
|
// Add the root as the initial leaf.
|
|
leaves_.push(beam_root_->RootEntry());
|
|
|
|
// Call initialize state on the root object.
|
|
beam_scorer_->InitializeState(&beam_root_->RootEntry()->state);
|
|
}
|
|
|
|
template <typename CTCBeamState, typename CTCBeamComparer>
|
|
bool CTCBeamSearchDecoder<CTCBeamState, CTCBeamComparer>::TopPaths(
|
|
int n, std::vector<std::vector<int>>* paths, std::vector<float>* log_probs,
|
|
bool merge_repeated) const {
|
|
TFLITE_DCHECK(paths);
|
|
TFLITE_DCHECK(log_probs);
|
|
paths->clear();
|
|
log_probs->clear();
|
|
if (n > beam_width_) {
|
|
return false;
|
|
}
|
|
if (n > leaves_.size()) {
|
|
return false;
|
|
}
|
|
|
|
gtl::TopN<BeamEntry*, CTCBeamComparer> top_branches(n);
|
|
|
|
// O(beam_width_ * log(n)), space complexity is O(n)
|
|
for (auto it = leaves_.unsorted_begin(); it != leaves_.unsorted_end(); ++it) {
|
|
top_branches.push(*it);
|
|
}
|
|
// O(n * log(n))
|
|
std::unique_ptr<std::vector<BeamEntry*>> branches(top_branches.Extract());
|
|
|
|
for (int i = 0; i < n; ++i) {
|
|
BeamEntry* e((*branches)[i]);
|
|
paths->push_back(e->LabelSeq(merge_repeated));
|
|
log_probs->push_back(e->newp.total);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
} // namespace ctc
|
|
} // namespace experimental
|
|
} // namespace tflite
|
|
|
|
#endif // TENSORFLOW_LITE_EXPERIMENTAL_KERNELS_CTC_BEAM_SEARCH_H_
|