Call the logits probs in create_inference_graph after they go thru softmax

This commit is contained in:
Catalin Voss 2020-11-03 09:49:27 -08:00 committed by Reuben Morais
parent 4270e22fe0
commit c5410fc550

View File

@ -730,7 +730,7 @@ def create_inference_graph(batch_size=1, n_steps=16, tflite=False):
logits = tf.squeeze(logits, [1])
# Apply softmax for CTC decoder
logits = tf.nn.softmax(logits, name='logits')
probs = tf.nn.softmax(logits, name='logits')
if batch_size <= 0:
if tflite:
@ -743,7 +743,7 @@ def create_inference_graph(batch_size=1, n_steps=16, tflite=False):
'input_lengths': seq_length,
},
{
'outputs': logits,
'outputs': probs,
},
layers
)
@ -763,7 +763,7 @@ def create_inference_graph(batch_size=1, n_steps=16, tflite=False):
inputs['input_lengths'] = seq_length
outputs = {
'outputs': logits,
'outputs': probs,
'new_state_c': new_state_c,
'new_state_h': new_state_h,
'mfccs': mfccs,