From 4eec977db75debc9dd36f4653e8041b3f963ed01 Mon Sep 17 00:00:00 2001 From: Suharsh Sivakumar Date: Wed, 27 Feb 2019 20:08:21 -0800 Subject: [PATCH] Fix codeblock in quantization docs. PiperOrigin-RevId: 236043945 --- tensorflow/lite/g3doc/convert/quantization.md | 11 +++++------ 1 file changed, 5 insertions(+), 6 deletions(-) diff --git a/tensorflow/lite/g3doc/convert/quantization.md b/tensorflow/lite/g3doc/convert/quantization.md index 2afdd3eb040..895f3e637e7 100644 --- a/tensorflow/lite/g3doc/convert/quantization.md +++ b/tensorflow/lite/g3doc/convert/quantization.md @@ -28,12 +28,11 @@ Currently, this requires training a model with Convert the graph: ``` -with tf.Session() as sess: - converter = tf.lite.TFLiteConverter.from_saved_model(saved_model_dir) - converter.inference_type = tf.lite.constants.QUANTIZED_UINT8 - input_arrays = converter.get_input_arrays() - converter.quantized_input_stats = {input_arrays[0] : (0., 1.)} # mean, std_dev - tflite_model = converter.convert() +converter = tf.lite.TFLiteConverter.from_saved_model(saved_model_dir) +converter.inference_type = tf.lite.constants.QUANTIZED_UINT8 +input_arrays = converter.get_input_arrays() +converter.quantized_input_stats = {input_arrays[0] : (0., 1.)} # mean, std_dev +tflite_model = converter.convert() ``` For fully integer models, the inputs are uint8. The `mean` and `std_dev values`