Update quantization docs to use TFLiteConverter.from_saved_model() API instead of .from_keras_model() API
PiperOrigin-RevId: 317251205 Change-Id: Ia8166decfa76327e3fd44871b194ffcae0f049f8
This commit is contained in:
parent
158d4be42d
commit
539e9cb3a2
@ -34,7 +34,7 @@ weights from floating point to integer, which has 8-bits of precision:
|
||||
|
||||
<pre>
|
||||
import tensorflow as tf
|
||||
converter = tf.lite.TFLiteConverter.from_keras_model(keras_model)
|
||||
converter = tf.lite.TFLiteConverter.from_saved_model(saved_model_dir)
|
||||
<b>converter.optimizations = [tf.lite.Optimize.DEFAULT]</b>
|
||||
tflite_quant_model = converter.convert()
|
||||
</pre>
|
||||
@ -68,7 +68,7 @@ the following steps:
|
||||
|
||||
<pre>
|
||||
import tensorflow as tf
|
||||
converter = tf.lite.TFLiteConverter.from_keras_model(keras_model)
|
||||
converter = tf.lite.TFLiteConverter.from_saved_model(saved_model_dir)
|
||||
<b>converter.optimizations = [tf.lite.Optimize.DEFAULT]
|
||||
def representative_dataset_gen():
|
||||
for _ in range(num_calibration_steps):
|
||||
@ -96,7 +96,7 @@ the following steps:
|
||||
|
||||
<pre>
|
||||
import tensorflow as tf
|
||||
converter = tf.lite.TFLiteConverter.from_keras_model(keras_model)
|
||||
converter = tf.lite.TFLiteConverter.from_saved_model(saved_model_dir)
|
||||
converter.optimizations = [tf.lite.Optimize.DEFAULT]
|
||||
def representative_dataset_gen():
|
||||
for _ in range(num_calibration_steps):
|
||||
@ -120,7 +120,7 @@ quantization of weights, use the following steps:
|
||||
|
||||
<pre>
|
||||
import tensorflow as tf
|
||||
converter = tf.lite.TFLiteConverter.from_keras_model(keras_model)
|
||||
converter = tf.lite.TFLiteConverter.from_saved_model(saved_model_dir)
|
||||
<b>converter.optimizations = [tf.lite.Optimize.DEFAULT]
|
||||
converter.target_spec.supported_types = [tf.float16]</b>
|
||||
tflite_quant_model = converter.convert()
|
||||
|
Loading…
Reference in New Issue
Block a user