Skip to content

Latest commit

 

History

History
31 lines (24 loc) · 1.11 KB

model_conversion.md

File metadata and controls

31 lines (24 loc) · 1.11 KB

Model Conversion

Introduction

Model conversion is used to convert different TensorFlow model format to another.

Now it supports QAT(quantization aware training) model to default(quantized) model. In the future, we will add tflite to default model support.

How to use it

See the following example which demonstrate model conversion API usage.

    from neural_compressor.experimental import ModelConversion, common
    conversion = ModelConversion()
    conversion.source = 'QAT'
    conversion.destination = 'default'
    conversion.model = common.Model('/path/to/trained/saved_model')
    q_model = conversion()
    q_model.save('/path/to/quantized/saved_model')

After this conversion is done, user could measure the accuracy or performance on quantized model.

  from neural_compressor.experimental import Benchmark, common
  evaluator = Benchmark('/path/to/yaml')
  evaluator.model = common.Model('/path/to/quantized/saved_model')
  evaluator.b_dataloader = ...       # create benchmark dataloader like examples/tensorflow/qat/benchmark.py
  evaluator('accuracy')