github huggingface/optimum-intel v1.5.0
v1.5.0: OpenVINO quantization

latest releases: v1.19.0, v1.18.3, push...
23 months ago

Quantization

  • Add OVQuantizer enabling OpenVINO NNCF post-training static quantization (#50)
  • Add OVTrainer enabling OpenVINO NNCF quantization aware training (#67)
  • Add OVConfig the configuration which contains the quantization process informations (#65)

The quantized model resulting from the OVQuantizer and the OVTrainer are exported to the OpenVINO IR and can be loaded with the corresponding OVModelForXxx to perform inference with OpenVINO Runtime.

OVModel

Add OVModelForCausalLM enabling OpenVINO Runtime for models with a causal language modeling head (#76)

Don't miss a new optimum-intel release

NewReleases is sending notifications on new releases.