github huggingface/optimum-intel v1.16.0
v1.16.0: OpenVINO config, SD hybrid quantization

latest releases: v1.20.1, v1.20.0, v1.19.0...
7 months ago

Add hybrid quantization for Stable Diffusion pipelines by @l-bat in #584

from optimum.intel import OVStableDiffusionPipeline, OVWeightQuantizationConfig

model_id = "echarlaix/stable-diffusion-v1-5-openvino"
quantization_config = OVWeightQuantizationConfig(bits=8, dataset="conceptual_captions")
model = OVStableDiffusionPipeline.from_pretrained(model_id, quantization_config=quantization_config)

Add openvino export configs by @eaidova in #568

Enabling OpenVINO export for the following architectures enabled : Mixtral, ChatGLM, Baichuan, MiniCPM, Qwen, Qwen2, StableLM

Add support for export and inference for StarCoder2 models by @eaidova in #619

Don't miss a new optimum-intel release

NewReleases is sending notifications on new releases.