github huggingface/optimum-intel v1.12.0
v1.12.0: Weight only quantization, LCM, Pix2Struct , GPTBigCode

latest releases: v1.19.0, v1.18.3, push...
10 months ago

OpenVINO

New architectures

LCMs

optimum-cli export openvino --model gpt2 ov_model
  • Enable Latent Consistency models OpenVINO export and inference by @echarlaix in #463
from optimum.intel import OVLatentConsistencyModelPipeline

pipe = OVLatentConsistencyModelPipeline.from_pretrained("SimianLuo/LCM_Dreamshaper_v7", export=True)
prompt = "sailing ship in storm by Leonardo da Vinci"
images = pipe(prompt=prompt, num_inference_steps=4, guidance_scale=8.0).images

Pix2Struct

  • Add support for export and inference for pix2struct models by @eaidova in #450

GPTBigCode

  • Add support for export and inference for GPTBigCode models by @echarlaix in #459

Changes and bugfixes

model = OVModelForCausalLM.from_pretrained(model_id, load_in_8bit=True)
  • Create default attention mask when needed but not provided by @eaidova in #457
  • Do not automatically cache models when exporting a model in a temporary directory by @helena-intel in #462

Neural Compressor

Full Changelog: https://github.com/huggingface/optimum-intel/commits/v1.12.0

Don't miss a new optimum-intel release

NewReleases is sending notifications on new releases.