Changes
- Bump bitsandbytes Windows wheel by @jllllll in #3097 --
--load-in-4bit
is now a lot faster - Add support low vram mode on llama.cpp module by @gabriel-pena in #3076
- Add links/reference to new multimodal instructblip-pipeline in multimodal readme by @kjerk in #2947
- Add token authorization for downloading model by @fahadh4ilyas in #3067
- Add default environment variable values to docker compose file by @Josh-XT in #3102
- models/config.yaml: +platypus/gplatty, +longchat, +vicuna-33b, +Redmond-Hermes-Coder, +wizardcoder, +more by @matatonic in #2928
- Add context_instruct to API. Load default model instruction template … by @atriantafy in #2688
- Chat history download creates more detailed file names by @UnskilledWolf in #3051
- Disable wandb remote HTTP requests
- Add Feature to Log Sample of Training Dataset for Inspection by @practicaldreamer in #1711
- Add ability to load all text files from a subdirectory for training by @kizinfo in #1997
- Add Tensorboard/Weights and biases integration for training by @kabachuha in #2624
- Fix: Fixed the tokenization process of a raw dataset and improved its efficiency by @Nan-Do in #3035
- More robust and error prone training by @FartyPants in #3058
Bug fixes
- [Fixed] wbits and groupsize values from model not shown by @set-soft in #2977
- Fix API example for loading models by @vadi2 in #3101
- google flan T5 tokenizer download fix by @FartyPants in #3080
- Changed FormComponent to IOComponent by @ricardopinto in #3017
- respect model dir for downloads by @micsthepick in #3079
Extensions
- Fix send_pictures extension
- Elevenlabs tts fixes by @set-soft in #2959
- [extensions/openai]: Major openai extension updates & fixes by @matatonic in #3049
- substitu superboog Beatiful Soup Parser by @juhenriquez in #2996