What's Changed
- Fix for using Torch with CUDA 11.8 by @sammcj in #4298
- openai: fix wrong models list on query present in /v1/models by @hronoas in #4139
- More silero languages by @missionfloyd in #3950
- ExLlamav2_HF: Convert logits to FP32 by @turboderp in #4310
- Support LLaVA v1.5 by @haotian-liu in #4305
- Structured requirements && Python 3.11 support by @mjbogusz in #4233
- Enable special token support for exllamav2 by @JohanAR in #4314
- Add flash-attention 2 for windows by @bdashore3 in #4235
- Docker: Remove explicit CUDA 11.8 Reference by @whiteadam in #4343
- Add a proper documentation by @oobabooga in #3885
- USE_CUDA118 from ENV remains null one_click.py + cuda-toolkit by @mongolu in #4352
- Training PRO a month worth of updates by @FartyPants in #4345
- Support LLaVA v1.5 7B by @cnut1648 in #4348
- Option to select/target additional linear modules/layers in LORA training by @computerman00 in #4178
- Allow multiple loaded LoRAs to simultaneously influence the output by @Googulator in #3120
New Contributors
- @hronoas made their first contribution in #4139
- @mjbogusz made their first contribution in #4233
- @whiteadam made their first contribution in #4343
- @mongolu made their first contribution in #4352
- @cnut1648 made their first contribution in #4348
- @computerman00 made their first contribution in #4178
- @Googulator made their first contribution in #3120
Full Changelog: snapshot-2023-10-15...snapshot-2023-10-22