New features
- Support LongLoRA for the LLaMA models
- Support training the Qwen-14B and InternLM-20B models
- Support training state recovery for the all-in-one Web UI
- Support Ascend NPU by @statelesshz in #975
- Integrate MMLU, C-Eval and CMMLU benchmarks
Modifications
- Rename repository to LLaMA Factory (former LLaMA Efficient Tuning)
- Use the
cutoff_lenargument instead ofmax_source_lengthandmax_target_length#944 - Add a
train_on_promptoption #1184