github unslothai/unsloth v0.1.25-beta
New Important Updates!

5 hours ago

Hey guys, it's only been 2 days since our last release, but we’ve got a lot more important updates:

  • Inference is now 20–30% faster. Previously, tool-calling and repeat penalty could slow inference below normal speeds. Inference tokens/s should now perform similar to llama-server / llama.cpp.
  • Now Auto-detects older or pre-existing models downloaded from LM Studio, Hugging Face, and similar sources.
  • Inference token/s speed is now calculated correctly. Previously, tokens/s included startup time, which made the displayed speed look slower than it actually was. It should now reflect 'true' inference speed.
  • CPU usage no longer spikes. Previously, inline querier identity changed every render, causing useLiveQuery to resubscribe continuously.
  • Unsloth Studio now has a shutdown x button and shuts down properly. Previously, closing it after opening from the desktop icon would not close it properly. Now, launching from the shortcut also opens the terminal, and closing that terminal fully exits Unsloth Studio. If you still have it open from a previous session you can restart your computer or run lsof -i :8888 then kill -9 <PID>.
  • Even better tool-calling and websearch with reduced errors.
  • Updated documentation with lots of new info on deleting models, uninstalling etc.
  • Cleaner, smarter install and setup logging across Windows and Linux. Output is now easier to read with consistent formatting, quieter by default for a smoother experience, and supports richer --verbose diagnostics when you want full technical detail.
    {% endupdate %}
  • You can now view your training history

What's Changed

New Contributors

Full Changelog: v0.1.2-beta...v0.1.25-beta

Don't miss a new unsloth release

NewReleases is sending notifications on new releases.