[0.0.1] (05 Oct 2022)
We are excited to announce the first binary release of Dorado. This Dorado release introduces important new features such as support for modified base calling, and significant improvements to basecalling performance and usability, taking it to the state of the art for speed and accuracy.
Major changes
- d3ddd1f, 37e28f7, - Support for mod base calling, performance enhancements.
- dd79bf5, 2a7fc17, 465cb4a, 56482fb - Support for basecalling on M1 family Apple Silicon.
- bd6014e - Support for basecalling on systems with multiple Nvidia GPUs.
- 41fdb11 - Support for POD5 file format.
- 0750654 - Addition of new “Quantile” - based read scaling algorithm for higher accuracy.
- 8acf2ba - Upgrade to torch 1.12.1
- 9955d0d - Added fast int8-quantization optimisation for LSTM networks with layer size of 96 or 128
- f2e993d - New cuBLAS-based implementation for LSTMs leading to state-of-the-art performance.
- 6ec50dc - Major rearchitecture of CUDA model runners for higher basecalling speed and lower GPU memory utilisation.
- a0a197f - Accuracy improvements to handling of short reads (<1Kb) with an upgraded padding strategy.
- d01bf04 - Ability to download basecalling models from Dorado.
- 7c7e59c - Support for SAM output
Minor Changes
- 0e89d63 - Automatic selection of batch size if user does not specify.
- 6afceea - Dorado version added to SAM output, including commit hash
- 339b2fc - Scaling information recorded in SAM output
- afbfab9 - Timestamps added to SAM output
- 9ec2d97 - Support for multi-threaded read scaling, preventing CPU bottlenecks and improving basecall speed.
- 7cbdbe0 - Support for multi-threaded POD5 reading for higher data ingestion rate and improved performance.
- 5a33e83 - Automatic querying of M1 device type for detection of optimal basecalling parameters.
- 42703d0 - Basecalling progress (Number of reads basecalled) printed out to terminal on Linux.