March 23, 2026
- Improve pickle checkpoint handling security. Default all loading to
weights_only=True, add safe_global for ArgParse. - Improve attention mask handling for core ViT/EVA models & layers. Resolve bool masks, pass
is_causalthrough for SSL tasks. - Fix class & register token uses with ViT and no pos embed enabled.
- Add Patch Representation Refinement (PRR) as a pooling option in ViT. Thanks Sina (https://github.com/sinahmr).
- Improve consistency of output projection / MLP dimensions for attention pooling layers.
- Hiera model F.SDPA optimization to allow Flash Attention kernel use.
- Caution added to SGDP optimizer.
- Release 1.0.26. First maintenance release since my departure from Hugging Face.
What's Changed
- fix: replace 5 bare except clauses with except Exception by @haosenwang1018 in #2672
- Add timmx model export tool to README by @Boulaouaney in #2673
- Enhance SGDP optimizer with caution parameter by @Yuan-Jinghui in #2675
- Fix CLS and Reg tokens usage when pos_embed is disabled by @sinahmr in #2676
- default weights_only=True for load fns by @rwightman in #2679
- Fix Hiera global attention to use 4D tensors for efficient SDPA dispatch by @Raiden129 in #2680
- Improve 2d and latent attention pool dimension handling. Fix #2682 by @rwightman in #2684
- Improve attention mask handling for vision_transformer and eva and related blocks by @rwightman in #2686
- Implement PRR as a pooling module. Alternative to #2678 by @rwightman in #2685
New Contributors
- @haosenwang1018 made their first contribution in #2672
- @Raiden129 made their first contribution in #2680
Full Changelog: v1.0.25...v1.0.26