- Features
- Performance
- Fused QKV preprocessing with precomputed RoPE caches (3x preprocessing speedup, 10-14% E2E) (MR !3912)
- Use new TE interface for user buffers (MR !3886)
- Add CPU activation offloading via TE (MR !4286)
- Add configurable double buffering (MR !4026)
- Add Muon optimizer and distributed optimizer support (MR !4106)
- Add setting to support Adam or AdamW optimizer (MR !3866)
- MoE
- Model support
- Add YaRN support for GPT-OSS (MR !4044)
- Add support for Qwen3-Next arguments (MR !4070)
- Add FP8 init for MTP (MR !3958)
- Add fp8_dpa option for FP8 scaling (MR !4053)
- Add RADIO-g support to converter and tester (MR !4371)
- Add audio semantic reasoning data for voice chat and speech instructions (MR !4397)
- FSDP
- Inference
- Add CUDA Graph runner lookup table cache (up to 2x E2E speedup) (MR !4082)
- Add MoE dropping and padding router for CUDA Graph + decode (MR !3816)
- Dynamic audio shapes with variable sequence lengths (2.5x throughput improvement) (MR !4274)
- Integrate unified memory for dynamic inference context (MR !3985)
- Post-training
- RL
- Ease of use
- Performance
- Bug fixes
- Fix convergence bug in MXFP8 parameter gradient buffer reuse (MR !3999)
- Fix loss mask cloning to prevent incorrect updates (MR !4164)
- Fix metadata loss in checkpoints (MR !4182)
- Fix FSDP grad accum fusion support (MR !4018)
- Fix non-TE optimizer checkpoint issue (MR !3931)
- Fix BERT virtual pipeline parallelism (MR !3993)
- Fix gc.freeze() slowdown by adding gc.collect() on last layer (MR !4003)
- Fix full iteration CUDA graph non-tensor handling (MR !4019)
- Fix model_auto_sync mis-set and add gradient assertion (MR !4062)
- Fix HF import dtype and checkpoint loading issues (MR !4095)
- Fix missing initialization in ProcessGroupCollection (MR !4159)
- Fix sink attention TP (MR !4173)
- Fix num_microbatches calculation (MR !4199)
- Fix 1f1b overlap unit tests for MTP standalone (MR !4210)
- Fix stale state dict handling (MR !4226)
- Fix dataset divergence with tokenizer PAD handling (MR !4231)
- Fix parameter initialization (MR !4296)
- Ensure tensor-parallel attributes set regardless of initialization flag (MR !4312)
- Known issues
NVIDIA Megatron Core 0.14.0
- Features
- Inference
- Post-training
- ModelOpt updates (MR !3268)
- Add speculative decoding AR validation feature
- Add DeepSeek and Qwen model configs
- ModelOpt updates (MR !3268)
- Performance
- MoE
- We're actively optimizing large-scale fine-grained MoE performance on Blackwell Platform.
- Features:
- Memory Optimization
- Performance Optimization
- Bug fixes:
- Fix router input jitter dtype (MR !3774)
- Model support
- Ease of use
- Bug fixes
- Use mscale_all_dim for softmax_factor (MR !2800)
- Fix FP8 param blockwise scaling unit test (MR !3480)
- Fix unit test blockwise scaling (MR !3491)
- Optimize prefill for token-less requests (MR !3499)
- Add default values for Fp8Padding and Fp8Unpadding (MR !3501)
- Fix CUDA graph logic for flexible pp layout (MR !3505)
- Load FP8 models with strict=False (MR !3508)
- Skip rope check for torch < 1.4.0 (MR !3528)
- Disable Apex tests for stability (MR !3539)
- Fix typo in parallel_state expert parallelism (MR !3548)
- Guard modelopt on macOS (MR !3549)
- Retry on CUDA function failure (MR !3554)
- Fix NCCL mem pool creation error (MR !3557)
- Fix get_rotary_seq_len return type (MR !3559)
- Retry on CUDA function failure (MR !3560)
- Fix NCCL allocator attribute error (MR !3565)
- Ensure multi-prompt inference works (MR !3568)
- Fix MD5 on FIPS systems (MR !3577)
- Fixes dynamic context and inference bugs (MR !3582)
- Fix TE version for interleaved fused RoPE (MR !3586)
- Fix MTP with MoE and TP logging (MR !3594)
- Guard TE import fix (MR !3596)
- Add assertion for NCCL UB case (MR !3599)
- Remove Encoder PP related Functions (MR !3604)
- Fix segfaults in tests (MR !3605)
- Fix TE error in distributed optimizer (MR !3625)
- Remove redundant barrier in checkpoint flow (MR !3626)
- Support VPP MTP, fix logging (MR !3630)
- Retry mechanism for free(): invalid pointer errors (MR !3632)
- Fix test_replication.py issues (MR !3633)
- Fix typo in parallel_state (MR !3634)
- Fix CUDA graph logic determination (MR !3635)
- Fix TE installation error (MR !3636)
- Ensure correct sharding type in local tests (MR !3643)
- Fix cudagraphed backward buffer reuse for last layer (MR !3645)
- Set default for packed_seq_params in get_rotary_seq_len (MR !3651)
- Fix dynamic example script errors (MR !3653)
- Guard TE import fix (MR !3666)
- Breaking changes:
megatron.core.distributed.custom_fsdprefactored as breaking change tomegatron.core.distributed.fsdp.src.megatron_fsdp
- Known issues