Patch Release 4.51.2
This is another round of bug fixes, but they are a lot more minor and outputs were not really affected!
- Fix Llama4 offset (#37414) by @Cyrilvallez
- Attention Quantization with FBGemm & TP (#37384) by @MekkCyber
- use rms_norm_eps for the L2Norm for Llama4 (#37418) by @danielhanchen
- mark llama4 as not supported with fa2 (#37416) by @winglian