Skip to content

Pull requests: NVIDIA/TransformerEngine

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[PyTorch] Quantizer as API
#2039 opened Aug 7, 2025 by negvet Loading…
5 of 11 tasks
Ubnext
#2038 opened Aug 6, 2025 by nv-akorzh Loading…
Symmetric Memory Pool
#2037 opened Aug 6, 2025 by wdykas Loading…
13 tasks
[JAX] TE Gemm custom call clean up
#2030 opened Aug 5, 2025 by phu0ngng Loading…
4 of 13 tasks
Add backward RMSNorm+Add fusion
#2028 opened Aug 5, 2025 by janekb04 Loading…
8 of 13 tasks
[PyTorch] Register weight and bias params in linear op bug Something isn't working
#2027 opened Aug 4, 2025 by timmoon10 Loading…
6 of 13 tasks
Offloading support for multiple attention layouts
#2024 opened Aug 3, 2025 by sanandaraj5597 Loading…
[PyTorch] Fix bug when deducing dtype in linear functional API bug Something isn't working
#2017 opened Aug 1, 2025 by timmoon10 Loading…
6 of 13 tasks
Dropout with 8-bit RNG
#2014 opened Jul 30, 2025 by vasunvidia Loading…
13 tasks
[Draft] Add FP8 attention with current scaling
#2012 opened Jul 30, 2025 by cyanguwa Loading…
8 of 13 tasks
[PyTorch] Disable fused dbias-quantize kernel for unsupported recipes bug Something isn't working
#2007 opened Jul 30, 2025 by timmoon10 Loading…
6 of 13 tasks
Remove if-else and torch.tensor to meet cudagraph requirement
#1997 opened Jul 25, 2025 by katec846 Loading…
13 tasks
Use userbuffers for MXFP8 wgrad all-gather overlap
#1982 opened Jul 22, 2025 by djns99 Loading…
1 of 13 tasks
[C][PyTorch] NVFP4 forward MXFP8 backward recipe
#1970 opened Jul 18, 2025 by ksivaman Draft
2 of 13 tasks
[PyTorch Debug] Debug support for GroupedLinear
#1953 opened Jul 15, 2025 by pggPL Draft
13 tasks
[PyTorch] Support delay_wgrad_compute cudagraph
#1948 opened Jul 14, 2025 by buptzyb Loading…
2 of 13 tasks
Fix import error when flash attention 3 is installed
#1913 opened Jun 30, 2025 by HollowMan6 Loading…
7 of 13 tasks
[PyTorch debug] Improve precision debug tools performance
#1909 opened Jun 30, 2025 by pggPL Loading…
9 of 13 tasks
[PyTorch] Support FA3 MLA CP feature
#1907 opened Jun 28, 2025 by zhujian19891203 Loading…
7 of 13 tasks
[PyTorch Debug] Support log fp8 tensor stats for blockwise recipe
#1905 opened Jun 27, 2025 by lengerfulluse Loading…
12 tasks
[Common] NVFP4 kernels enhancement New feature or request
#1904 opened Jun 27, 2025 by Oleg-Goncharov Draft
5 of 13 tasks
Fix fp8_calibration path
#1903 opened Jun 27, 2025 by sudhakarsingh27 Draft
1 of 13 tasks
[PyTorch Debug] More advanced stats for Quantized Tensors
#1897 opened Jun 26, 2025 by pggPL Loading…
2 of 13 tasks
ProTip! Mix and match filters to narrow down what you’re looking for.