-
Notifications
You must be signed in to change notification settings - Fork 346
Pull requests: NVIDIA/TransformerEngine
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Make it an option to compile activation functions with fast math
#1410
opened Jan 15, 2025 by
guyueh1
Loading…
1 of 13 tasks
Support
store_param_remainders
feature from Apex in TE Fused Adam
#1408
opened Jan 13, 2025 by
sanandaraj5597
Loading…
[JAX] Support segment_ids/pos as FA inputs
#1406
opened Jan 13, 2025 by
zlsh80826
Loading…
9 of 13 tasks
[JAX] Consolidate the distributed fused attention test code
#1405
opened Jan 12, 2025 by
mgoldfarb-nvidia
Loading…
8 of 13 tasks
[PyTorch] Avoid Something isn't working
parameters
function in op backward pass
bug
#1403
opened Jan 11, 2025 by
timmoon10
Loading…
3 of 13 tasks
[PyTorch] Fix AttentionParams comparison logic
#1397
opened Jan 9, 2025 by
cyanguwa
Loading…
8 of 13 tasks
Don't touch nor send messages to the root logger.
#1380
opened Dec 19, 2024 by
sagostinho-nvidia
Loading…
4 of 13 tasks
[MoE][PyTorch] Add mask-based MoE permutation
#1373
opened Dec 13, 2024 by
hxbai
Loading…
8 of 13 tasks
[PyTorch] Bugfix for wgrad bulk overlap conflict when dgrad overlap is reduce-scatter
bug
Something isn't working
#1341
opened Nov 18, 2024 by
denera
Loading…
6 of 13 tasks
TP communication overlap: enable the overlap between GEMM chunk at Ho…
#1311
opened Nov 4, 2024 by
erhoo82
Loading…
1 of 13 tasks
[JAX] Collective GEMM custom op with
nvte_cublas_gemm
(no comm. overlap)
jax
#1307
opened Nov 2, 2024 by
denera
Loading…
7 of 17 tasks
[PyTorch] Add heuristics for intializing FP8 params
enhancement
New feature or request
#1300
opened Oct 30, 2024 by
timmoon10
Loading…
8 of 13 tasks
attention_mask fill with -inf for UnfusedDotProductAttention
#1268
opened Oct 18, 2024 by
Agoniii
Loading…
1 of 13 tasks
Previous Next
ProTip!
Follow long discussions with comments:>50.