-
Notifications
You must be signed in to change notification settings - Fork 352
Issues: NVIDIA/TransformerEngine
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
FP8 execution requires 2D input matrices with height divisible by 8 and width divisible by 16
#1422
opened Jan 25, 2025 by
Liufeiran123
Deadline or schedule new update supporting blackwell and fp4?
#1421
opened Jan 24, 2025 by
johnnynunez
Problem when install transformers_engine with nvcc11.8 and nvcc12.0
#1420
opened Jan 23, 2025 by
chwenjun225
Questions about accuracy alignment between BF16 and FP8
question
Further information is requested
#1419
opened Jan 22, 2025 by
zigzagcai
Questions on DotProductAttention API Usage in Flash Attention thd Mode
#1409
opened Jan 14, 2025 by
pipSu
Import fails when working from a TE directory
good first issue
Good for newcomers
#1400
opened Jan 10, 2025 by
ksivaman
_NoopCatFunc in transformer layer
bug
Something isn't working
#1384
opened Dec 22, 2024 by
robot-transformer
AttributeError: module 'transformer_engine' has no attribute 'pytorch'
#1379
opened Dec 17, 2024 by
carrot0117
Support more than 1 shape/attention_params for DotProductAttention decision cache
#1349
opened Nov 29, 2024 by
parthmannan
the max error of moe_permute/unpermute.grad could reach 3.6e+00
#1336
opened Nov 15, 2024 by
NiuMa-1234
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.