-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Report AVX512_BF16 support in CPUID features #10991
Labels
type: enhancement
New feature or request
Comments
I think it should look like this:
|
Awesome, I'll give that patch a test this week and try to get a PR together with some tests soon if that goes well 👍 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Description
We've been doing some performance analysis and have noticed that on bare-metal, a PyTorch image conversion from RGB to YUV will take over 1s for a sample image and on bare-metal it takes less than 50ms. This is fully on the CPU, no CUDA involved. I'm not super familiar with the CPU features that PyTorch uses or how to profile it, but looking at the CPU flags reported via
/proc/cpuinfo
on the host with what gVisor is reporting, the big missing AVX512 flag I'm seeing is for bfloat16 support,avx512_bf16
.From what I can tell reading the Intel manual, this flag is in the Structured Extended Feature Enumeration Sub-leaf, where EAX=0x07H and ECX=1 and Linux calls this block 12. I don't see support in
features_amd64.go
for this block at all and am wondering if there's already a tracking ticket for surfacing bf16 support in gVisor or a reason it's not supported (beyond just "hasn't been done yet" 😄).I can take a pass at implementing it if there's any advice around adding entire new blocks, especially when it looks like there's a gap going from block 7 to block 12.
Is this feature related to a specific bug?
No response
Do you have a specific solution in mind?
No response
The text was updated successfully, but these errors were encountered: