Skip to content

vulkan: Add logging for bf16 features to ggml_vk_print_gpu_info (#13274) #14707

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jul 19, 2025

Conversation

Peter0x44
Copy link
Contributor

@Peter0x44 Peter0x44 commented Jul 16, 2025

Issue #13274 requested logging whether the VK_KHR_bfloat16 is supported at runtime. I have added the detection for the feature to ggml_vk_print_gpu_info, and tested it on windows with the latest Nvidia GeForce Game Ready Driver (where it reported 0) as well as the Nvidia Vulkan 1.4 Developer Beta Driver (where it reported 1).

I have not tested Linux or any GPUs other than my laptop RTX 4060.

@github-actions github-actions bot added Vulkan Issues specific to the Vulkan backend ggml changes relating to the ggml tensor library for machine learning labels Jul 16, 2025
Copy link
Collaborator

@0cc4m 0cc4m left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The code looks fine to me. I don't think bf16 has major performance implications currently, unless maybe if it is used extensively in a model, but knowing whether it's supported natively can't hurt.

@0cc4m 0cc4m merged commit d4b91ea into ggml-org:master Jul 19, 2025
45 of 48 checks passed
@oscarbg
Copy link
Contributor

oscarbg commented Jul 29, 2025

nice! don't forget to add similar fp8 Vulkan support reporting once llama.cpp adds use of VK_EXT_shader_float8 to it's Vulkan backend!

@Peter0x44 Peter0x44 deleted the vulkan_bf16_logging branch July 29, 2025 22:45
@0cc4m
Copy link
Collaborator

0cc4m commented Jul 30, 2025

nice! don't forget to add similar fp8 Vulkan support reporting once llama.cpp adds use of VK_EXT_shader_float8 to it's Vulkan backend!

This is only relevant if the new datatype makes any difference in the backend. bf16 is being used in specific cases, float8 would need a shader to use it first.

Edit: On second read, that is what you said. I misread it, sorry.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ggml changes relating to the ggml tensor library for machine learning Vulkan Issues specific to the Vulkan backend
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants