-
-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Why not support Tesla P100 and limit computing power to greater than 7? #1284
Comments
As the runtime error states, the P100 only has a compute capability of 6.0, which is lower than the minimum requirement of 8.0 for bfloat16 support. This means that Tesla P100 cannot use bfloat16 efficiently or correctly. By limiting the computing power to greater than 8, the software can ensure that the GPU has the right hardware and software to use bfloat16 with no problem. |
But P100 is widely used in school teaching, with lower performance and no problem. If Vllm is discarded during the teaching phase, and company employees do not learn Vllm well during the school phase, would it be a loss for Vllm. |
A software that is not only used in production environments. Before being used in production, it is used for teaching, validation of technical routes, and so on. |
A software that targets the world, and many developing and underdeveloped countries also need to use good technology to develop their own countries. |
add parameter in command line |
which means loading the model using fp16 (v100 support), but I'm not sure if it performs the same as bf16 loading. fp16 is less precise than bf16. (fp16 has 5 bits for exponent, bf16 has 8 bits) |
You can use |
Bfloat16 is only supported on GPUs with compute capability of at least 8.0. Your Tesla P100-PCIE-16GB GPU has compute capability 6.0.
RuntimeError: GPUs with compute capability below 7.0 are not supported.
The text was updated successfully, but these errors were encountered: