Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

only support hopper GPU? #319

Open
QingshuiL opened this issue Feb 20, 2025 · 1 comment
Open

only support hopper GPU? #319

QingshuiL opened this issue Feb 20, 2025 · 1 comment

Comments

@QingshuiL
Copy link

QingshuiL commented Feb 20, 2025

Very good work, but I have some questions to consult.

When I tried to run the code, I encountered the following error.

[rank0]: Traceback (most recent call last):
[rank0]: File "/.conda/envs/torch2.4/lib/python3.10/site-packages/triton/language/core.py", line 35, in wrapper
[rank0]: return fn(*args, **kwargs)
[rank0]: File "/.conda/envs/torch2.4/lib/python3.10/site-packages/triton/language/core.py", line 993, in to
[rank0]: return semantic.cast(self, dtype, _builder, fp_downcast_rounding)
[rank0]: File "/.conda/envs/torch2.4/lib/python3.10/site-packages/triton/language/semantic.py", line 759, in cast
[rank0]: assert builder.options.allow_fp8e4nv, "fp8e4nv data type is not supported on CUDA arch < 89"
[rank0]: AssertionError: fp8e4nv data type is not supported on CUDA arch < 89

So the project is still not out of the fp8 restrictions?

@QingshuiL
Copy link
Author

I can only use the a100 GPU now. Will the quantization and inference of the A100 GPU be supported in the future? How many A100-40G GPUs are required if supported?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant