Skip to content

Commit

Permalink
fix is_fp8_supported function
Browse files Browse the repository at this point in the history
  • Loading branch information
ajtejankar committed Oct 28, 2024
1 parent 488d428 commit 813b64b
Showing 1 changed file with 3 additions and 2 deletions.
5 changes: 3 additions & 2 deletions server/lorax_server/utils/torch_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,8 +15,9 @@ def is_quantized(quantize):


def is_fp8_supported():
return (torch.cuda.get_device_capability()[0] >= 9) \
or (torch.cuda.get_device_capability()[0] == 8 and torch.cuda.get_device_capability()[1] >= 9)
return torch.cuda.is_available() and \
(torch.cuda.get_device_capability()[0] >= 9) or \
(torch.cuda.get_device_capability()[0] == 8 and torch.cuda.get_device_capability()[1] >= 9)


def is_fp8_kv(quantize):
Expand Down

0 comments on commit 813b64b

Please sign in to comment.