You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Quantization routines use unsigned 32 bit integers to pack quantized data. Since PyTorch doesn't support this datatype, we're use numpy arrays before casting back to a Tensor, which currently limits us to running on the CPU. This restriction means we are forced to move data to/from the CPU when sending/receiving data between stages, rather than a more efficient direct network transfer to/from GPUs and NICs.
The text was updated successfully, but these errors were encountered:
Quantization routines use unsigned 32 bit integers to pack quantized data. Since PyTorch doesn't support this datatype, we're use numpy arrays before casting back to a Tensor, which currently limits us to running on the CPU. This restriction means we are forced to move data to/from the CPU when sending/receiving data between stages, rather than a more efficient direct network transfer to/from GPUs and NICs.
The text was updated successfully, but these errors were encountered: