You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Nodetool already has the ability to add extra-index-urls for Torch, which means that reusing this functionality to support nexa-sdk should be easy. In this scenario, the CUDA version for Torch would need to increase to 12.4, matching the installable version of ComfyUI.
Nexa SDK would be a great alternative or addition to Ollama. It would also give us a wider range of models to give users more options that better fit their hardware.
Do you know of any benchmarks? In particular, how well does it run on ONNX or MPS?
I don't think the ONNX support is well developed. However, there is an evaluation component for GGUF that you can use to measure performance.
Regarding additional models, T2I support is lagging behind at the moment (issues#358) and TTS support is currently disabled (pulls#359). While both of these are temporary issues that will be addressed in time, for now the most useful parts of Nexa SDK are the Python integration and the well-developed TGI backend.
Nodetool currently relies on Ollama, an unmanaged dependency that requires manual user installation.
It would be helpful to consider replacing this dependency with nexa-sdk, which can be installed and managed from within Python.
Refactor
Like Torch, nexa-sdk uses custom repositories for different platforms:
Nodetool already has the ability to add extra-index-urls for Torch, which means that reusing this functionality to support nexa-sdk should be easy. In this scenario, the CUDA version for Torch would need to increase to 12.4, matching the installable version of ComfyUI.
Functionality
Looking at src/nodetool/providers/ollama, I observe that nexa-sdk:
EDIT: removed erroneous commentary regarding Torch version.
The text was updated successfully, but these errors were encountered: