VapourSynth ML filter runtimes.
Please see the wiki for supported models.
OpenVINO is an AI inference runtime developed by Intel, mainly targeting x86 CPUs and Intel GPUs.
The vs-openvino plugin provides optimized pure CPU runtime for some popular AI filters, with Intel GPU support planned in the future.
To install, download the latest release and extract them into your VS plugins
directory.
Please visit the vsov directory for details.
ONNX Runtime is an AI inference runtime with many backends.
The vs-onnxruntime plugin provides optimized CPU and CUDA GPU runtime for some popular AI filters.
To install, download the latest release and extract them into your VS plugins
directory.
Please visit the vsort directory for details.
TensorRT is a highly optimized AI inference runtime for NVidia GPUs. It uses benchmarking to find the optimal kernel to use for your specific GPU, and so there is an extra step to build an engine from ONNX network on the machine you are going to use the vstrt filter, and this extra step makes deploying models a little harder than the other runtimes. However, the resulting performance is also typically much much better than the CUDA backend of vsort.
To install, download the latest release and extract them into your VS plugins
directory.
Please visit the vstrt directory for details.