You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Stable diffusion with W8A8 quantization- This is important because the stable diffusion demo which I saw uses fp16 weights with transformers.js as the engine and requires webgpu/webnn support. Onnxstream with W8A8 quantization(and without webgpu/webnn requirement) should be much lighter.
LLM(llama) with int4-int8 quantization- This will become a feasible alternative to wllama
The text was updated successfully, but these errors were encountered:
Please create the following browser wasm demos-
The text was updated successfully, but these errors were encountered: