Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

explore WebGPT for LLMs #642

Open
martinjaggi opened this issue Feb 28, 2024 · 4 comments
Open

explore WebGPT for LLMs #642

martinjaggi opened this issue Feb 28, 2024 · 4 comments

Comments

@martinjaggi
Copy link
Member

explore the WebGPT library as a potential (seems full?) replacement for TFJS.
try on small standard architecures like nanoGPT

we can explore this in parallel to our tfjs based prototype #641

who would be interested in giving this a shot? (can first be tried outside disco). maybe robin / albert potentially?

@JulienVig
Copy link
Collaborator

WebGPT is currently only compatible with Chrome Canary. It seems to be a big restriction for DISCO to have, what do you think @martinjaggi?

@martinjaggi
Copy link
Member Author

no issue at the moment
(huge audience for LLMs, and huge audience of chrome, so sufficiently many people interested)

@tharvik
Copy link
Collaborator

tharvik commented Mar 5, 2024

to add to @JulienVig's answer, it's an expertimental, still in draft, technology that's not supported around. there is currently only 25% of support for WebGPU, 0% on mobile. that effectively drops support for most of our targetted audience.
there might be way to optionally support it (isomorphic-webgpu that'll write ourself) but I don't know how well WebGPT supports it.

there are various backend of tfjs: WebGL (≠ WebGPU), wasm, cpu; the first one is the one running in the webapp (at least on my stable firefox). there is an ongoing project to add a WebGPU backend to tfjs which we can try out and would allow us to fallback on a slower backend if not supported.

all in all, I don't see the point of adding WebGPT for the only benefit of faster computation. are there other reasons maybe? (models based on it that's are not trivially converted to tfjs for eg?)

@martinjaggi
Copy link
Member Author

yea best would be to optionally support (only for LLM tasks, not affecting the rest of disco), and if not available just show an error message to the user

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants