-
-
Notifications
You must be signed in to change notification settings - Fork 3.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: browser wasm llm #2970
Comments
https://github.com/mlc-ai/web-llm or something like it? |
This is useful if you can only do a limited amount of processing on the server, and where the client has more processing power. WASM approaches to running Llama.cpp exist, as well as web GPU approaches (that only work on certain systems and and certain browser versions). There may be a WASM memory size issue with normal LLM sizes though. Working around it may well be possible though. |
@justinh-rahb yep, seems promising! |
Yeah MLC-LLM is the ship! Would love to plug-n-play with it |
No description provided.
The text was updated successfully, but these errors were encountered: