Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: browser wasm llm #2970

Open
tjbck opened this issue Jun 10, 2024 · 4 comments
Open

feat: browser wasm llm #2970

tjbck opened this issue Jun 10, 2024 · 4 comments

Comments

@tjbck
Copy link
Contributor

tjbck commented Jun 10, 2024

No description provided.

@justinh-rahb
Copy link
Collaborator

https://github.com/mlc-ai/web-llm or something like it?

@lukestanley
Copy link

lukestanley commented Jun 12, 2024

This is useful if you can only do a limited amount of processing on the server, and where the client has more processing power. WASM approaches to running Llama.cpp exist, as well as web GPU approaches (that only work on certain systems and and certain browser versions). There may be a WASM memory size issue with normal LLM sizes though. Working around it may well be possible though.

@tjbck
Copy link
Contributor Author

tjbck commented Jun 13, 2024

@justinh-rahb yep, seems promising!

@BuildBackBuehler
Copy link

@justinh-rahb yep, seems promising!

Yeah MLC-LLM is the ship!
#1270

Would love to plug-n-play with it

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants