-
-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: allow setting num_gpu
parameter
#2877
Comments
bumping this |
derpyhue/openwebui_num_gpu@fff91f7 |
1 |
Is your feature request related to a problem? Please describe.
To avoid creating a new modelfile only for adjusting/finetuning the number of layers offloaded to the GPU, make this setting (
num_gpu
) user-configurable, which ollama considers as one of the most common parameters.Describe the solution you'd like
Implement and add 'num_gpu (Ollama)' in the 'Advanced Params' section of a model in Workspace > Models.
(I would not add it to the 'Advanced Parameters' section of Settings > General, as the number and size of layers is model- and-quant-specific.
Describe alternatives you've considered
Well, creating a new ollama model(file) with an adjusted
num_gpu
, but this is cumbersome if one wants to adjust/finetunenum_gpu
(or modify it quickly if the GPU runs other things/models).Additional context
num_gpu
is not specified1 in the official ollama docs as valid PARAMETER, but is supported by the API.Footnotes
like
use_mmap
,use_mlock
, andnum_thread
, which are already configurable in open_webui. ↩The text was updated successfully, but these errors were encountered: