You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
Not really a problem, but rather an optimisation. I'm currently hosting Llama3 and would like to reuse this model for every task like User queries, Chat title generation, Web search. The issue I have is that I have set larger context length for the model in the Workspaces, which differs to the one (default) used in Task Model. The result is that Ollama constantly reloads the model, since the LLM parameters are different. If they are the same, the model will be kept in the GPU memory and will be much efficient.
Describe the solution you'd like
A best solultion would be to have an option to use the workspace model's parameters in Set Task Model, or less ideal to have a way to override the default LLM parameters using environment variables.
Describe alternatives you've considered
The only alternative I have found is to use very small model that can fit in the GPU memory along with the larger and use it as Task Model.
The text was updated successfully, but these errors were encountered:
boshk0
changed the title
Task Model to use the same LLM parameters as in Workspaces
Feature request: Task Model to use the same LLM parameters as in Workspaces
Jun 19, 2024
Is your feature request related to a problem? Please describe.
Not really a problem, but rather an optimisation. I'm currently hosting Llama3 and would like to reuse this model for every task like User queries, Chat title generation, Web search. The issue I have is that I have set larger context length for the model in the Workspaces, which differs to the one (default) used in Task Model. The result is that Ollama constantly reloads the model, since the LLM parameters are different. If they are the same, the model will be kept in the GPU memory and will be much efficient.
Describe the solution you'd like
A best solultion would be to have an option to use the workspace model's parameters in Set Task Model, or less ideal to have a way to override the default LLM parameters using environment variables.
Describe alternatives you've considered
The only alternative I have found is to use very small model that can fit in the GPU memory along with the larger and use it as Task Model.
The text was updated successfully, but these errors were encountered: