Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feat: Adjust model hub v0.4.13 #2879

Merged
merged 39 commits into from
May 13, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
39 commits
Select commit Hold shift click to select a range
500ed5d
fix: correct phi3
hahuyhoang411 May 8, 2024
33e53e5
redundant phi2 dolphin
hahuyhoang411 May 8, 2024
4489413
add: hermes llama3
hahuyhoang411 May 8, 2024
e345931
add: ngl settings
hahuyhoang411 May 8, 2024
fe79079
correct ctx len
hahuyhoang411 May 8, 2024
1c4d256
correct ngl
hahuyhoang411 May 8, 2024
cd41e5c
correct maxlen ngl
hahuyhoang411 May 8, 2024
a5fd13d
disable phi3
hahuyhoang411 May 8, 2024
a369581
add ngl
hahuyhoang411 May 8, 2024
0ba3dde
add ngl
hahuyhoang411 May 8, 2024
627013e
add ngl
hahuyhoang411 May 8, 2024
7fd1637
add ngl
hahuyhoang411 May 8, 2024
e173a40
add ngl
hahuyhoang411 May 8, 2024
8e471a3
add ngl
hahuyhoang411 May 8, 2024
8bd880c
add ngl
hahuyhoang411 May 8, 2024
fc6b075
remove redundant hermes pro
hahuyhoang411 May 8, 2024
8b511e1
add ngl
hahuyhoang411 May 8, 2024
1a8f96e
add ngl
hahuyhoang411 May 8, 2024
ee95ffe
add ngl
hahuyhoang411 May 8, 2024
2645df7
remove miqu
hahuyhoang411 May 8, 2024
a2587a1
add ngl
hahuyhoang411 May 8, 2024
15c8146
add ngl
hahuyhoang411 May 8, 2024
baa2fb4
add ngl
hahuyhoang411 May 8, 2024
e840a3f
add ngl
hahuyhoang411 May 8, 2024
13f80dc
remove redundant
hahuyhoang411 May 8, 2024
8f276fc
add ngl
hahuyhoang411 May 8, 2024
1b9f302
add ngl
hahuyhoang411 May 8, 2024
710fc25
add ngl
hahuyhoang411 May 8, 2024
fd38f75
add ngl
hahuyhoang411 May 8, 2024
4484231
add ngl
hahuyhoang411 May 8, 2024
5747c4d
add ngl
hahuyhoang411 May 8, 2024
2ff1a16
add ngl
hahuyhoang411 May 8, 2024
ba34c43
add ngl
hahuyhoang411 May 8, 2024
bb2b440
add ngl
hahuyhoang411 May 8, 2024
6d58a2e
version package bump
hahuyhoang411 May 8, 2024
9c07126
feat: resolve issue of cannot found model in the extensions due to th…
May 13, 2024
3f25920
feat: completely remove hermes-pro-7b
May 13, 2024
f3b84b9
feat: completely remove openhermes-neural-7b and miqu-70b, and add ll…
May 13, 2024
18eb3b5
fix: correct description
hahuyhoang411 May 13, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion extensions/inference-nitro-extension/package.json
Original file line number Diff line number Diff line change
@@ -1,7 1,7 @@
{
"name": "@janhq/inference-nitro-extension",
"productName": "Nitro Inference Engine",
"version": "1.0.5",
"version": "1.0.6",
"description": "This extension embeds Nitro, a lightweight (3mb) inference engine written in C . See https://nitro.jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
"main": "dist/index.js",
"node": "dist/node/index.cjs.js",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 8192,
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
"llama_model_path": "codeninja-1.0-openchat-7b.Q4_K_M.gguf"
"llama_model_path": "codeninja-1.0-openchat-7b.Q4_K_M.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 131072,
"prompt_template": "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",
"llama_model_path": "c4ai-command-r-v01-Q4_K_M.gguf"
"llama_model_path": "c4ai-command-r-v01-Q4_K_M.gguf",
"ngl": 40
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 16384,
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
"llama_model_path": "deepseek-coder-1.3b-instruct.Q8_0.gguf"
"llama_model_path": "deepseek-coder-1.3b-instruct.Q8_0.gguf",
"ngl": 24
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 16384,
"prompt_template": "### Instruction:\n{prompt}\n### Response:",
"llama_model_path": "deepseek-coder-33b-instruct.Q4_K_M.gguf"
"llama_model_path": "deepseek-coder-33b-instruct.Q4_K_M.gguf",
"ngl": 62
},
"parameters": {
"temperature": 0.7,
Expand Down

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 8192,
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
"llama_model_path": "gemma-2b-it-q4_k_m.gguf"
"llama_model_path": "gemma-2b-it-q4_k_m.gguf",
"ngl": 18
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 8192,
"prompt_template": "<start_of_turn>user\n{prompt}<end_of_turn>\n<start_of_turn>model",
"llama_model_path": "gemma-7b-it-q4_K_M.gguf"
"llama_model_path": "gemma-7b-it-q4_K_M.gguf",
"ngl": 28
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 4096,
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
"llama_model_path": "llama-2-70b-chat.Q4_K_M.gguf"
"llama_model_path": "llama-2-70b-chat.Q4_K_M.gguf",
"ngl": 80
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 4096,
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]",
"llama_model_path": "llama-2-7b-chat.Q4_K_M.gguf"
"llama_model_path": "llama-2-7b-chat.Q4_K_M.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,19 8,20 @@
"id": "llama3-8b-instruct",
"object": "model",
"name": "Llama 3 8B Q4",
"version": "1.0",
"version": "1.1",
"description": "Meta's Llama 3 excels at general usage situations, including chat, general world knowledge, and coding.",
"format": "gguf",
"settings": {
"ctx_len": 8192,
"prompt_template": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"llama_model_path": "Meta-Llama-3-8B-Instruct-Q4_K_M.gguf"
"llama_model_path": "Meta-Llama-3-8B-Instruct-Q4_K_M.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"max_tokens": 8192,
"stop": ["<|end_of_text|>","<|eot_id|>"],
"frequency_penalty": 0,
"presence_penalty": 0
Expand Down
Original file line number Diff line number Diff line change
@@ -1,35 1,38 @@
{
"sources": [
{
"filename": "Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf",
"url": "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B-GGUF/resolve/main/Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf"
"filename": "Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf",
"url": "https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF/resolve/main/Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf"
}
],
"id": "hermes-pro-7b",
"id": "llama3-hermes-8b",
"object": "model",
"name": "Hermes Pro 7B Q4",
"name": "Hermes Pro Llama 3 8B Q4",
"version": "1.1",
"description": "Hermes Pro is superior in Roleplaying, Reasoning and Explaining problem.",
"description": "Hermes Pro is well-designed for General chat and JSON output.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"ctx_len": 8192,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "Hermes-2-Pro-Mistral-7B.Q4_K_M.gguf"
"llama_model_path": "Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"max_tokens": 8192,
"stop": [],
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "NousResearch",
"tags": ["7B", "Finetuned"],
"size": 4370000000
"tags": [
"7B",
"Finetuned"
],
"size": 4920000000
},
"engine": "nitro"
}

Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 2048,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "llamacorn-1.1b-chat.Q8_0.gguf"
"llama_model_path": "llamacorn-1.1b-chat.Q8_0.gguf",
"ngl": 22
},
"parameters": {
"temperature": 0.7,
Expand Down

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 32768,
"prompt_template": "[INST] {prompt} [/INST]",
"llama_model_path": "mistral-7b-instruct-v0.2.Q4_K_M.gguf"
"llama_model_path": "mistral-7b-instruct-v0.2.Q4_K_M.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 32768,
"prompt_template": "[INST] {prompt} [/INST]",
"llama_model_path": "mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf"
"llama_model_path": "mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf",
"ngl": 100
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 32768,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "Noromaid-7B-0.4-DPO.q4_k_m.gguf"
"llama_model_path": "Noromaid-7B-0.4-DPO.q4_k_m.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 8192,
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:",
"llama_model_path": "openchat-3.5-0106.Q4_K_M.gguf"
"llama_model_path": "openchat-3.5-0106.Q4_K_M.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down

This file was deleted.

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 16384,
"prompt_template": "### System Prompt\n{system_message}\n### User Message\n{prompt}\n### Assistant",
"llama_model_path": "phind-codellama-34b-v2.Q4_K_M.gguf"
"llama_model_path": "phind-codellama-34b-v2.Q4_K_M.gguf",
"ngl": 48
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 32768,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "qwen1_5-7b-chat-q4_k_m.gguf"
"llama_model_path": "qwen1_5-7b-chat-q4_k_m.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 4096,
"prompt_template": "<|user|>\n{prompt}<|endoftext|>\n<|assistant|>",
"llama_model_path": "stablelm-zephyr-3b.Q8_0.gguf"
"llama_model_path": "stablelm-zephyr-3b.Q8_0.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 32768,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant",
"llama_model_path": "stealth-v1.3.Q4_K_M.gguf"
"llama_model_path": "stealth-v1.3.Q4_K_M.gguf",
"ngl": 32
},
"parameters": {
"temperature": 0.7,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 14,8 @@
"settings": {
"ctx_len": 4096,
"prompt_template": "<|system|>\n{system_message}<|user|>\n{prompt}<|assistant|>",
"llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf"
"llama_model_path": "tinyllama-1.1b-chat-v1.0.Q4_K_M.gguf",
"ngl": 22
},
"parameters": {
"temperature": 0.7,
Expand Down
Loading
Loading