Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Chore: Update model.json for UI #1448

Merged
merged 1 commit into from
Jan 8, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion extensions/model-extension/package.json
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
{
"name": "@janhq/model-extension",
"version": "1.0.19",
"version": "1.0.20",
"description": "Model Management Extension provides model exploration and seamless downloads",
"main": "dist/index.js",
"module": "dist/module.js",
Expand Down
9 changes: 7 additions & 2 deletions models/capybara-34b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "Capybara 200k 34B Q5",
"version": "1.0",
"description": "Nous Capybara 34B, a variant of the Yi-34B model, is the first Nous model with a 200K context length, trained for three epochs on the innovative Capybara dataset.",
"description": "Nous Capybara 34B is a long context length model that supports 200K tokens.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "USER:\n{prompt}\nASSISTANT:"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "NousResearch, The Bloke",
Expand Down
9 changes: 7 additions & 2 deletions models/codeninja-1.0-7b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "CodeNinja 7B Q4",
"version": "1.0",
"description": "CodeNinja is finetuned on openchat/openchat-3.5-1210. It is good for codding tasks",
"description": "CodeNinja is good for coding tasks and can handle various languages including Python, C, C++, Rust, Java, JavaScript, and more.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Beowolx",
Expand Down
9 changes: 7 additions & 2 deletions models/deepseek-coder-1.3b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -5,14 +5,19 @@
"object": "model",
"name": "Deepseek Coder 1.3B Q8",
"version": "1.0",
"description": "Deepseek Coder trained on 2T tokens (87% code, 13% English/Chinese), excelling in project-level code completion with advanced capabilities across multiple programming languages.",
"description": "Deepseek Coder excelled in project-level code completion with advanced capabilities across multiple programming languages.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "### Instruction:\n{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Deepseek, The Bloke",
Expand Down
11 changes: 8 additions & 3 deletions models/deepseek-coder-34b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,19 +4,24 @@
"object": "model",
"name": "Deepseek Coder 33B Q5",
"version": "1.0",
"description": "Deepseek Coder trained on 2T tokens (87% code, 13% English/Chinese), excelling in project-level code completion with advanced capabilities across multiple programming languages.",
"description": "Deepseek Coder excelled in project-level code completion with advanced capabilities across multiple programming languages.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "### Instruction:\n{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Deepseek, The Bloke",
"tags": ["34B", "Foundational Model"],
"size": 26040000000
"size": 19940000000
},
"engine": "nitro"
}
Expand Down
9 changes: 7 additions & 2 deletions models/dolphin-2.7-mixtral-8x7b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "Dolphin 8x7B Q4",
"version": "1.0",
"description": "This model is an uncensored model based on Mixtral-8x7b. Dolphin is really good at coding",
"description": "Dolphin is an uncensored model built on Mixtral-8x7b. It is good at programming tasks.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Cognitive Computations, TheBloke",
Expand Down
9 changes: 7 additions & 2 deletions models/llama2-chat-70b-q4/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "Llama 2 Chat 70B Q4",
"version": "1.0",
"description": "This is a 4-bit quantized version of Meta AI's Llama 2 Chat 70b model.",
"description": "Llama 2 Chat 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "MetaAI, The Bloke",
Expand Down
9 changes: 7 additions & 2 deletions models/llama2-chat-7b-q4/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "Llama 2 Chat 7B Q4",
"version": "1.0",
"description": "This is a 4-bit quantized iteration of Meta AI's Llama 2 Chat 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"description": "Llama 2 Chat 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "[INST] <<SYS>>\n{system_message}<</SYS>>\n{prompt}[/INST]"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "MetaAI, The Bloke",
Expand Down
23 changes: 0 additions & 23 deletions models/magicoder-s-ds-7b/model.json

This file was deleted.

11 changes: 8 additions & 3 deletions models/mistral-ins-7b-q4/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "Mistral Instruct 7B Q4",
"version": "1.0",
"description": "This is a 4-bit quantized iteration of MistralAI's Mistral Instruct 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"description": "Mistral Instruct 7b model, specifically designed for a comprehensive understanding through training on extensive internet data.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "<s>[INST]{prompt}\n[/INST]"
"prompt_template": "[INST] {prompt} [/INST]"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "MistralAI, The Bloke",
Expand Down
9 changes: 7 additions & 2 deletions models/mixtral-8x7b-instruct/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "Mixtral 8x7B Instruct Q4",
"version": "1.0",
"description": "The Mixtral-8x7B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. The Mixtral-8x7B outperforms Llama 2 70B on most benchmarks we tested.",
"description": "The Mixtral-8x7B is a pretrained generative Sparse Mixture of Experts. The Mixtral-8x7B outperforms 70B models on most benchmarks.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "[INST] {prompt} [/INST]"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "MistralAI, TheBloke",
Expand Down
11 changes: 8 additions & 3 deletions models/noromaid-7b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,19 +4,24 @@
"object": "model",
"name": "Noromaid 7B Q5",
"version": "1.0",
"description": "The Noromaid 7b model is designed for role-playing and general use, featuring a unique touch with the no_robots dataset that enhances human-like behavior.",
"description": "The Noromaid 7b model is designed for role-playing with human-like behavior.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "### Instruction:{prompt}\n### Response:"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "NeverSleep",
"tags": ["7B", "Merged"],
"size": 5130000000
"size": 4370000000
},
"engine": "nitro"
}
Expand Down
9 changes: 7 additions & 2 deletions models/openchat-3.5-7b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -11,11 +11,16 @@
"prompt_template": "GPT4 Correct User: {prompt}<|end_of_turn|>GPT4 Correct Assistant:"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Openchat",
"tags": ["7B", "Finetuned"],
"tags": ["Recommended", "7B", "Finetuned"],
"size": 4370000000
},
"engine": "nitro"
Expand Down
9 changes: 7 additions & 2 deletions models/openhermes-neural-7b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,19 @@
"object": "model",
"name": "OpenHermes Neural 7B Q4",
"version": "1.0",
"description": "OpenHermes Neural is a merged model using the TIES method.",
"description": "OpenHermes Neural is a merged model using the TIES method. It performs well in various benchmarks.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "<|im_start|>system\n{system_message}<|im_end|>\n<|im_start|>user\n{prompt}<|im_end|>\n<|im_start|>assistant"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Intel, Jan",
Expand Down
Binary file removed models/pandora-10.7b-v1/cover.png
Binary file not shown.
23 changes: 0 additions & 23 deletions models/pandora-10.7b-v1/model.json

This file was deleted.

13 changes: 9 additions & 4 deletions models/phi-2-3b/model.json
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
{
"source_url": "https://huggingface.co/TheBloke/phi-2-GGUF/resolve/main/phi-2.Q4_K_M.gguf",
"source_url": "https://huggingface.co/TheBloke/phi-2-GGUF/resolve/main/phi-2.Q8_0.gguf",
"id": "phi-2-3b",
"object": "model",
"name": "Phi-2 3B Q4",
"name": "Phi-2 3B Q8",
"version": "1.0",
"description": "Phi-2 is a 2.7B model, excelling in common sense and logical reasoning benchmarks, trained with synthetic texts and filtered websites.",
"format": "gguf",
Expand All @@ -11,12 +11,17 @@
"prompt_template": "Intruct:\n{prompt}\nOutput:"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Microsoft",
"tags": ["3B","Foundational Model"],
"size": 1790000000
"size": 2960000000
},
"engine": "nitro"
}
Expand Down
11 changes: 8 additions & 3 deletions models/phind-34b/model.json
Original file line number Diff line number Diff line change
Expand Up @@ -4,19 +4,24 @@
"object": "model",
"name": "Phind 34B Q5",
"version": "1.0",
"description": "Phind-CodeLlama-34B-v2 is an AI model fine-tuned on 1.5B tokens of high-quality programming data. It's a SOTA open-source model in coding. This multi-lingual model excels in various programming languages, including Python, C/C++, TypeScript, Java, and is designed to be steerable and user-friendly.",
"description": "Phind 34B is fine-tuned on 1.5B tokens of high-quality programming data. This multi-lingual model excels in various programming languages and is designed to be steerable and user-friendly.",
"format": "gguf",
"settings": {
"ctx_len": 4096,
"prompt_template": "### System Prompt\n{system_message}\n### User Message\n{prompt}\n### Assistant"
},
"parameters": {
"max_tokens": 4096
"temperature": 0.7,
"top_p": 0.95,
"stream": true,
"max_tokens": 4096,
"frequency_penalty": 0,
"presence_penalty": 0
},
"metadata": {
"author": "Phind, The Bloke",
"tags": ["34B", "Finetuned"],
"size": 24320000000
"size": 20220000000
},
"engine": "nitro"
}
Expand Down
Loading