mirror of
https://github.com/nomic-ai/gpt4all.git
synced 2024-10-01 01:06:10 -04:00
144 lines
5.7 KiB
JSON
144 lines
5.7 KiB
JSON
|
[
|
||
|
{
|
||
|
"order": "a",
|
||
|
"md5sum": "5aff90007499bce5c64b1c0760c0b186",
|
||
|
"name": "Wizard v1.2",
|
||
|
"filename": "wizardlm-13b-v1.2.Q4_0.gguf",
|
||
|
"filesize": "7365834624",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "16",
|
||
|
"parameters": "13 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "LLaMA",
|
||
|
"systemPrompt": " ",
|
||
|
"description": "<strong>Best overall model</strong><br><ul><li>Instruction based<li>Gives very long responses<li>Finetuned with only 1k of high-quality data<li>Trained by Microsoft and Peking University<li>Cannot be used commercially</ul",
|
||
|
"url": "https://gpt4all.io/models/gguf/wizardlm-13b-v1.2.Q4_0.gguf"
|
||
|
},
|
||
|
{
|
||
|
"order": "a",
|
||
|
"md5sum": "97463be739b50525df56d33b26b00852",
|
||
|
"name": "Mistral Instruct",
|
||
|
"filename": "mistral-7b-instruct-v0.1.Q4_0.gguf",
|
||
|
"filesize": "4108916384",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "8",
|
||
|
"parameters": "7 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "LLaMA",
|
||
|
"systemPrompt": " ",
|
||
|
"description": " ",
|
||
|
"url": "https://gpt4all.io/models/gguf/mistral-7b-instruct-v0.1.Q4_0.gguf"
|
||
|
},
|
||
|
{
|
||
|
"order": "a",
|
||
|
"md5sum": "48de9538c774188eb25a7e9ee024bbd3",
|
||
|
"name": "Mistral OpenOrca",
|
||
|
"filename": "mistral-7b-openorca.Q4_0.gguf",
|
||
|
"filesize": "4108927744",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "8",
|
||
|
"parameters": "7 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "LLaMA",
|
||
|
"systemPrompt": " ",
|
||
|
"description": " ",
|
||
|
"url": "https://gpt4all.io/models/gguf/mistral-7b-openorca.Q4_0.gguf"
|
||
|
},
|
||
|
{
|
||
|
"order": "b",
|
||
|
"md5sum": "31cb6d527bd3bfb5e73c2e9dfbc75033",
|
||
|
"name": "GPT4All Falcon",
|
||
|
"filename": "gpt4all-falcon-q4_0.gguf",
|
||
|
"filesize": "4210419040",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "8",
|
||
|
"parameters": "7 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "Falcon",
|
||
|
"systemPrompt": " ",
|
||
|
"description": "<strong>Best overall smaller model</strong><br><ul><li>Fast responses</li><li>Instruction based</li><li>Trained by TII<li>Finetuned by Nomic AI<li>Licensed for commercial use</ul>",
|
||
|
"url": "https://gpt4all.io/models/gguf/gpt4all-falcon-q4_0.gguf",
|
||
|
"promptTemplate": "### Instruction:\n%1\n### Response:\n"
|
||
|
},
|
||
|
{
|
||
|
"order": "c",
|
||
|
"md5sum": "3d12810391d04d1153b692626c0c6e16",
|
||
|
"name": "Hermes",
|
||
|
"filename": "nous-hermes-llama2-13b.Q4_0.gguf",
|
||
|
"filesize": "7366062080",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "16",
|
||
|
"parameters": "13 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "LLaMA",
|
||
|
"systemPrompt": " ",
|
||
|
"description": "<strong>Extremely good model</strong><br><ul><li>Instruction based<li>Gives long responses<li>Curated with 300,000 uncensored instructions<li>Trained by Nous Research<li>Cannot be used commercially</ul>",
|
||
|
"url": "https://gpt4all.io/models/gguf/nous-hermes-llama2-13b.Q4_0.gguf",
|
||
|
"promptTemplate": "### Instruction:\n%1\n### Response:\n"
|
||
|
},
|
||
|
{
|
||
|
"order": "f",
|
||
|
"md5sum": "40388eb2f8d16bb5d08c96fdfaac6b2c",
|
||
|
"name": "Snoozy",
|
||
|
"filename": "gpt4all-13b-snoozy-q4_0.gguf",
|
||
|
"filesize": "7365834624",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "16",
|
||
|
"parameters": "13 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "LLaMA",
|
||
|
"systemPrompt": " ",
|
||
|
"description": "<strong>Very good overall model</strong><br><ul><li>Instruction based<li>Based on the same dataset as Groovy<li>Slower than Groovy, with higher quality responses<li>Trained by Nomic AI<li>Cannot be used commercially</ul>",
|
||
|
"url": "https://gpt4all.io/models/gguf/gpt4all-13b-snoozy-q4_0.gguf"
|
||
|
},
|
||
|
{
|
||
|
"order": "g",
|
||
|
"md5sum": "f5bc6a52f72efd9128efb2eeed802c86",
|
||
|
"name": "MPT Chat",
|
||
|
"filename": "mpt-7b-chat-q4_0.gguf",
|
||
|
"filesize": "3911522272",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "8",
|
||
|
"parameters": "7 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "MPT",
|
||
|
"description": "<strong>Best overall smaller model</strong><br><ul><li>Fast responses<li>Chat based<li>Trained by Mosaic ML<li>Cannot be used commercially</ul>",
|
||
|
"url": "https://gpt4all.io/models/gguf/mpt-7b-chat-q4_0.gguf",
|
||
|
"promptTemplate": "<|im_start|>user\n%1<|im_end|><|im_start|>assistant\n",
|
||
|
"systemPrompt": "<|im_start|>system\n- You are a helpful assistant chatbot trained by MosaicML.\n- You answer questions.\n- You are excited to be able to help the user, but will refuse to do anything that could be considered harmful to the user.\n- You are more than just an information source, you are also able to write poetry, short stories, and make jokes.<|im_end|>"
|
||
|
},
|
||
|
{
|
||
|
"order": "i",
|
||
|
"md5sum": "aae346fe095e60139ca39b3fda4ac7ae",
|
||
|
"name": "Mini Orca (Small)",
|
||
|
"filename": "orca-mini-3b.q4_0.gguf",
|
||
|
"filesize": "1928648352",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "4",
|
||
|
"parameters": "3 billion",
|
||
|
"quant": "q4_0",
|
||
|
"type": "OpenLLaMa",
|
||
|
"description": "<strong>Small version of new model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Licensed for commercial use</ul>",
|
||
|
"url": "https://gpt4all.io/models/gguf/orca-mini-3b.q4_0.gguf",
|
||
|
"promptTemplate": "### User:\n%1\n### Response:\n",
|
||
|
"systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
|
||
|
},
|
||
|
{
|
||
|
"order": "s",
|
||
|
"md5sum": "51c627fac9062e208f9b386f105cbd48",
|
||
|
"disableGUI": "true",
|
||
|
"name": "Replit",
|
||
|
"filename": "replit-code-v1-3b-q4_0.gguf",
|
||
|
"filesize": "1532949760",
|
||
|
"requires": "2.5.0",
|
||
|
"ramrequired": "4",
|
||
|
"parameters": "3 billion",
|
||
|
"quant": "f16",
|
||
|
"type": "Replit",
|
||
|
"systemPrompt": " ",
|
||
|
"promptTemplate": "%1",
|
||
|
"description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based<li>Licensed for commercial use</ul>",
|
||
|
"url": "https://gpt4all.io/models/gguf/replit-code-v1-3b-q4_0.gguf"
|
||
|
}
|
||
|
]
|