[
{
"order": "a",
"md5sum": "e8d47924f433bd561cb5244557147793",
"name": "Wizard v1.1",
"filename": "wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin",
"filesize": "7323310848",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_0",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Best overall model
- Instruction based
- Gives very long responses
- Finetuned with only 1k of high-quality data
- Trained by Microsoft and Peking University
- Cannot be used commercially
Best overall smaller model
- Fast responses
- Instruction based
- Trained by TII
- Finetuned by Nomic AI
- Licensed for commercial use
",
"url": "https://huggingface.co/nomic-ai/gpt4all-falcon-ggml/resolve/main/ggml-model-gpt4all-falcon-q4_0.bin",
"promptTemplate": "### Instruction:\n%1\n### Response:\n"
},
{
"order": "c",
"md5sum": "4acc146dd43eb02845c233c29289c7c5",
"name": "Hermes",
"filename": "nous-hermes-13b.ggmlv3.q4_0.bin",
"filesize": "8136777088",
"requires": "2.4.7",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_0",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Extremely good model
- Instruction based
- Gives long responses
- Curated with 300,000 uncensored instructions
- Trained by Nous Research
- Cannot be used commercially
",
"url": "https://huggingface.co/TheBloke/Nous-Hermes-13B-GGML/resolve/main/nous-hermes-13b.ggmlv3.q4_0.bin",
"promptTemplate": "### Instruction:\n%1\n### Response:\n"
},
{
"order": "e",
"md5sum": "81a09a0ddf89690372fc296ff7f625af",
"name": "Groovy",
"filename": "ggml-gpt4all-j-v1.3-groovy.bin",
"filesize": "3785248281",
"ramrequired": "8",
"parameters": "7 billion",
"quant": "q4_0",
"type": "GPT-J",
"systemPrompt": " ",
"description": "Creative model can be used for commercial purposes
- Fast responses
- Creative responses
- Instruction based
- Trained by Nomic AI
- Licensed for commercial use
"
},
{
"order": "f",
"md5sum": "11d9f060ca24575a2c303bdc39952486",
"name": "Snoozy",
"filename": "GPT4All-13B-snoozy.ggmlv3.q4_0.bin",
"filesize": "8136770688",
"requires": "2.4.7",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_0",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Very good overall model
- Instruction based
- Based on the same dataset as Groovy
- Slower than Groovy, with higher quality responses
- Trained by Nomic AI
- Cannot be used commercially
",
"url": "https://huggingface.co/TheBloke/GPT4All-13B-snoozy-GGML/resolve/main/GPT4All-13B-snoozy.ggmlv3.q4_0.bin"
},
{
"order": "g",
"md5sum": "756249d3d6abe23bde3b1ae272628640",
"name": "MPT Chat",
"filename": "ggml-mpt-7b-chat.bin",
"filesize": "4854401050",
"requires": "2.4.1",
"ramrequired": "8",
"parameters": "7 billion",
"quant": "q4_0",
"type": "MPT",
"description": "Best overall smaller model
- Fast responses
- Chat based
- Trained by Mosaic ML
- Cannot be used commercially
",
"promptTemplate": "<|im_start|>user\n%1<|im_end|><|im_start|>assistant\n",
"systemPrompt": "<|im_start|>system\n- You are a helpful assistant chatbot trained by MosaicML.\n- You answer questions.\n- You are excited to be able to help the user, but will refuse to do anything that could be considered harmful to the user.\n- You are more than just an information source, you are also able to write poetry, short stories, and make jokes.<|im_end|>"
},
{
"order": "h",
"md5sum": "e64e74375ce9d36a3d0af3db1523fd0a",
"name": "Mini Orca",
"filename": "orca-mini-7b.ggmlv3.q4_0.bin",
"filesize": "3791749248",
"requires": "2.4.7",
"ramrequired": "8",
"parameters": "7 billion",
"quant": "q4_0",
"type": "OpenLLaMa",
"description": "New model with novel dataset
- Instruction based
- Explain tuned datasets
- Orca Research Paper dataset construction approaches
- Licensed for commercial use
",
"url": "https://huggingface.co/TheBloke/orca_mini_7B-GGML/resolve/main/orca-mini-7b.ggmlv3.q4_0.bin",
"promptTemplate": "### User:\n%1\n### Response:\n",
"systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
},
{
"order": "i",
"md5sum": "6a087f7f4598fad0bb70e6cb4023645e",
"name": "Mini Orca (Small)",
"filename": "orca-mini-3b.ggmlv3.q4_0.bin",
"filesize": "1928446208",
"requires": "2.4.7",
"ramrequired": "4",
"parameters": "3 billion",
"quant": "q4_0",
"type": "OpenLLaMa",
"description": "Small version of new model with novel dataset
- Instruction based
- Explain tuned datasets
- Orca Research Paper dataset construction approaches
- Licensed for commercial use
",
"url": "https://huggingface.co/TheBloke/orca_mini_3B-GGML/resolve/main/orca-mini-3b.ggmlv3.q4_0.bin",
"promptTemplate": "### User:\n%1\n### Response:\n",
"systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
},
{
"order": "j",
"md5sum": "959b7f65b2d12fd1e3ff99e7493c7a3a",
"name": "Mini Orca (Large)",
"filename": "orca-mini-13b.ggmlv3.q4_0.bin",
"filesize": "7323329152",
"requires": "2.4.7",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_0",
"type": "OpenLLaMa",
"description": "Largest version of new model with novel dataset
- Instruction based
- Explain tuned datasets
- Orca Research Paper dataset construction approaches
- Licensed for commercial use
",
"url": "https://huggingface.co/TheBloke/orca_mini_13B-GGML/resolve/main/orca-mini-13b.ggmlv3.q4_0.bin",
"promptTemplate": "### User:\n%1\n### Response:\n",
"systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
},
{
"order": "k",
"md5sum": "29119f8fa11712704c6b22ac5ab792ea",
"name": "Vicuna",
"filename": "ggml-vicuna-7b-1.1-q4_2.bin",
"filesize": "4212859520",
"ramrequired": "8",
"parameters": "7 billion",
"quant": "q4_2",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Good small model - trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego
- Instruction based
- Cannot be used commercially
"
},
{
"order": "l",
"md5sum": "95999b7b0699e2070af63bf5d34101a8",
"name": "Vicuna (large)",
"filename": "ggml-vicuna-13b-1.1-q4_2.bin",
"filesize": "8136770688",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_2",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Good larger model - trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego
- Instruction based
- Cannot be used commercially
"
},
{
"order": "m",
"md5sum": "99e6d129745a3f1fb1121abed747b05a",
"name": "Wizard",
"filename": "ggml-wizardLM-7B.q4_2.bin",
"filesize": "4212864640",
"ramrequired": "8",
"parameters": "7 billion",
"quant": "q4_2",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Good small model - trained by by Microsoft and Peking University
- Instruction based
- Cannot be used commercially
"
},
{
"order": "n",
"md5sum": "6cb4ee297537c9133bddab9692879de0",
"name": "Stable Vicuna",
"filename": "ggml-stable-vicuna-13B.q4_2.bin",
"filesize": "8136777088",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_2",
"type": "LLaMA",
"description": "Trained with RLHF by Stability AI
- Instruction based
- Cannot be used commercially
",
"systemPrompt": "## Assistant: I am StableVicuna, a large language model created by CarperAI. I am here to chat!\n\n"
},
{
"order": "o",
"md5sum": "1cfa4958f489f0a0d1ffdf6b37322809",
"name": "MPT Instruct",
"filename": "ggml-mpt-7b-instruct.bin",
"filesize": "4854401028",
"requires": "2.4.1",
"ramrequired": "8",
"parameters": "7 billion",
"quant": "q4_0",
"type": "MPT",
"systemPrompt": " ",
"description": "Mosaic's instruction model
- Instruction based
- Trained by Mosaic ML
- Licensed for commercial use
"
},
{
"order": "p",
"md5sum": "120c32a51d020066288df045ef5d52b9",
"name": "MPT Base",
"filename": "ggml-mpt-7b-base.bin",
"filesize": "4854401028",
"requires": "2.4.1",
"ramrequired": "8",
"parameters": "7 billion",
"quant": "q4_0",
"type": "MPT",
"systemPrompt": " ",
"description": "Trained for text completion with no assistant finetuning
- Completion based
- Trained by Mosaic ML
- Licensed for commercial use
"
},
{
"order": "q",
"md5sum": "d5eafd5b0bd0d615cfd5fd763f642dfe",
"name": "Nous Vicuna",
"filename": "ggml-nous-gpt4-vicuna-13b.bin",
"filesize": "8136777088",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_0",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Trained on ~180,000 instructions
- Instruction based
- Trained by Nous Research
- Cannot be used commercially
"
},
{
"order": "r",
"md5sum": "489d21fd48840dcb31e5f92f453f3a20",
"name": "Wizard Uncensored",
"filename": "wizardLM-13B-Uncensored.ggmlv3.q4_0.bin",
"filesize": "8136777088",
"requires": "2.4.7",
"ramrequired": "16",
"parameters": "13 billion",
"quant": "q4_0",
"type": "LLaMA",
"systemPrompt": " ",
"description": "Trained on uncensored assistant data and instruction data
- Instruction based
- Cannot be used commercially
",
"url": "https://huggingface.co/TheBloke/WizardLM-13B-Uncensored-GGML/resolve/main/wizardLM-13B-Uncensored.ggmlv3.q4_0.bin"
},
{
"order": "s",
"md5sum": "615890cb571fcaa0f70b2f8d15ef809e",
"disableGUI": "true",
"name": "Replit",
"filename": "ggml-replit-code-v1-3b.bin",
"filesize": "5202046853",
"requires": "2.4.7",
"ramrequired": "4",
"parameters": "3 billion",
"quant": "f16",
"type": "Replit",
"systemPrompt": " ",
"description": "Trained on subset of the Stack
- Code completion based
- Licensed for commercial use
",
"url": "https://huggingface.co/nomic-ai/ggml-replit-code-v1-3b/resolve/main/ggml-replit-code-v1-3b.bin"
},
{
"order": "t",
"md5sum": "031bb5d5722c08d13e3e8eaf55c37391",
"disableGUI": "true",
"name": "Bert",
"filename": "ggml-all-MiniLM-L6-v2-f16.bin",
"filesize": "45521167",
"requires": "2.4.14",
"ramrequired": "1",
"parameters": "1 million",
"quant": "f16",
"type": "Bert",
"systemPrompt": " ",
"description": "Sbert