diff --git a/gpt4all-chat/metadata/models.json b/gpt4all-chat/metadata/models.json index 3fe931fe..15854e03 100644 --- a/gpt4all-chat/metadata/models.json +++ b/gpt4all-chat/metadata/models.json @@ -54,6 +54,48 @@ }, { "order": "e", + "md5sum": "e64e74375ce9d36a3d0af3db1523fd0a", + "name": "Orca", + "filename": "orca-mini-7b.ggmlv3.q4_0.bin", + "filesize": "3791749248", + "requires": "2.4.7", + "ramrequired": "8", + "parameters": "7 billion", + "quant": "q4_0", + "type": "OpenLLaMa", + "description": "New model with novel dataset
", + "url": "https://huggingface.co/TheBloke/orca_mini_7B-GGML/resolve/main/orca-mini-7b.ggmlv3.q4_0.bin" + }, + { + "order": "f", + "md5sum": "6a087f7f4598fad0bb70e6cb4023645e", + "name": "Orca (Small)", + "filename": "orca-mini-3b.ggmlv3.q4_0.bin", + "filesize": "1928446208", + "requires": "2.4.7", + "ramrequired": "4", + "parameters": "3 billion", + "quant": "q4_0", + "type": "OpenLLaMa", + "description": "Small version of new model with novel dataset
", + "url": "https://huggingface.co/TheBloke/orca_mini_3B-GGML/resolve/main/orca-mini-3b.ggmlv3.q4_0.bin" + }, + { + "order": "g", + "md5sum": "959b7f65b2d12fd1e3ff99e7493c7a3a", + "name": "Orca (Large)", + "filename": "orca-mini-13b.ggmlv3.q4_0.bin", + "filesize": "7323329152", + "requires": "2.4.7", + "ramrequired": "16", + "parameters": "13 billion", + "quant": "q4_0", + "type": "OpenLLaMa", + "description": "Largest version of new model with novel dataset
", + "url": "https://huggingface.co/TheBloke/orca_mini_13B-GGML/resolve/main/orca-mini-13b.ggmlv3.q4_0.bin" + }, + { + "order": "h", "md5sum": "29119f8fa11712704c6b22ac5ab792ea", "name": "Vicuna", "filename": "ggml-vicuna-7b-1.1-q4_2.bin", @@ -65,7 +107,7 @@ "description": "Good small model - trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego
" }, { - "order": "f", + "order": "i", "md5sum": "95999b7b0699e2070af63bf5d34101a8", "name": "Vicuna (large)", "filename": "ggml-vicuna-13b-1.1-q4_2.bin", @@ -77,7 +119,7 @@ "description": "Good larger model - trained by teams from UC Berkeley, CMU, Stanford, MBZUAI, and UC San Diego
" }, { - "order": "g", + "order": "j", "md5sum": "99e6d129745a3f1fb1121abed747b05a", "name": "Wizard", "filename": "ggml-wizardLM-7B.q4_2.bin", @@ -89,7 +131,7 @@ "description": "Good small model - trained by by Microsoft and Peking University
" }, { - "order": "h", + "order": "k", "md5sum": "6cb4ee297537c9133bddab9692879de0", "name": "Stable Vicuna", "filename": "ggml-stable-vicuna-13B.q4_2.bin", @@ -101,7 +143,7 @@ "description": "Trained with RHLF by Stability AI
" }, { - "order": "i", + "order": "l", "md5sum": "1cfa4958f489f0a0d1ffdf6b37322809", "name": "MPT Instruct", "filename": "ggml-mpt-7b-instruct.bin", @@ -114,7 +156,7 @@ "description": "Mosaic's instruction model
" }, { - "order": "j", + "order": "m", "md5sum": "120c32a51d020066288df045ef5d52b9", "name": "MPT Base", "filename": "ggml-mpt-7b-base.bin", @@ -127,7 +169,7 @@ "description": "Trained for text completion with no assistant finetuning
" }, { - "order": "k", + "order": "n", "md5sum": "d5eafd5b0bd0d615cfd5fd763f642dfe", "name": "Nous Vicuna", "filename": "ggml-nous-gpt4-vicuna-13b.bin", @@ -139,7 +181,7 @@ "description": "Trained on ~180,000 instructions
" }, { - "order": "l", + "order": "o", "md5sum": "489d21fd48840dcb31e5f92f453f3a20", "name": "Wizard Uncensored", "filename": "wizardLM-13B-Uncensored.ggmlv3.q4_0.bin", @@ -153,7 +195,7 @@ "url": "https://huggingface.co/TheBloke/WizardLM-13B-Uncensored-GGML/resolve/main/wizardLM-13B-Uncensored.ggmlv3.q4_0.bin" }, { - "order": "m", + "order": "p", "md5sum": "615890cb571fcaa0f70b2f8d15ef809e", "disableGUI": "true", "name": "Replit",