Skip to content
This repository has been archived by the owner on Sep 27, 2024. It is now read-only.

Commit

Permalink
Merge pull request #48 from lmstudio-ai/catalog-update-dec23
Browse files Browse the repository at this point in the history
Catalog update
  • Loading branch information
yagil authored Dec 16, 2023
2 parents db32bf6 + 061c4fb commit c9c262c
Show file tree
Hide file tree
Showing 7 changed files with 833 additions and 426 deletions.
964 changes: 540 additions & 424 deletions catalog.json

Large diffs are not rendered by default.

59 changes: 59 additions & 0 deletions models/Mistral-7B-Instruct-v0.2.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,59 @@
{
"_descriptorVersion": "0.0.1",
"datePublished": "2023-12-12T10:12:59",
"name": "Mistral 7B Instruct v0.2",
"description": "The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an improved instruct fine-tuned version of Mistral-7B-Instruct-v0.1. For full details of this model read MistralAI's blog post and paper.",
"author": {
"name": "Mistral AI",
"url": "https://mistral.ai/",
"blurb": "Mistral AI's mission is to spearhead the revolution of open models."
},
"numParameters": "7B",
"resources": {
"canonicalUrl": "https://mistral.ai/news/la-plateforme/",
"paperUrl": "https://arxiv.org/abs/2310.06825",
"downloadUrl": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF"
},
"trainedFor": "chat",
"arch": "mistral",
"files": {
"highlighted": {
"economical": {
"name": "mistral-7b-instruct-v0.2.Q4_K_S.gguf"
},
"most_capable": {
"name": "mistral-7b-instruct-v0.2.Q6_K.gguf"
}
},
"all": [
{
"name": "mistral-7b-instruct-v0.2.Q4_K_S.gguf",
"url": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF/resolve/main/mistral-7b-instruct-v0.2.Q4_K_S.gguf",
"sizeBytes": 4140374304,
"quantization": "Q4_K_S",
"format": "gguf",
"sha256checksum": "1213e19b3e103932fdfdc82e3b6dee765f57ad5756e0f673e7d36514a5b60d0a",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/Mistral-7B-Instruct-v0.2-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF"
},
{
"name": "mistral-7b-instruct-v0.2.Q6_K.gguf",
"url": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF/resolve/main/mistral-7b-instruct-v0.2.Q6_K.gguf",
"sizeBytes": 5942065440,
"quantization": "Q6_K",
"format": "gguf",
"sha256checksum": "a4643671c92f47eb6027d0eff50b9875562e8e172128a4b10b2be250bb4264de",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/Mistral-7B-Instruct-v0.2-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF"
}
]
}
}
58 changes: 58 additions & 0 deletions models/NexusRaven-V2-13B.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
{
"_descriptorVersion": "0.0.1",
"datePublished": "2023-12-11T06:26:58",
"name": "NexusRaven-V2-13B",
"description": "NexusRaven-V2 accepts a list of python functions. These python functions can do anything (e.g. sending GET/POST requests to external APIs). The two requirements include the python function signature and the appropriate docstring to generate the function call. *** Follow NexusRaven's prompting guide found on the model's Hugging Face page. ***",
"author": {
"name": "Nexusflow",
"url": "https://nexusflow.ai",
"blurb": "Nexusflow is democratizing Cyber Intelligence with Generative AI, fully on top of open-source large language models (LLMs)"
},
"numParameters": "13B",
"resources": {
"canonicalUrl": "https://huggingface.co/Nexusflow/NexusRaven-V2-13B",
"downloadUrl": "https://huggingface.co/TheBloke/NexusRaven-V2-13B-GGUF"
},
"trainedFor": "other",
"arch": "llama",
"files": {
"highlighted": {
"economical": {
"name": "nexusraven-v2-13b.Q4_K_S.gguf"
},
"most_capable": {
"name": "nexusraven-v2-13b.Q6_K.gguf"
}
},
"all": [
{
"name": "nexusraven-v2-13b.Q4_K_S.gguf",
"url": "https://huggingface.co/TheBloke/NexusRaven-V2-13B-GGUF/resolve/main/nexusraven-v2-13b.Q4_K_S.gguf",
"sizeBytes": 7414501952,
"quantization": "Q4_K_S",
"format": "gguf",
"sha256checksum": "bc2e1ce9fa064e675690d4c6f2c441d922f24241764241aa013d0ca8a87ecbfe",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/NexusRaven-V2-13B-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/NexusRaven-V2-13B-GGUF"
},
{
"name": "nexusraven-v2-13b.Q6_K.gguf",
"url": "https://huggingface.co/TheBloke/NexusRaven-V2-13B-GGUF/resolve/main/nexusraven-v2-13b.Q6_K.gguf",
"sizeBytes": 10679342592,
"quantization": "Q6_K",
"format": "gguf",
"sha256checksum": "556ae244f4c69c603b7cda762d003d09f68058c671f304c2e011214ce754acb4",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/NexusRaven-V2-13B-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/NexusRaven-V2-13B-GGUF"
}
]
}
}
58 changes: 58 additions & 0 deletions models/OpenHermes-2.5-Mistral-7B.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
{
"_descriptorVersion": "0.0.1",
"datePublished": "2023-10-29T21:27:30",
"name": "OpenHermes 2.5 Mistral 7B",
"description": "OpenHermes 2.5 Mistral 7B is an advanced iteration of the OpenHermes 2 language model, enhanced by training on a significant proportion of code datasets. This additional training improved performance across several benchmarks, notably TruthfulQA, AGIEval, and the GPT4All suite, while slightly decreasing the BigBench score. Notably, the model's ability to handle code-related tasks, measured by the humaneval score, increased from 43% to 50.7%. The training data consisted of one million entries, primarily sourced from GPT-4 outputs and other high-quality open datasets. This data was rigorously filtered and standardized to the ShareGPT format and subsequently processed using ChatML by the axolotl tool.",
"author": {
"name": "Teknium",
"url": "https://twitter.com/Teknium1",
"blurb": "Creator of numerous chart topping fine-tunes and a Co-founder of NousResearch"
},
"numParameters": "7B",
"resources": {
"canonicalUrl": "https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B",
"downloadUrl": "https://huggingface.co/TheBloke/OpenHermes-2.5-Mistral-7B-GGUF"
},
"trainedFor": "chat",
"arch": "mistral",
"files": {
"highlighted": {
"economical": {
"name": "openhermes-2.5-mistral-7b.Q4_K_S.gguf"
},
"most_capable": {
"name": "openhermes-2.5-mistral-7b.Q6_K.gguf"
}
},
"all": [
{
"name": "openhermes-2.5-mistral-7b.Q4_K_S.gguf",
"url": "https://huggingface.co/TheBloke/OpenHermes-2.5-Mistral-7B-GGUF/resolve/main/openhermes-2.5-mistral-7b.Q4_K_S.gguf",
"sizeBytes": 4140385024,
"quantization": "Q4_K_S",
"format": "gguf",
"sha256checksum": "5ae9c3c11ce520a2360dcfca1f4e38392dc0b7a49413ce6695857a5148a71d35",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/OpenHermes-2.5-Mistral-7B-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/OpenHermes-2.5-Mistral-7B-GGUF"
},
{
"name": "openhermes-2.5-mistral-7b.Q6_K.gguf",
"url": "https://huggingface.co/TheBloke/OpenHermes-2.5-Mistral-7B-GGUF/resolve/main/openhermes-2.5-mistral-7b.Q6_K.gguf",
"sizeBytes": 5942078272,
"quantization": "Q6_K",
"format": "gguf",
"sha256checksum": "cd4caa42229e973636e9d4c8db50a89593353c521e0342ca615756ded2b977a2",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/OpenHermes-2.5-Mistral-7B-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/OpenHermes-2.5-Mistral-7B-GGUF"
}
]
}
}
58 changes: 58 additions & 0 deletions models/deepseek-coder-6.7b-instruct.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
{
"_descriptorVersion": "0.0.1",
"datePublished": "2023-10-29T11:30:13",
"name": "Deepseek Coder",
"description": "Deepseek Coder is a collection of code language models with sizes ranging from 1B to 33B parameters, trained on a dataset comprising 2 trillion tokens (87% code, 13% natural language in English and Chinese). It is designed for project-level code completion and infilling, utilizing a 16K token window size and an additional fill-in-the-blank task. The models demonstrate leading performance on several programming benchmarks. The 6.7B parameter variant, deepseek-coder-6.7b-instruct, is fine-tuned on 2 billion tokens of instructional data. The code repository is MIT licensed, and the models support commercial use under the Model License.",
"author": {
"name": "DeepSeek",
"url": "https://huggingface.co/deepseek-ai",
"blurb": "DeepSeek (深度求索), founded in 2023, is a Chinese company dedicated to making AGI a reality"
},
"numParameters": "6.7B",
"resources": {
"canonicalUrl": "https://github.com/deepseek-ai/deepseek-coder",
"downloadUrl": "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-GGUF"
},
"trainedFor": "chat",
"arch": "llama",
"files": {
"highlighted": {
"economical": {
"name": "deepseek-coder-6.7b-instruct.Q4_K_S.gguf"
},
"most_capable": {
"name": "deepseek-coder-6.7b-instruct.Q6_K.gguf"
}
},
"all": [
{
"name": "deepseek-coder-6.7b-instruct.Q4_K_S.gguf",
"url": "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-GGUF/resolve/main/deepseek-coder-6.7b-instruct.Q4_K_S.gguf",
"sizeBytes": 3858751712,
"quantization": "Q4_K_S",
"format": "gguf",
"sha256checksum": "d5d4b757645ce359a52d25584d29f1ff0d89580075edc35d87a20b89e65a5313",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/deepseek-coder-6.7B-instruct-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-GGUF"
},
{
"name": "deepseek-coder-6.7b-instruct.Q6_K.gguf",
"url": "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-GGUF/resolve/main/deepseek-coder-6.7b-instruct.Q6_K.gguf",
"sizeBytes": 5531476192,
"quantization": "Q6_K",
"format": "gguf",
"sha256checksum": "113fba500e4feb1313ce80d72cf381330b51460d265a7719bba626d6a461f9eb",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/deepseek-coder-6.7B-instruct-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-GGUF"
}
]
}
}
58 changes: 58 additions & 0 deletions models/stablelm-zephyr-3b.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
{
"_descriptorVersion": "0.0.1",
"datePublished": "2023-11-21T16:28:30",
"name": "StableLM Zephyr 3B",
"description": "StableLM Zephyr 3B is an English-language, auto-regressive language model with 3 billion parameters, developed by Stability AI. It's an instruction-tuned model influenced by HuggingFace's Zephyr 7B training approach and is built on transformer decoder architecture. It was trained using a mix of public and synthetic datasets, including SFT and Preference Datasets from the HuggingFace Hub with Direct Preference Optimization (DPO). Its performance has been evaluated using the MT Bench and Alpaca Benchmark, achieving a score of 6.64 and a win rate of 76% respectively. For fine-tuning, it utilizes the StabilityAI's stablelm-3b-4e1t model and is available under the StabilityAI Non-Commercial Research Community License. Commercial use requires contacting Stability AI for more information. The model was trained on a Stability AI cluster with 8 nodes, each equipped with 8 A100 80GB GPUs, using internal scripts for SFT steps and HuggingFace's Alignment Handbook scripts for DPO training.",
"author": {
"name": "Stability AI",
"url": "https://stability.ai/",
"blurb": "Stability AI is developing cutting-edge open AI models for Image, Language, Audio, Video, 3D and Biology."
},
"numParameters": "3B",
"resources": {
"canonicalUrl": "https://huggingface.co/stabilityai/stablelm-zephyr-3b",
"downloadUrl": "https://huggingface.co/TheBloke/stablelm-zephyr-3b-GGUF"
},
"trainedFor": "chat",
"arch": "stablelm",
"files": {
"highlighted": {
"economical": {
"name": "stablelm-zephyr-3b.Q4_K_S.gguf"
},
"most_capable": {
"name": "stablelm-zephyr-3b.Q6_K.gguf"
}
},
"all": [
{
"name": "stablelm-zephyr-3b.Q4_K_S.gguf",
"url": "https://huggingface.co/TheBloke/stablelm-zephyr-3b-GGUF/resolve/main/stablelm-zephyr-3b.Q4_K_S.gguf",
"sizeBytes": 1620695488,
"quantization": "Q4_K_S",
"format": "gguf",
"sha256checksum": "748f9fa7b893df8383467c7f28affef3489e20f2da351441b0dd112c43ddb587",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/stablelm-zephyr-3b-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/stablelm-zephyr-3b-GGUF"
},
{
"name": "stablelm-zephyr-3b.Q6_K.gguf",
"url": "https://huggingface.co/TheBloke/stablelm-zephyr-3b-GGUF/resolve/main/stablelm-zephyr-3b.Q6_K.gguf",
"sizeBytes": 2295985088,
"quantization": "Q6_K",
"format": "gguf",
"sha256checksum": "d51685399c77b1dfe2dafa53ac7e6272b414bbc529c0f3bf0bdd15f90559c049",
"publisher": {
"name": "TheBloke",
"socialUrl": "https://twitter.com/TheBlokeAI"
},
"respository": "TheBloke/stablelm-zephyr-3b-GGUF",
"repositoryUrl": "https://huggingface.co/TheBloke/stablelm-zephyr-3b-GGUF"
}
]
}
}
4 changes: 2 additions & 2 deletions schema.json
Original file line number Diff line number Diff line change
Expand Up @@ -51,15 +51,15 @@
},
"numParameters": {
"type": "string",
"enum": ["3B", "7B", "13B", "15B", "30B", "65B", "unknown"]
"enum": ["3B", "6.7B", "7B", "13B", "15B", "30B", "65B", "unknown"]
},
"trainedFor": {
"type": "string",
"enum": ["chat", "instruct", "code_completion", "other"]
},
"arch": {
"type": "string",
"enum": ["llama", "pythia", "gpt-neo-x", "gpt-j", "mpt", "replit", "starcoder", "falcon", "mistral"]
"enum": ["llama", "pythia", "gpt-neo-x", "gpt-j", "mpt", "replit", "starcoder", "falcon", "mistral", "stablelm"]
},
"description": {
"type": "string"
Expand Down

0 comments on commit c9c262c

Please sign in to comment.