models.json 4.3 KB
Newer Older
Jeffrey Morgan's avatar
Jeffrey Morgan committed
1
2
[
  {
Michael Chiang's avatar
Michael Chiang committed
3
    "name": "orca_mini_3b",
Michael Chiang's avatar
Michael Chiang committed
4
5
6
7
8
9
10
    "display_name": "Orca Mini 3B",
    "url": "huggingface.co/TheBloke/orca_mini_3B-GGML",
    "short_description": "Follow instructions. Great small model that runs fast even without GPU support.",
    "description": "An OpenLLaMa-3B model trained on explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying Orca Research Paper dataset construction approaches.",
    "published_by": "TheBloke",
    "original_author": "psmathur",
    "original_url": "https://huggingface.co/psmathur/orca_mini_3b",
Bruce MacDonald's avatar
Bruce MacDonald committed
11
    "license": "CC-BY-SA-4.0"
Jeffrey Morgan's avatar
Jeffrey Morgan committed
12
13
  },
  {
Michael Chiang's avatar
Michael Chiang committed
14
    "name": "orca_mini_7b", 
Michael Chiang's avatar
Michael Chiang committed
15
16
17
18
19
20
21
    "display_name": "Orca Mini 7B",
    "url": "huggingface.co/TheBloke/orca_mini_7B-GGML",
    "short_description": "Follow instructions",
    "description": "An OpenLLaMa-7B model trained on explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying Orca Research Paper dataset construction approaches.",
    "published_by": "TheBloke",
    "original_author": "psmathur",
    "original_url": "https://huggingface.co/psmathur/orca_mini_7b",
Bruce MacDonald's avatar
Bruce MacDonald committed
22
    "license": "CC-BY-SA-4.0"
Jeffrey Morgan's avatar
Jeffrey Morgan committed
23
24
  },
  {
Michael Chiang's avatar
Michael Chiang committed
25
    "name": "orca_mini_13b",
Michael Chiang's avatar
Michael Chiang committed
26
27
28
29
30
31
32
    "display_name": "Orca Mini 13B",
    "url": "huggingface.co/TheBloke/orca_mini_13B-GGML",
    "short_description": "Follow instructions",
    "description": "An OpenLLaMa-7B model trained on explain tuned datasets, created using Instructions and Input from WizardLM, Alpaca & Dolly-V2 datasets and applying Orca Research Paper dataset construction approaches.",
    "published_by": "TheBloke",
    "original_author": "psmathur",
    "original_url": "https://huggingface.co/psmathur/orca_mini_13b",
Bruce MacDonald's avatar
Bruce MacDonald committed
33
    "license": "CC-BY-SA-4.0"
Michael Chiang's avatar
Michael Chiang committed
34
35
  },
  {
Michael Chiang's avatar
Michael Chiang committed
36
    "name": "replit_code_3b", 
Michael Chiang's avatar
Michael Chiang committed
37
38
39
40
41
42
43
    "display_name": "Replit Code V1 3B", 
    "url": "https://huggingface.co/nomic-ai/ggml-replit-code-v1-3b", 
    "short_description": "Code Completion",
    "description": "This model focuses on code completion. The training mixture includes 20 different languages, listed here in descending order of number of tokens: Markdown, Java, JavaScript, Python, TypeScript, PHP, SQL, JSX, reStructuredText, Rust, C, CSS, Go, C++, HTML, Vue, Ruby, Jupyter Notebook, R, and Shell. This model binary is converted by Nomic AI with the original Replit model code before it was refactored to use MPT configurations.",
    "published_by": "Nomic AI", 
    "original_author": "Replit, Inc.",
    "original_url": "https://huggingface.co/replit/replit-code-v1-3b",
Bruce MacDonald's avatar
Bruce MacDonald committed
44
    "license": "CC-BY-SA-4.0"
Michael Chiang's avatar
Michael Chiang committed
45
46
  }, 
  {
Michael Chiang's avatar
Michael Chiang committed
47
    "name": "nous_hermes_13b", 
Michael Chiang's avatar
Michael Chiang committed
48
49
50
51
52
53
54
    "display_name": "Nous Hermes 13B", 
    "url": "https://huggingface.co/TheBloke/Nous-Hermes-13B-GGML",
    "short_description": "Currently one of the best 13B general model.",
    "description": "It is suitable for a wide range of language tasks, from generating creative text to understanding and following complex instructions. This model was fine-tuned by Nous Research, with Teknium and Karan4D leading the fine tuning process and dataset curation, Redmond AI sponsoring the compute, and several other contributors. The result is an enhanced Llama 13b model that rivals GPT-3.5-turbo in performance across a variety of tasks. \n \n This model stands out for its long responses, low hallucination rate, and absence of OpenAI censorship mechanisms. The fine-tuning process was performed with a 2000 sequence length on an 8x a100 80GB DGX machine for over 50 hours.",
    "published_by": "TheBloke",
    "original_author": "NousResearch",
    "original_url": "https://huggingface.co/NousResearch/Nous-Hermes-13b",
Bruce MacDonald's avatar
Bruce MacDonald committed
55
    "license": "GPL"
Michael Chiang's avatar
Michael Chiang committed
56
57
  }, 
  {
Michael Chiang's avatar
Michael Chiang committed
58
    "name": "wizard_vicuna_13b_uncensored", 
Michael Chiang's avatar
Michael Chiang committed
59
60
61
62
63
64
65
66
    "display_name": "Wizard Vicuna 13B Uncensored", 
    "url": "https://huggingface.co/TheBloke/Wizard-Vicuna-13B-Uncensored-GGML", 
    "short_description": "An uncensored model with no guardrails.",
    "description": "This model is trained with a subset of the dataset - responses that contained alignment / moralizing were removed. The intent is to train a WizardLM that doesn't have alignment built-in, so that alignment (of any sort) can be added separately with for example with a RLHF LoRA.",
    "published_by":"TheBloke" ,
    "original_author": "ehartford", 
    "original_url": "https://huggingface.co/ehartford/Wizard-Vicuna-13B-Uncensored", 
    "license:": "GPL"
Jeffrey Morgan's avatar
Jeffrey Morgan committed
67
68
  }
]