Models LibraryDiscover amazing Large Language Models by the gaia community!
All LLMs
0.5B|1.5B|7B|72B
Qwen2
Qwen2 is a new series of large language models from Alibaba group. This is the base 0.5B parameter model.
34,928Nodes
1.6B
StableLM2
Stable LM 2 1.6B is a state-of-the-art 1.6 billion parameter small language model trained on multilingual data in English, Spanish, German, Italian, French, Portuguese, and Dutch.
186Nodes
8B|70B
Llama3
Meta released the Meta Llama 3 family of large language models (LLMs), including 8B-sized and 70B-sized pretrained and instruction-tuned generative text models. Optimized for dialogue, Llama 3 outperforms many open-source chat models on industry benchmarks, emphasizing helpfulness and safety.
487Nodes
0.5B|1.8B|14B|4B|7B
Qwen1.5
Qwen1.5-0.5B is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. It's the smallest model in the Qwen 1.5 series of models.
146,718Nodes
4B|14B|3.8B
Phi3
Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.
21,156Nodes
6B|9B|34B
Yi-1.5
The Yi series models are large language models trained from scratch by developers at 01.AI. This is the base 9B parameter model.
25Nodes
8B
Llama3.1
The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes (text in/text out).
78Nodes
9B|27B
Gemma2
Gemma 2 is the latest family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
139Nodes
22B
Codestral
Codestrall-22B-v0.1 is trained on a diverse dataset of 80+ programming languages, including the most popular ones, such as Python, Java, C, C++, JavaScript, and Bash.
32Nodes
7B
CodeGemma
CodeGemma-7B-it is a lightweight open code models built on top of Gemma, which specializes in code completion and code generation tasks.
11Nodes
3B
Llama 3.2
The Llama 3.2 3B models support context length of 128K tokens and are state-of-the-art in their class for on-device use cases like summarization, instruction following, and rewriting tasks running locally at the edge.
4Nodes