Models LibraryDiscover amazing Large Language Models by the gaia community!
All LLMs
Codestral
Codestral is Mistral AI's 22B code model, proficient in over 80 programming languages. It can generate code, write tests and complete partial code for tasks such as Java, Python or C++.
DeepSeekR1DistillLlama
DeepSeek-R1-Zero, a reinforcement learning model, shows remarkable reasoning performance but faces challenges. DeepSeek-R1 addresses these issues and matches OpenAI-o1's performance.
gemma2
A family of lightweight open language models from Google, based on Gemini technology. These efficient models excel at text generation tasks while being small enough to run on personal computers.
Llama3.1
The Meta Llama 3.1 collection includes pre-trained and instruction-tuned generative models in 8B, 70B, and 405B sizes. These text-only models excel in multilingual dialogue applications, outperforming many open-source and closed-source chat models on industry benchmarks.
Llama3.2
The Meta Llama 3.2 collection features pre-trained and instruction-tuned generative models in 1B and 3B sizes. These text-only models excel in multilingual dialogue applications, including agentic retrieval and summarization tasks, and outperform many open-source and closed-source chat models on industry benchmarks.
Llama3.3
The Meta Llama 3.3 model is a 70B-sized, instruction-tuned LLM optimized for multilingual dialogue applications. It outperforms many open-source and closed-source chat models on industry benchmarks, making it a top choice for dialogue use cases.
Phi3.5
This model, Phi-3.5-mini, is a compact version of the advanced Phi-3 family, trained on high-quality data and focusing on complex reasoning.
Phi4
Phi-4 is a cutting-edge model combining high-quality data from books, websites and Q&A datasets for advanced reasoning. It was refined through supervised fine-tuning and safety optimization.
Qwen2.5
The latest series of Code-Specific Qwen models, delivering significant improvements in code generation, reasoning, and fixing capabilities.
Qwen2.5Coder
The latest series of Code-Specific Qwen models, delivering significant improvements in code generation, reasoning, and fixing capabilities.
Yi1.5
Yi-1.5 is an improved version of Yi, trained on 500B tokens and fine-tuned on 3M examples. It exceeds Yi in coding, math, reasoning, and following instructions, while maintaining strong language understanding and reading skills.
YiCoder
An efficient open-source code language model delivering state-of-the-art performance under 10B parameters, with 128K token context length and support for 52 programming languages.