Llama 3.2 This collection hosts the transformers and original repos of the Llama 3.2 and Llama Guard 3 Collection by meta-llama 17 days ago 355 meta-llama/Llama-3.2-1B Text Generation • Updated 13 days ago • 247k • • 459 meta-llama/Llama-3.2-3B Text Generation • Updated 16 days ago • 82k • 187 meta-llama/Llama-3.2-1B-Instruct Text Generation • Updated 17 days ago • 320k • • 335 meta-llama/Llama-3.2-3B-Instruct Text Generation • Updated 17 days ago • 374k • • 319
Molmo Artifacts for open multimodal language models. Collection by allenai 17 days ago 245 allenai/Molmo-72B-0924 Image-Text-to-Text • Updated 2 days ago • 4.15k • 228 allenai/Molmo-7B-D-0924 Image-Text-to-Text • Updated 2 days ago • 32.7k • 353 allenai/Molmo-7B-O-0924 Image-Text-to-Text • Updated 2 days ago • 5.01k • 128 allenai/MolmoE-1B-0924 Image-Text-to-Text • Updated 2 days ago • 11.3k • 102
Qwen2.5 Qwen2.5 language models, including pretrained and instruction-tuned models of 7 sizes, including 0.5B, 1.5B, 3B, 7B, 14B, 32B, and 72B. Collection by Qwen 24 days ago 245 Running 353 🚀 Qwen2.5 Qwen/Qwen2.5-0.5B Text Generation • Updated 18 days ago • 93.6k • 71 Qwen/Qwen2.5-0.5B-Instruct Text Generation • Updated 18 days ago • 124k • 64 Qwen/Qwen2.5-1.5B Text Generation • Updated 5 days ago • 26.4k • 30
NVLM 1.0 A family of frontier-class multimodal large language models (LLMs) that achieve state-of-the-art results on vision-language tasks and text-only tasks. Collection by nvidia 12 days ago 38 nvidia/NVLM-D-72B Image-Text-to-Text • Updated 4 days ago • 22.4k • 611
Sapiens Foundation models for human tasks. Code: https://github.com/facebookresearch/sapiens Collection by facebook 24 days ago 38 Sapiens: Foundation for Human Vision Models Paper • 2408.12569 • Published Aug 22 • 86 facebook/sapiens Updated 23 days ago • 72 • 210 Running on Zero 19 📊 Sapiens Pose Running on Zero 94 🌍 Sapiens Segmentation
LLaVA-Video Models focus on video understanding (previously known as LLaVA-NeXT-Video). Collection by lmms-lab 8 days ago 47 Video Instruction Tuning With Synthetic Data Paper • 2410.02713 • Published 9 days ago • 33 lmms-lab/LLaVA-Video-178K Viewer • Updated 2 days ago • 1.63M • 538 • 49 lmms-lab/LLaVA-Video-7B-Qwen2 Video-Text-to-Text • Updated 3 days ago • 56.4k • 18 lmms-lab/LLaVA-Video-72B-Qwen2 Text Generation • Updated 6 days ago • 1.03k • 8
Salamandra 🦎 Collection by BSC-LT 12 days ago 28 BSC-LT/salamandra-2b Text Generation • Updated 3 days ago • 558 • 16 BSC-LT/salamandra-7b-instruct Text Generation • Updated 3 days ago • 1.74k • 19 BSC-LT/salamandra-7b Text Generation • Updated 3 days ago • 1.08k • 9 BSC-LT/salamandra-2b-instruct Text Generation • Updated 3 days ago • 497 • 10
INTELLECT-1 Dataset INTELLECT-1 Training dataset Collection by PrimeIntellect 5 days ago 5 PrimeIntellect/fineweb-edu Viewer • Updated 3 days ago • 1.2B • 3 PrimeIntellect/fineweb Preview • Updated 2 days ago • 4 PrimeIntellect/StackV1-popular Viewer • Updated 5 days ago • 93M • 2 • 1 open-web-math/open-web-math Viewer • Updated Oct 17, 2023 • 6.32M • 4.44k • 269
Moshi v0.1 Release MLX, Candle & PyTorch model checkpoints released as part of the Moshi release from Kyutai. Run inference via: https://github.com/kyutai-labs/moshi Collection by kyutai 25 days ago 206 kyutai/moshiko-pytorch-bf16 Updated 25 days ago • 15.9k • 138 kyutai/moshika-pytorch-bf16 Updated 25 days ago • 10.7k • 42 kyutai/moshiko-mlx-q4 Updated 25 days ago • 2.69k • 23 kyutai/moshika-mlx-q4 Updated 25 days ago • 550 • 9
Qwen2.5-Coder Code-specific model series based on Qwen2.5 Collection by Qwen 18 days ago 74 Running 92 🥸 Qwen2.5-Coder-7B-Instruct Qwen2.5-Coder Technical Report Paper • 2409.12186 • Published 24 days ago • 122 Qwen/Qwen2.5-Coder-1.5B Text Generation • Updated 18 days ago • 4.77k • 20 Qwen/Qwen2.5-Coder-1.5B-Instruct Text Generation • Updated 18 days ago • 11.6k • 24