Reasoning Vectors: Transferring Chain-of-Thought Capabilities via Task Arithmetic Paper • 2509.01363 • Published Sep 1, 2025 • 58
MiniMax-M1 Collection MiniMax-M1, the world's first open-weight, large-scale hybrid-attention reasoning model. • 6 items • Updated Oct 21, 2025 • 120
YuE Collection YuE: Open Full-song Generation Foundation Model • 11 items • Updated Mar 18, 2025 • 28
AIMv2 Collection A collection of AIMv2 vision encoders that supports a number of resolutions, native resolution, and a distilled checkpoint. • 19 items • Updated Aug 25, 2025 • 82
Molmo Collection Artifacts for open multimodal language models. • 5 items • Updated 11 days ago • 309
Moshi v0.1 Release Collection MLX, Candle & PyTorch model checkpoints released as part of the Moshi release from Kyutai. Run inference via: https://github.com/kyutai-labs/moshi • 16 items • Updated 11 days ago • 242
Qwen2-Audio Collection Audio-language model series based on Qwen2 • 4 items • Updated 4 days ago • 65
Parler-TTS: fully open-source high-quality TTS Collection If you want to find out more about how these models were trained and even fine-tune them yourself, check-out the Parler-TTS repository on GitHub. • 8 items • Updated Dec 2, 2024 • 51
Minitron Collection A family of compressed models obtained via pruning and knowledge distillation • 12 items • Updated 11 days ago • 62
Chameleon Collection Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR. • 2 items • Updated Jul 9, 2024 • 33
xLAM models Collection xLAM: A Family of Large Action Models to Empower AI Agent Systems: https://github.com/SalesforceAIResearch/xLAM • 22 items • Updated Nov 5, 2025 • 59
SEED-Story: Multimodal Long Story Generation with Large Language Model Paper • 2407.08683 • Published Jul 11, 2024 • 24