inclusionAI
Team
community
AI & ML interests
None defined yet.
Recent Activity
Papers
VenusBench-GD: A Comprehensive Multi-Platform GUI Benchmark for Diverse Grounding Tasks
TwinFlow: Realizing One-step Generation on Large Models with Self-adversarial Flows
-
inclusionAI/Ming-flash-omni-Preview
Any-to-Any • 104B • Updated • 2.98k • 65 -
Ming-Flash-Omni: A Sparse, Unified Architecture for Multimodal Perception and Generation
Paper • 2510.24821 • Published • 38 -
inclusionAI/MingTok-Vision
Image Feature Extraction • 0.7B • Updated • 279 • 31 -
inclusionAI/Ming-UniVision-16B-A3B
Any-to-Any • 19B • Updated • 56 • 61
GroveMoE is an open-source family of large language models developed by the AGI Center, Ant Research Institute.
AReaL-boba-2
-
LLaDA2.0: Scaling Up Diffusion Language Models to 100B
Paper • 2512.15745 • Published • 73 -
inclusionAI/LLaDA2.0-flash
Text Generation • 103B • Updated • 433 • 58 -
inclusionAI/LLaDA2.0-mini
Text Generation • 16B • Updated • 4.96k • 47 -
inclusionAI/LLaDA2.0-flash-preview
Text Generation • 103B • Updated • 132 • 69
A collection of TwinFlow-accelerated diffusion models
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI, derived from Ling.
The Agent Runtime for Self-Improvement
-
UI-Venus Technical Report: Building High-performance UI Agents with RFT
Paper • 2508.10833 • Published • 44 -
inclusionAI/UI-Venus-Ground-7B
Image-Text-to-Text • 8B • Updated • 258 • 19 -
inclusionAI/UI-Venus-Ground-72B
Image-Text-to-Text • 73B • Updated • 214 • 11 -
inclusionAI/UI-Venus-Navi-7B
Image-Text-to-Text • 8B • Updated • 77 • 10
-
Ming-Omni: A Unified Multimodal Model for Perception and Generation
Paper • 2506.09344 • Published • 28 -
inclusionAI/Ming-Lite-Omni
Any-to-Any • 19B • Updated • 119 • 196 -
inclusionAI/Ming-Lite-Omni-1.5
Any-to-Any • 19B • Updated • 378 • 81 -
inclusionAI/Ming-UniAudio-16B-A3B
Any-to-Any • 18B • Updated • 144 • 73
-
LLaDA2.0: Scaling Up Diffusion Language Models to 100B
Paper • 2512.15745 • Published • 73 -
inclusionAI/LLaDA2.0-flash
Text Generation • 103B • Updated • 433 • 58 -
inclusionAI/LLaDA2.0-mini
Text Generation • 16B • Updated • 4.96k • 47 -
inclusionAI/LLaDA2.0-flash-preview
Text Generation • 103B • Updated • 132 • 69
A collection of TwinFlow-accelerated diffusion models
-
inclusionAI/Ming-flash-omni-Preview
Any-to-Any • 104B • Updated • 2.98k • 65 -
Ming-Flash-Omni: A Sparse, Unified Architecture for Multimodal Perception and Generation
Paper • 2510.24821 • Published • 38 -
inclusionAI/MingTok-Vision
Image Feature Extraction • 0.7B • Updated • 279 • 31 -
inclusionAI/Ming-UniVision-16B-A3B
Any-to-Any • 19B • Updated • 56 • 61
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI, derived from Ling.
The Agent Runtime for Self-Improvement
GroveMoE is an open-source family of large language models developed by the AGI Center, Ant Research Institute.
-
UI-Venus Technical Report: Building High-performance UI Agents with RFT
Paper • 2508.10833 • Published • 44 -
inclusionAI/UI-Venus-Ground-7B
Image-Text-to-Text • 8B • Updated • 258 • 19 -
inclusionAI/UI-Venus-Ground-72B
Image-Text-to-Text • 73B • Updated • 214 • 11 -
inclusionAI/UI-Venus-Navi-7B
Image-Text-to-Text • 8B • Updated • 77 • 10
AReaL-boba-2
-
Ming-Omni: A Unified Multimodal Model for Perception and Generation
Paper • 2506.09344 • Published • 28 -
inclusionAI/Ming-Lite-Omni
Any-to-Any • 19B • Updated • 119 • 196 -
inclusionAI/Ming-Lite-Omni-1.5
Any-to-Any • 19B • Updated • 378 • 81 -
inclusionAI/Ming-UniAudio-16B-A3B
Any-to-Any • 18B • Updated • 144 • 73