This organization contains the series of open-source projects from Ant Group with dedicated efforts to work towards Artificial General Intelligence (AGI).

Ling: A MoE LLM Provided and Open-sourced by InclusionAI

🤗 Hugging Face   |   🤖 ModelScope Introduction Ling is a MoE LLM provided and open-sourced by InclusionAI. We introduce two different sizes, which are Ling-lite and Ling-plus. Ling-lite has 16.8 billion parameters with 2.75 billion activated parameters, while Ling-plus has 290 billion parameters with 28.8 billion activated parameters. Both models demonstrate impressive performance compared to existing models in the industry. Their structure makes it easy to scale up and down and adapt to different tasks, so users can use these models for a wide range of tasks, from processing natural language to solving complex problems....

May 8, 2025 Â· 8 min Â· 1574 words Â· inclusionAI, Ant Group

Ming-Lite-Uni: Advancements in Unified Architecture for Natural Multimodal Interaction

GITHUB 📑 Paper|🤗 Hugging Face|🤖 ModelScope Introduction Ming-Lite-Uni is an open-source multimodal framework that includes a newly developed unified visual generator, and a native multimodal autoregressive model meant to integrate vision and language. This project offers an open-source implementation of the integrated MetaQueries and M2-omni framework, while offering the innovative multi-scale learnable tokens and multi-scale representation alignment strategy. Ming-Lite-Uni utilizes a fixed MLLM and a learnable diffusion model, allowing native multimodal AR models to execute text-to-image production and instruction-based image editing tasks, hence enhancing their functionalities beyond mere visual comprehension....

May 7, 2025 Â· 6 min Â· 1133 words Â· inclusionAI, Ant Group

Ming-Lite-Omni-Preview: A MoE Model Designed to Perceive a Wide Range of Modalities

GITHUB 🤗 Hugging Face | 🤖 ModelScope Introduction Ming-Lite-Omni-Preview is built upon Ling-Lite, which is a MoE model designed to perceive a wide range of modalities, including text, images, audio, and video, while generating text and natural speech in a streaming manner. To naturely handle the diverse modalities, we have enhanced Ling-Lite by incorporating modality-specific routers for each modality. As a result, Ming-Omni excels at handling information from diverse modalities and is highly scalable....

May 5, 2025 Â· 6 min Â· 1105 words Â· inclusionAI, Ant Group

Agentic Learning

Introduction Agent exhibits powerful capabilities by interacting with the external environment and making decisions based on the feedback it receives from the environment. For complex problems, it is often necessary for an agent to have multi-turn interactions with the environment to reach a solution. The complexity and dynamism of environments, coupled with the necessity for multi-turn interactions, pose numerous challenges in training agents. We introduce AgenticLearning, an open-source agent training paradigm designed to empower researchers to train and evaluate autonomous agents effectively....

April 1, 2025 Â· 3 min Â· 446 words Â· inclusionAI, Ant Group

AReaL: Ant Reasoning Reinforcement Learning for LLMs

| Paper | Documentation | Ask DeepWiki | 🤗 Models & Data | WeChat Group | AReaL (Ant Reasoning RL) is an open-source fully asynchronous reinforcement learning training system for large reasoning models developed at the RL Lab, Ant Research. Built upon the open-source project RealHF, we are fully committed to open-source by providing training details, data, and infrastructure required to reproduce results along with the model itself. AReaL aims to help everyone build their own AI agents easily and affordably....

April 1, 2025 Â· 7 min Â· 1431 words Â· inclusionAI, Ant Group