Lang ZH

SMALL SIZE, SUPER POWER

Edge Model for Everyone, Everyday, Everywhere
MiniCPM InsidePhones
MiniCPM InsideAIPC
MiniCPM InsideIntelligent Cabins
MiniCPM InsideEmbodied Robots
MiniCPM InsideWearable Devices

Put ChatGPT, GPT-4V Level LLMs on Your Phone, Pad and PC

Learn More

The 'MiniCPM' edge model series is a world-leading, lightweight, and high-performance LLM. Since its release in February 2024, it has been widely tested and acclaimed by the global open-source community for its "achieving more with less" efficiency and outstanding on-device performance. It has repeatedly topped GitHub and Hugging Face trending charts, becoming one of the most popular LLMs on Hugging Face in 2024. The 'MiniCPM' has partnered with industry benchmark leaders, emerging as an indispensable player in driving innovation across sectors such as AIPC, AI phones, intelligent cabins, and embodied robots.

High Efficiency, Low Cost, Achieving More with LessFundation Model MiniCPM
4B2.4B1.2B
The On-Device ChatGPT Moment
4B 2.4B 1.2B
githubhuggingface
Unbelievably Strong for 4B size edge Model on your device!
ChatGPT-level Basic Performance Surpassing GPT-3.5, Qwen2-7B, GLM4-9B


New Architecture, New Benchmark of LLM Knowledge Density

Light! Fast! On-Device Friendly
Only 2GB of memory after quantization
Versatile and Sharp as a Swiss Army KnifeSurpassing Kimi! Infinite Long Text
32, 128, 256, 512K... Unlimited Context Expansion

GPT-4o-level Function Calling
Surpassing GPT-3.5, GLM4-9B, Close to GPT-4o

Superior RAG External Attachment Set Number One in Chinese Retrieval, Results Generation Surpassing Llama3-8B
Learn More
View the detailed features of each version
GPT-4o level Omni Model runs on deviceMultimodal Model MiniCPM-V
8B Full-Modal8B Live Video8B2.8B
The On-Device GPT-4o New Era
8B Full-Modal 8B Live Video 8B 2.8B
githubhuggingface
Edge-Side GPT-4oReal-time streaming, end-to-end Full-modal, all SOTA The best edge visual general model The best audio general model
Continuous watching, real videos Not just a single frame-based model Real-time listening, truly smooth Hear clearly, understand distinctly Natural speaking, emotional engagement Real-time interruptions without confusionFull Capability, End-to-EndHigh performance, low latency More natural, more coherent Context understanding Interruptible at any time Noise resistance Easy deployment and maintenanceLearn More
View the detailed features of each version
Compare the functionalities of various versions

Global Partner

amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
amd
Technical Blog

端侧GPT-4o来了!全新面壁小钢炮,流式全模态+端到端!是她!端侧GPT-4O,端到端,多模态。
双榜首!登顶Hugging Face和GitHub趋势榜Top1MiniCPM-Llama3-V 2.5: 8B 参数,8G 显存,4070 轻松推理,手机端 6-8 tokens/s 高效运行。 当前 MiniCPM-V 系列下载总量已超 13 万,GitHub 星标 3k+。
感谢社区厚爱,面壁小钢炮 MiniCPM 免费商用感谢全球开源社区的朋友们,一路厚爱与支持! 作为开源社区的贡献者和受益者, 面壁智能, OpenBMB&清华 NLP 实验室 认真讨论决定: 将面壁「小钢炮」 MiniCPM 免费商用。
多图、视频首上端!3 SOTA 面壁小钢炮,创 GPT-4V 端侧全面对标新时代!再次刷新端侧多模态天花板,面壁「小钢炮」 MiniCPM-V 2.6 模型重磅上新! 仅 8B 参数,取得 20B 以下单图、多图、视频理解 3 SOTA 成绩,一举将端侧AI多模态能力拉升至全面对标 GPT-4V 水平。 更有多项功能首次上「端」:小钢炮一口气将实时视频理解、多图联合理解、多图 ICL 等能力首次搬上端侧多模态模型,更接近充斥着复杂、模糊、连续实时视觉信息的多模态真实世界,更能充分发挥端侧 AI 传感器富集、贴近用户的优势。
星标破万!小钢炮2.6登顶GitHub,Hugging Face TOP3, 燃爆开源社区!想到了直升机,没想到的是火箭! MiniCPM-V 2.6 一经发布,火箭登顶全球著名开源社区 GitHub 与 HuggingFace 趋势榜 Top 3。 至此,面壁小钢炮 MiniCPM-V系列,GitHub 星标破万! 小钢炮MiniCPM系列自今年2月1日面世以来,累计下载量已超百万!
双登顶!面壁小钢炮3.0 GitHub Top 1,Hugging Face Top 3面壁小钢炮 MiniCPM 3.0 持续引领端侧 ChatGPT 时代!
端侧 ChatGPT 时刻到来!面壁小钢炮 3.0 重磅发布面壁发布小钢炮3.0
WAIC 2024,面壁打开大模型新定律、新架构、新生态!面壁智能联合创始人、首席科学家刘知远在WAIC 2024 “模型即服务(Mass)加速大模型应用落地”论坛进行了《大模型时代的摩尔定律,迈入更高效的大模型时代》主题演讲,并首次对外介绍
面壁新模型:早于Llama3、比肩 Llama3、推理超越 Llama3!面壁发布领先的开源大模型「Eurux-8x22B 」。 相比口碑之作 Llama3-70B,发布时间更早,综合性能相当,尤其是拥有更强的推理性能——刷新开源大模型推理性能 SOTA,堪称开源大模型中「理科状元」。 激活参数仅 39B,支持 64k 上下文,相比 Llama3 速度更快、可处理更长文本。
Efficiency FirstWe believe the best model is the one with superior power, faster speed and lower costEfficiency comes from mastering the science of large language models (LLMs), with knowledge density as the key principle. As knowledge density grows, it becomes a core competitive advantage, unlocking vast potential for edge intelligence and applications.
Modelbest LawMoore’s Law
The capability density of LLMs increases exponentially over time. Since 2023, the maximum capability density of LLMs doubles approximately every 3.3 months.Capability density: The ratio of effective parameter size to actual parameter size. Effective parameter size refers to the minimum number of parameters required for the reference model (e.g., MiniCPM) to achieve performance equivalent to the given target model.
News

A G I  F O R  L I V E S