目录

🌐 Language: English | 简体中文

FlagRelease

FlagRelease 是由智源研究院研发的一套面向多架构人工智能芯片的大模型自动迁移、适配与发布平台,平台旨在通过自动化、标准化和智能化的适配流程,使主流大模型能够在不同国产人工智能芯片上以更低成本、更高效率完成模型迁移、验证与发版。依托统一、开源的 AI 系统软件栈 FlagOS 提供的跨芯适配能力,FlagRelease 构建了一套标准化流程,实现大模型自动迁移至不同硬件架构、自动评测迁移效果、内置自动部署与调优,并生成多芯版本模型进行发布。基于FlagRelease平台发布的产物,用户可以在 魔搭(ModelScope)或 Hugging Face 的 FlagRelease 页面找到开源大模型的不同芯片版本,下载即可在对应硬件环境上直接使用,无需自行迁移,降低用户迁移成本。 FlagRelease 平台目前的产出物包括经过迁移和验证的模型文件以及一体化 Docker 镜像,镜像中包括FlagOS核心组件和模型依赖包,用户可以直接应用到对应芯片。同时每个模型版本都提供评测结果作为技术参考,让用户清楚不同硬件上的正确性表现。另外,每个模型发布附带AnythingLLM配置使用方式,帮助用户快速验证迁移模型的可用性,方便用户基于模型进行二次开发。 FlagOS 整体架构如下图所示:

模型发布日志

Model Name 原始模型
DeepSeek-R1-Distill-Qwen-32B Huggingface: DeepSeek-R1-Distill-Qwen-32B-FlagOS-NVIDIA
Modalscope: DeepSeek-R1-Distill-Qwen-32B-FlagOS-NVIDIA
Huggingface: DeepSeek-R1-Distill-Qwen-32B-FlagOS-Cambricon
Modalscope: DeepSeek-R1-Distill-Qwen-32B-FlagOS-Cambricon
MiniMax-M1-80k Huggingface: MiniMax-M1-80k-FlagOS
Modalscope: MiniMax-M1-80k-FlagOS
Qwen3.5-35B-A3B Huggingface: Qwen3.5-35B-A3B-FlagOS
Modalscope: Qwen3.5-35B-A3B-FlagOS
Qwen2-7B-Instruct Huggingface: Qwen2-7B-Instruct-FlagOS
Modalscope: Qwen2-7B-Instruct-FlagOS
Qwen3-235B-A22B Huggingface: Qwen3-235B-A22B-FlagOS-nvidia
Modalscope: Qwen3-235B-A22B-FlagOS-nvidia
phi-4 Huggingface: phi-4-FlagOS
Modalscope: phi-4-FlagOS
Huggingface: phi-4-hygon-FlagOS
Modalscope: phi-4-hygon-FlagOS
Huggingface: phi-4-metax-FlagOS
Modalscope: phi-4-metax-FlagOS
Qwen2.5-32B-Instruct Huggingface: Qwen2.5-32B-Instruct-FlagOS-Nvidia
Modalscope: Qwen2.5-32B-Instruct-FlagOS-Nvidia
RoboBrain2.0-7B-W8A16 Huggingface: RoboBrain2.0-7B-W8A16-FlagOS
Modalscope: RoboBrain2.0-7B-W8A16-FlagOS
pi0 Huggingface: pi0-FlagOS
Modalscope: pi0-FlagOS
DeepSeek-R1-INT8 Huggingface: DeepSeek-R1-FlagOS-Iluvatar-INT8
Modalscope: DeepSeek-R1-FlagOS-Iluvatar-INT8
Huggingface: DeepSeek-R1-FlagOS-Kunlunxin-INT8
Modalscope: DeepSeek-R1-FlagOS-Kunlunxin-INT8
DeepSeek-R1-INT4 Huggingface: DeepSeek-R1-INT4-FlagOS-Iluvatar
Modalscope: DeepSeek-R1-INT4-FlagOS-Iluvatar
grok-2 Huggingface: grok-2-FlagOS
Modalscope: grok-2-FlagOS
RoboBrain-X0 Huggingface: RoboBrain-X0-FlagOS
Modalscope: RoboBrain-X0-FlagOS
MiniCPM-V-4 Huggingface: MiniCPM-V-4-FlagOS
Modalscope: MiniCPM-V-4-FlagOS
Huggingface: MiniCPM-V-4-metax-FlagOS
Modalscope: MiniCPM-V-4-metax-FlagOS
Qwen3-VL-235B-A22B-Instruct Huggingface: Qwen3-VL-235B-A22B-Instruct-FlagOS
Modalscope: Qwen3-VL-235B-A22B-Instruct-FlagOS
GLM-4.5 Huggingface: GLM-4.5-FlagOS
Modalscope: GLM-4.5-FlagOS
step3 Huggingface: step3-FlagOS
Modalscope: step3-FlagOS
Qwen3.5-397B-A17B-zhenwu Huggingface: Qwen3.5-397B-A17B-zhenwu-FlagOS
Modalscope: Qwen3.5-397B-A17B-zhenwu-FlagOS
RoboBrain2.0-7B Huggingface: RoboBrain2.0-7B-FlagOS
Modalscope: RoboBrain2.0-7B-FlagOS
Huggingface: RoboBrain2.0-7B-FlagOS-Ascend
Modalscope: RoboBrain2.0-7B-FlagOS-Ascend
Huggingface: RoboBrain2.0-7B-metax-FlagOS
Modalscope: RoboBrain2.0-7B-metax-FlagOS
Kimi-K2-Instruct Huggingface: Kimi-K2-Instruct-FlagOS
Modalscope: Kimi-K2-Instruct-FlagOS
Hunyuan-A13B-Instruct Huggingface: Hunyuan-A13B-Instruct-FlagOS
Modalscope: Hunyuan-A13B-Instruct-FlagOS
RoboBrain2.0-7B-FP8Dynamic Huggingface: RoboBrain2.0-7B-FP8Dynamic-FlagOS
Modalscope: RoboBrain2.0-7B-FP8Dynamic-FlagOS
RoboBrain-X0-Preview Huggingface: RoboBrain-X0-Preview-FlagOS
Modalscope: RoboBrain-X0-Preview-FlagOS
Huggingface: RoboBrain-X0-Preview-ascend-FlagOS
Modalscope: RoboBrain-X0-Preview-ascend-FlagOS
Kimi-K2-Thinking Huggingface: Kimi-K2-Thinking-FlagOS
Modalscope: Kimi-K2-Thinking-FlagOS
MiniCPM-o-4.5-zhenwu Huggingface: MiniCPM-o-4.5-zhenwu-FlagOS
Modalscope: MiniCPM-o-4.5-zhenwu-FlagOS
DeepSeek-R1-BF16 Huggingface: DeepSeek-R1-FlagOS-Nvidia-BF16
Modalscope: DeepSeek-R1-FlagOS-Nvidia-BF16
Huggingface: DeepSeek-R1-FlagOS-Metax-BF16
Modalscope: DeepSeek-R1-FlagOS-Metax-BF16
Huggingface: DeepSeek-R1-FlagOS-Cambricon-BF16
Modalscope: DeepSeek-R1-FlagOS-Cambricon-BF16
MiniMax-M2 Huggingface: MiniMax-M2-FlagOS
Modalscope: MiniMax-M2-FlagOS
Qwen3-Omni-30B-A3B-Instruct Huggingface: Qwen3-Omni-30B-A3B-Instruct-FlagOS
Modalscope: Qwen3-Omni-30B-A3B-Instruct-FlagOS
Qwen2-7B Huggingface: Qwen2-7B-FlagOS-Arm
Modalscope: Qwen2-7B-FlagOS-Arm
QwQ-32B Huggingface: QwQ-32B-FlagOS-Cambricon
Modalscope: QwQ-32B-FlagOS-Cambricon
Huggingface: QwQ-32B-FlagOS-Nvidia
Modalscope: QwQ-32B-FlagOS-Nvidia
Huggingface: QwQ-32B-FlagOS-Iluvatar
Modalscope: QwQ-32B-FlagOS-Iluvatar
MiniCPM-o-4.5 Huggingface: MiniCPM-o-4.5-ascend-FlagOS
Modalscope: MiniCPM-o-4.5-ascend-FlagOS
Huggingface: MiniCPM-o-4.5-metax-FlagOS
Modalscope: MiniCPM-o-4.5-metax-FlagOS
Huggingface: MiniCPM-o-4.5-iluvatar-FlagOS
Modalscope: MiniCPM-o-4.5-iluvatar-FlagOS
Huggingface: MiniCPM-o-4.5-hygon-FlagOS
Modalscope: MiniCPM-o-4.5-hygon-FlagOS
Huggingface: MiniCPM-o-4.5-nvidia-FlagOS
Modalscope: MiniCPM-o-4.5-nvidia-FlagOS
Qwen3.5-397B-A17B Huggingface: Qwen3.5-397B-A17B-nvidia-FlagOS
Modalscope: Qwen3.5-397B-A17B-nvidia-FlagOS
Huggingface: Qwen3.5-397B-A17B-metax-FlagOS
Modalscope: Qwen3.5-397B-A17B-metax-FlagOS
Qwen3-Next-80B-A3B-Instruct Huggingface: Qwen3-Next-80B-A3B-Instruct-FlagOS
Modalscope: Qwen3-Next-80B-A3B-Instruct-FlagOS
Huggingface: Qwen3-Next-80B-A3B-Instruct-metax-FlagOS
Modalscope: Qwen3-Next-80B-A3B-Instruct-metax-FlagOS
Qwen3-32B Huggingface: Qwen3-32B-FlagOS
Modalscope: Qwen3-32B-FlagOS
Huggingface: Qwen3-32B-ascend-FlagOS
Modalscope: Qwen3-32B-ascend-FlagOS
Qwen3-8B Huggingface: Qwen3-8B-metax-FlagOS
Modalscope: Qwen3-8B-metax-FlagOS
Huggingface: Qwen3-8B-FlagOS
Modalscope: Qwen3-8B-FlagOS
Huggingface: Qwen3-8B-mthreads-FlagOS
Modalscope: Qwen3-8B-mthreads-FlagOS
Emu3.5 Huggingface: Emu3.5-FlagOS
Modalscope: Emu3.5-FlagOS
MiniCPM_o_2.6 Huggingface: MiniCPM_o_2.6-FlagOS-Cambricon
Modalscope: MiniCPM_o_2.6-FlagOS-Cambricon
Huggingface: MiniCPM_o_2.6-FlagOS-NVIDIA
Modalscope: MiniCPM_o_2.6-FlagOS-NVIDIA
DeepSeek-V3.2-Exp Huggingface: DeepSeek-V3.2-Exp-FlagOS
Modalscope: DeepSeek-V3.2-Exp-FlagOS
Qwen2.5-VL-32B-Instruct-BF16 Huggingface: Qwen2.5-VL-32B-Instruct-FlagOS-Metax-BF16
Modalscope: Qwen2.5-VL-32B-Instruct-FlagOS-Metax-BF16
Qwen2.5-VL-32B-Instruct Huggingface: Qwen2.5-VL-32B-Instruct-FlagOS-Nvidia
Modalscope: Qwen2.5-VL-32B-Instruct-FlagOS-Nvidia
RoboBrain2.5-8B Huggingface: RoboBrain2.5-8B-FlagOS
Modalscope: RoboBrain2.5-8B-FlagOS
Huggingface: RoboBrain2.5-8B-ascend-FlagOS
Modalscope: RoboBrain2.5-8B-ascend-FlagOS
gpt-oss-120b Huggingface: gpt-oss-120b-FlagOS
Modalscope: gpt-oss-120b-FlagOS
Qwen3-4B Huggingface: Qwen3-4B-FlagOS-cambricon
Modalscope: Qwen3-4B-FlagOS-cambricon
Huggingface: Qwen3-4B-FlagOS-Nvidia
Modalscope: Qwen3-4B-FlagOS-Nvidia
Huggingface: Qwen3-4B-FlagOS-Iluvatar
Modalscope: Qwen3-4B-FlagOS-Iluvatar
Huggingface: Qwen3-4B-FlagOS-Metax
Modalscope: Qwen3-4B-FlagOS-Metax
Huggingface: Qwen3-4B-FlagOS-Ascend
Modalscope: Qwen3-4B-FlagOS-Ascend
Huggingface: Qwen3-4B-hygon-FlagOS
Modalscope: Qwen3-4B-hygon-FlagOS
Qwen3-30B-A3B Huggingface: Qwen3-30B-A3B-FlagOS-nvidia
Modalscope: Qwen3-30B-A3B-FlagOS-nvidia
Huggingface: Qwen3-30B-A3B-Iluvatar-FlagOS
Modalscope: Qwen3-30B-A3B-Iluvatar-FlagOS
RoboBrain2.0-32B Huggingface: RoboBrain2.0-32B-FlagOS
Modalscope: RoboBrain2.0-32B-FlagOS
Huggingface: RoboBrain2.0-32B-Ascend-FlagOS
Modalscope: RoboBrain2.0-32B-Ascend-FlagOS
Qwen3-235B-A22B-Instruct-2507 Huggingface: Qwen3-235B-A22B-Instruct-2507-FlagOS
Modalscope: Qwen3-235B-A22B-Instruct-2507-FlagOS
Huggingface: Qwen3-235B-A22B-Instruct-2507-hygon-FlagOS
Modalscope: Qwen3-235B-A22B-Instruct-2507-hygon-FlagOS
Seed-OSS-36B-Instruct Huggingface: Seed-OSS-36B-Instruct-FlagOS
Modalscope: Seed-OSS-36B-Instruct-FlagOS
ERNIE-4.5-300B-A47B-PT Huggingface: ERNIE-4.5-300B-A47B-PT-FlagOS
Modalscope: ERNIE-4.5-300B-A47B-PT-FlagOS

产出物操作使用示例

FlagRelease 的产出物包括经过验证的大模型文件以及一体化的 FlagOS Docker 镜像。利用这些产出物,用户可以快速在不同硬件上部署和运行大模型,无需自行迁移或配置复杂环境。 操作步骤示例:

  1. 下载开源模型权重
  • 通过魔搭(ModelScope)或 Hugging Face 的 FlagRelease 页面,选择所需的大模型及对应芯片版本,直接下载模型权重文件。
  1. 下载 FlagOS 镜像
  • 获取官方提供的一体化 Docker 镜像,包含统一的软件栈和硬件适配支持。
  1. 部署与运行
  • 将下载的模型权重和 FlagOS 镜像结合,即可在目标硬件上直接运行模型。
  • FlagOS 会自动管理硬件资源,支持多芯片并行计算,无需手动配置环境。

示例应用场景:

  • 在科研实验中,快速部署大模型进行推理,无需关注底层硬件差异。
  • 在生产环境中,直接使用不同芯片版本的模型进行服务部署,保证性能和稳定性。
关于

No description, website, or topics provided.

870.0 KB
邀请码
    Gitlink(确实开源)
  • 加入我们
  • 官网邮箱:gitlink@ccf.org.cn
  • QQ群
  • QQ群
  • 公众号
  • 公众号

版权所有:中国计算机学会技术支持:开源发展技术委员会
京ICP备13000930号-9 京公网安备 11010802032778号