Open Neural Network Exchange (ONNX) is an open ecosystem that empowers AI developers
to choose the right tools as their project evolves. ONNX provides an open source format for AI models, both deep learning and traditional ML. It defines an extensible computation graph model, as well as definitions of built-in operators and standard
data types. Currently we focus on the capabilities needed for inferencing (scoring).
ONNX is widely supported and can be found in many frameworks, tools, and hardware. Enabling interoperability between different frameworks and streamlining the path from research to production helps increase the speed of innovation in the AI community. We invite the community to join us and further evolve ONNX.
ONNX is a community project. We encourage you to join the effort and contribute feedback, ideas, and code. You can participate in the SIGs and Working Groups to shape the future of ONNX.
Note: When installing in a non-Anaconda environment, make sure to install the Protobuf compiler before running the pip installation of onnx. For example, on Ubuntu:
If you are building ONNX from source on Windows, it is recommended that you also build Protobuf locally as a static library. The version distributed with conda-forge is a DLL and this is a conflict as ONNX expects it to be a static library.
Note that the instructions in this README assume you are using Visual Studio. It is recommended that you run all the commands from a shell started from “Developer Command Prompt for VS 2019” and keep the build system generator for cmake (e.g., cmake -G “Visual Studio 16 2019”) consistent.
Build Protobuf and ONNX on Windows
Step 1: Build Protobuf locally
git clone https://github.com/protocolbuffers/protobuf.git
cd protobuf
git checkout 3.11.x
cd cmake
# Explicitly set -Dprotobuf_MSVC_STATIC_RUNTIME=OFF to make sure protobuf does not statically link to runtime library
cmake -G -A -Dprotobuf_MSVC_STATIC_RUNTIME=OFF -Dprotobuf_BUILD_TESTS=OFF -Dprotobuf_BUILD_EXAMPLES=OFF -DCMAKE_INSTALL_PREFIX=<protobuf_install_dir>
# For example:
# cmake -G "Visual Studio 16 2019" -A x64 -Dprotobuf_MSVC_STATIC_RUNTIME=OFF -Dprotobuf_BUILD_TESTS=OFF -Dprotobuf_BUILD_EXAMPLES=OFF -DCMAKE_INSTALL_PREFIX=..\install
msbuild protobuf.sln /m /p:Configuration=Release
msbuild INSTALL.vcxproj /p:Configuration=Release
Step 2: Build ONNX
# Get ONNX
git clone https://github.com/onnx/onnx.git
cd onnx
git submodule update --init --recursive
# Set environment variables to find protobuf and turn off static linking of ONNX to runtime library.
# Even better option is to add it to user\system PATH so this step can be performed only once.
# For more details check https://docs.microsoft.com/en-us/cpp/build/reference/md-mt-ld-use-run-time-library?view=vs-2017
set PATH=<protobuf_install_dir>\bin;<protobuf_install_dir>\include;<protobuf_install_dir>\libs;%PATH%
set USE_MSVC_STATIC_RUNTIME=0
# use the static installed protobuf
set CMAKE_ARGS=-DONNX_USE_PROTOBUF_SHARED_LIBS=OFF -DProtobuf_USE_STATIC_LIBS=ON
# Optional: Set environment variable `ONNX_ML=1` for onnx-ml
# Build ONNX
python setup.py install
If you would prefer to use Protobuf from conda-forge instead of building Protobuf from source, you can use the following instructions.
Build ONNX on Windows with Anaconda
# Use conda-forge protobuf
conda install -c conda-forge numpy libprotobuf=3.11.3 protobuf
# Get ONNX
git clone https://github.com/onnx/onnx.git
cd onnx
git submodule update --init --recursive
# Set environment variable for ONNX to use protobuf shared lib
set USE_MSVC_STATIC_RUNTIME=0
set CMAKE_ARGS=-DONNX_USE_PROTOBUF_SHARED_LIBS=ON -DProtobuf_USE_STATIC_LIBS=OFF -DONNX_USE_LITE_PROTO=ON
# Build ONNX
# Optional: Set environment variable `ONNX_ML=1` for onnx-ml
python setup.py install
Build ONNX on ARM 64
If you are building ONNX on an ARM 64 device, please make sure to install the dependencies appropriately.
If ONNX_USE_PROTOBUF_SHARED_LIBS is ON then Protobuf_USE_STATIC_LIBS must be OFF and USE_MSVC_STATIC_RUNTIME must be 0.
If ONNX_USE_PROTOBUF_SHARED_LIBS is OFF then Protobuf_USE_STATIC_LIBS must be ON and USE_MSVC_STATIC_RUNTIME can be 1 or 0.
Note that the import onnx command does not work from the source checkout directory; in this case you’ll see ModuleNotFoundError: No module named 'onnx.onnx_cpp2py_export'. Change into another directory to fix this error.
Building ONNX on Ubuntu works well, but on CentOS/RHEL and other ManyLinux systems, you might need to open the CMakeLists file and replace all instances of /lib with /lib64.
If you want to build ONNX on Debug mode, remember to set the environment variable DEBUG=1. For debug versions of the dependencies, you need to open the CMakeLists file and append a letter d at the end of the package name lines. For example, NAMES protobuf-lite would become NAMES protobuf-lited.
You can also use the onnx-dev docker image for a Linux-based installation without having to worry about dependency versioning.
Testing
ONNX uses pytest as test driver. In order to run tests, you will first need to install pytest:
pip install pytest nbval
After installing pytest, use the following command to run tests.
Open Neural Network Exchange (ONNX) is an open ecosystem that empowers AI developers to choose the right tools as their project evolves. ONNX provides an open source format for AI models, both deep learning and traditional ML. It defines an extensible computation graph model, as well as definitions of built-in operators and standard data types. Currently we focus on the capabilities needed for inferencing (scoring).
ONNX is widely supported and can be found in many frameworks, tools, and hardware. Enabling interoperability between different frameworks and streamlining the path from research to production helps increase the speed of innovation in the AI community. We invite the community to join us and further evolve ONNX.
Use ONNX
Learn about the ONNX spec
Programming utilities for working with ONNX Graphs
Contribute
ONNX is a community project. We encourage you to join the effort and contribute feedback, ideas, and code. You can participate in the SIGs and Working Groups to shape the future of ONNX.
Check out our contribution guide to get started.
If you think some operator should be added to ONNX specification, please read this document.
Discuss
We encourage you to open Issues, or use Slack for more real-time discussion
Follow Us
Stay up to date with the latest ONNX news. [Facebook] [Twitter]
Installation
Binaries
A binary build of ONNX is available from Conda, in conda-forge:
Source
If you have installed onnx on your machine, please
pip uninstall onnxfirst before the following process of build from source.Linux and MacOS
You will need an install of Protobuf and NumPy to build ONNX. One easy way to get these dependencies is via Anaconda:
You can then install ONNX from PyPi (Note: Set environment variable
ONNX_ML=1for onnx-ml):Alternatively, you can also build and install ONNX locally from source code:
Note: When installing in a non-Anaconda environment, make sure to install the Protobuf compiler before running the pip installation of onnx. For example, on Ubuntu:
Windows
If you are building ONNX from source on Windows, it is recommended that you also build Protobuf locally as a static library. The version distributed with conda-forge is a DLL and this is a conflict as ONNX expects it to be a static library.
Note that the instructions in this README assume you are using Visual Studio. It is recommended that you run all the commands from a shell started from “Developer Command Prompt for VS 2019” and keep the build system generator for cmake (e.g., cmake -G “Visual Studio 16 2019”) consistent.
Build Protobuf and ONNX on Windows
Step 1: Build Protobuf locally
Step 2: Build ONNX
If you would prefer to use Protobuf from conda-forge instead of building Protobuf from source, you can use the following instructions.
Build ONNX on Windows with Anaconda
Build ONNX on ARM 64
If you are building ONNX on an ARM 64 device, please make sure to install the dependencies appropriately.
Verify Installation
After installation, run
to verify it works.
Common Errors
Environment variables:
USE_MSVC_STATIC_RUNTIME(should be 1 or 0, not ON or OFF)CMake variables:
ONNX_USE_PROTOBUF_SHARED_LIBS,Protobuf_USE_STATIC_LIBSIf
ONNX_USE_PROTOBUF_SHARED_LIBSis ON thenProtobuf_USE_STATIC_LIBSmust be OFF andUSE_MSVC_STATIC_RUNTIMEmust be 0. IfONNX_USE_PROTOBUF_SHARED_LIBSis OFF thenProtobuf_USE_STATIC_LIBSmust be ON andUSE_MSVC_STATIC_RUNTIMEcan be 1 or 0.Note that the
import onnxcommand does not work from the source checkout directory; in this case you’ll seeModuleNotFoundError: No module named 'onnx.onnx_cpp2py_export'. Change into another directory to fix this error.Building ONNX on Ubuntu works well, but on CentOS/RHEL and other ManyLinux systems, you might need to open the CMakeLists file and replace all instances of
/libwith/lib64.If you want to build ONNX on Debug mode, remember to set the environment variable
DEBUG=1. For debug versions of the dependencies, you need to open the CMakeLists file and append a letterdat the end of the package name lines. For example,NAMES protobuf-litewould becomeNAMES protobuf-lited.You can also use the onnx-dev docker image for a Linux-based installation without having to worry about dependency versioning.
Testing
ONNX uses pytest as test driver. In order to run tests, you will first need to install pytest:
After installing pytest, use the following command to run tests.
Development
Check out the contributor guide for instructions.
License
Apache License v2.0
Code of Conduct
ONNX Open Source Code of Conduct