| | --- |
| | myst: |
| | substitutions: |
| | onnxscript: '*ONNX Script*' |
| | --- |
| | |
| | # ONNX Script |
| |
|
| | For instructions on how to install **ONNX Script** refer to [ONNX Script Github Repo](https://github.com/microsoft/onnxscript) |
| |
|
| |
|
| | ## Overview |
| |
|
| | {{ onnxscript }} enables developers to naturally author ONNX functions and |
| | models using a subset of Python. It is intended to be: |
| |
|
| | - **Expressive:** enables the authoring of all ONNX functions. |
| | - **Simple and concise:** function code is natural and simple. |
| | - **Debuggable:** allows for eager-mode evaluation that enables |
| | debugging the code using standard python debuggers. |
| |
|
| | Note however that {{ onnxscript }} does **not** intend to support the entirety |
| | of the Python language. |
| |
|
| | {{ onnxscript }} provides a few major capabilities for authoring and debugging |
| | ONNX models and functions: |
| |
|
| | - A converter which translates a Python {{ onnxscript }} function into an |
| | ONNX graph, accomplished by traversing the Python Abstract Syntax Tree |
| | to build an ONNX graph equivalent of the function. |
| | - A runtime shim that allows such functions to be evaluated |
| | (in an "eager mode"). This functionality currently relies on |
| | ONNX Runtime for executing ONNX ops |
| | and there is a Python-only reference runtime for ONNX underway that |
| | will also be supported. |
| | - A converter that translates ONNX models and functions into {{ onnxscript }}. |
| | This capability can be used to fully round-trip ONNX Script ↔ ONNX graph. |
| |
|
| | Note that the runtime is intended to help understand and debug function definitions. |
| | Performance is not a goal here. |
| |
|
| |
|
| | ## Example |
| |
|
| | The following toy example illustrates how to use onnxscript. |
| |
|
| | ```python |
| | from onnxscript import script |
| | # We use ONNX opset 15 to define the function below. |
| | from onnxscript import opset15 as op |
| | |
| | # We use the script decorator to indicate that the following function is meant |
| | # to be translated to ONNX. |
| | @script() |
| | def MatmulAdd(X, Wt, Bias): |
| | return op.MatMul(X, Wt) + Bias |
| | ``` |
| |
|
| | The decorator parses the code of the function and converts it into an intermediate |
| | representation. If it fails, it produces an error message indicating the error detected. |
| | If it succeeds, the corresponding ONNX representation of the function |
| | (a value of type FunctionProto) can be generated as shown below: |
| |
|
| | ```python |
| | fp = MatmulAdd.to_function_proto() # returns an onnx.FunctionProto |
| | ``` |
| |
|
| | One can similarly generate an ONNX Model. There are a few differences between |
| | ONNX models and ONNX functions. For example, ONNX models must specify the |
| | type of inputs and outputs (unlike ONNX functions). |
| | The following example illustrates how we can generate an ONNX Model: |
| |
|
| | ```python |
| | from onnxscript import script |
| | from onnxscript import opset15 as op |
| | from onnxscript import FLOAT |
| | |
| | @script() |
| | def MatmulAddModel(X : FLOAT[64, 128] , Wt: FLOAT[128, 10], Bias: FLOAT[10]) -> FLOAT[64, 10]: |
| | return op.MatMul(X, Wt) + Bias |
| | |
| | model = MatmulAddModel.to_model_proto() # returns an onnx.ModelProto |
| | ``` |
| |
|
| | ## Eager mode |
| |
|
| | Eager evaluation mode is mostly use to debug and check intermediate results |
| | are as expected. The function defined earlier can be called as below, and this |
| | executes in an eager-evaluation mode. |
| |
|
| | ```python |
| | import numpy as np |
| | |
| | x = np.array([[0, 1], [2, 3]], dtype=np.float32) |
| | wt = np.array([[0, 1], [2, 3]], dtype=np.float32) |
| | bias = np.array([0, 1], dtype=np.float32) |
| | result = MatmulAdd(x, wt, bias) |
| | ``` |
| |
|
| | ```{toctree} |
| | :maxdepth: 1 |
| | |
| | Overview <self> |
| | tutorial/index |
| | api/index |
| | intermediate_representation/index |
| | auto_examples/index |
| | articles/index |
| | ``` |
| |
|
| | ## License |
| |
|
| | onnxscript comes with a [MIT](https://github.com/microsoft/onnxscript/blob/main/LICENSE) license. |
| |
|