Dorothydu's picture
Upload 50 random repository samples
9d3c8f5 verified
from typing import Optional
from onnx import TensorProto
from onnx.helper import make_tensor
from onnxscript import script
from onnxscript.onnx_opset import opset15 as op
from onnxscript.onnx_types import FLOAT
# Design choices for optional input:
# Consider the layer-normalization function, which has an optional Bias input.
# As a toy-version of this, assume we want to compute "Log(X) + Bias", where Bias
# is optional.
# The function-implementation option1 below has the following advantages:
# (1) Easier for backend to optimize away an unnecessary addition (when no bias is present)
# (2) It is more general (than option2) since it can handle cases where a simple default
# value like zero cannot be used.
# However, this requires us to add a new (pseudo) primitive-op to ONNX that does the
# check "Bias != None"
def option1(X, Bias: FLOAT[...] = None):
Y = op.Log(X)
if Bias != None:
Y = Y + Bias
return Y
# The pros/cons of the option2 implementation below are just the dual of option1.
# (1) This leads to an unnecessary tensor creation and add operation.
# (2) It could be difficult to do in the general-case. E.g., down-below, we need
# to introduce a `CastLike` op to handle the typing aspect.
# This requires introducing default-values for parameters in FunctionProto,
# similar to initializers in GraphProto
def option2(X, Bias=op.Constant(value=make_tensor("zero", TensorProto.FLOAT, [1], [0]))):
Y = op.Log(X)
Bias = CastLike(Bias, Y)
Y = Y + Bias
return Y
# The implementation option3 is similar to option1, but differs in one aspect.
# It changes the type-signature of the op/function, namely Bias is declared to
# be an optional-type. This means that we don't need to introduce a new primitive
# op, since we already have an op to check if "Bias != None" in this case.
# (See the op OptionalHasElement.)
# While this is a cleaner version of option1, it is challenging to go back and
# change the type signature of all optional parameters in ONNX ops. That would be
# quite disruptive. In short, ONNX originally introduced a limited form of
# optional inputs (before the full-fledged optional-type was introduced) and that
# is the source of our problem.
# does not work yet.
# TypeError: typing.Optional requires a single type. Got FLOAT.
# def option3(X, Bias: Optional[FLOAT[...]] = None):
# Y = op.Log(X)
# if (Bias != None):
# Y = Y + Bias
# return Y
# Proposal: The proposal is use option 1, and define a variant of the OptionalHasElement
# op in ONNX to enable this.