zhenyuzhao's picture
Add files using upload-large-folder tool
a82517c verified
# SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
# SPDX-License-Identifier: Apache-2.0
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import atexit
import ctypes
import os
import tensorrt as trt
import torch
def torch_type(trt_type):
mapping = {
trt.float32: torch.float32,
trt.float16: torch.float16,
trt.int8: torch.int8,
trt.int32: torch.int32,
trt.bool: torch.bool,
trt.uint8: torch.uint8,
trt.int64: torch.int64,
}
if trt_type in mapping:
return mapping[trt_type]
raise TypeError(
f"Could not resolve TensorRT datatype to an equivalent numpy datatype. {trt_type}"
)
class Engine(object):
def __init__(self, file, plugins=[]):
super().__init__()
self.logger = trt.Logger(trt.Logger.ERROR)
trt.init_libnvinfer_plugins(self.logger, "")
self.plugins = [ctypes.CDLL(plugin, ctypes.RTLD_GLOBAL) for plugin in plugins]
self.file = file
self.load(file)
def destroy(self):
del self.execution_context
del self.handle
atexit.register(destroy, self)
self.print()
def print(self):
if int(os.getenv("LOCAL_RANK", -1)) not in [0, -1]:
return
print("============= TRT Engine Detail =============")
print(f"Engine file: {self.file}")
print(f"Inputs: {len(self.in_meta)}")
for ib, item in enumerate(self.in_meta):
tensor_name, shape, dtype = item[:3]
print(f" {ib}. {tensor_name}: {'x'.join(map(str, shape))} [{dtype}]")
print(f"Outputs: {len(self.out_meta)}")
for ib, item in enumerate(self.out_meta):
tensor_name, shape, dtype = item[:3]
print(f" {ib}. {tensor_name}: {'x'.join(map(str, shape))} [{dtype}]")
print("=============================================")
def load(self, file):
runtime = trt.Runtime(self.logger)
with open(file, "rb") as f:
self.handle = runtime.deserialize_cuda_engine(f.read())
assert (
self.handle is not None
), f"Failed to deserialize the cuda engine from file: {file}"
self.execution_context = self.handle.create_execution_context()
self.meta, self.in_meta, self.out_meta = [], [], []
for tensor_name in self.handle:
shape = self.handle.get_tensor_shape(tensor_name)
dtype = torch_type(self.handle.get_tensor_dtype(tensor_name))
if self.handle.get_tensor_mode(tensor_name) == trt.TensorIOMode.INPUT:
self.in_meta.append([tensor_name, shape, dtype])
else:
self.out_meta.append([tensor_name, shape, dtype])
def __call__(self, *args, **inputs):
return self.forward(*args, **inputs)
def set_runtime_tensor_shape(self, name, shape):
self.execution_context.set_input_shape(name, shape)
def forward(self, *args, **kwargs):
return_list = kwargs.pop("return_list", False)
reference_tensors = []
stream = torch.cuda.current_stream()
for iarg, x in enumerate(args):
name, shape, dtype = self.in_meta[iarg]
runtime_shape = self.execution_context.get_tensor_shape(name)
assert isinstance(x, torch.Tensor), f"Unsupported tensor type: {type(x)}"
assert runtime_shape == x.shape, f"Invalid input shape: {runtime_shape} != {x.shape}"
assert (
dtype == x.dtype
), f"Invalid tensor dtype, excepted dtype is {dtype}, but got {x.dtype}"
assert x.is_cuda, f"Invalid tensor device, excepted device is cuda, but got {x.device}"
x = x.cuda().contiguous()
self.execution_context.set_tensor_address(name, x.data_ptr())
reference_tensors.append(x)
for name, shape, dtype in self.in_meta:
if name not in kwargs:
continue
runtime_shape = self.execution_context.get_tensor_shape(name)
x = kwargs[name]
assert isinstance(x, torch.Tensor), f"Unsupported tensor[{name}] type: {type(x)}"
assert (
runtime_shape == x.shape
), f"Invalid input[{name}] shape: {x.shape}, but the expected shape is: {runtime_shape}"
assert (
dtype == x.dtype
), f"Invalid tensor[{name}] dtype, expected dtype is {dtype}, but got {x.dtype}"
assert (
x.is_cuda
), f"Invalid tensor[{name}] device, expected device is cuda, but got {x.device}"
x = x.cuda().contiguous()
self.execution_context.set_tensor_address(name, x.data_ptr())
reference_tensors.append(x)
for item in self.out_meta:
name = item[0]
runtime_shape = self.execution_context.get_tensor_shape(name)
output_tensor = torch.zeros(
*runtime_shape, dtype=item[2], device=reference_tensors[0].device
)
self.execution_context.set_tensor_address(name, output_tensor.data_ptr())
reference_tensors.append(output_tensor)
self.execution_context.execute_async_v3(stream.cuda_stream)
stream.synchronize()
assert len(reference_tensors) == len(self.in_meta) + len(
self.out_meta
), f"Invalid input tensors. The expected I/O tensors are {len(self.in_meta) + len(self.out_meta)}, but got {len(reference_tensors)}"
if return_list:
return [
reference_tensors[len(self.in_meta) + i] for i, item in enumerate(self.out_meta)
]
else:
return {
item[0]: reference_tensors[len(self.in_meta) + i]
for i, item in enumerate(self.out_meta)
}