|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import os |
|
|
from functools import partial |
|
|
|
|
|
import torch |
|
|
import torch.utils.checkpoint as cp |
|
|
from transformers.feature_extraction_utils import BatchFeature |
|
|
|
|
|
import deployment_scripts.trt_torch as trt |
|
|
|
|
|
|
|
|
def eagle_tensorrt_forward(self, vl_input): |
|
|
eagle_prefix = "eagle_" |
|
|
eagle_input = { |
|
|
k.removeprefix(eagle_prefix): v for k, v in vl_input.items() if k.startswith(eagle_prefix) |
|
|
} |
|
|
del eagle_input["image_sizes"] |
|
|
vl_input = eagle_input |
|
|
|
|
|
self.set_frozen_modules_to_eval_mode() |
|
|
batch_size = vl_input["pixel_values"].shape[0] |
|
|
position_ids = torch.arange(self.num_patches, device="cuda").expand((batch_size, -1)) |
|
|
if vl_input["pixel_values"].dtype != torch.float16: |
|
|
vl_input["pixel_values"] = vl_input["pixel_values"].to(torch.float16) |
|
|
|
|
|
assert ( |
|
|
vl_input["pixel_values"].shape[0] <= 8 |
|
|
), "Batch size must be <= 8 because TensorRT engine was built with max_batch_size=8, you can try to adjust the max_batch_size in the build_engine.sh script and rebuild the engine." |
|
|
|
|
|
self.vit_engine.set_runtime_tensor_shape("pixel_values", vl_input["pixel_values"].shape) |
|
|
self.vit_engine.set_runtime_tensor_shape("position_ids", position_ids.shape) |
|
|
vit_embeds = self.vit_engine(vl_input["pixel_values"], position_ids)["vit_embeds"] |
|
|
vit_embeds = vit_embeds.view(1, -1, vit_embeds.shape[-1]) |
|
|
|
|
|
if self.eagle_model.use_pixel_shuffle: |
|
|
h = w = int(vit_embeds.shape[1] ** 0.5) |
|
|
vit_embeds = vit_embeds.reshape(vit_embeds.shape[0], h, w, -1) |
|
|
vit_embeds = self.pixel_shuffle( |
|
|
vit_embeds, scale_factor=self.downsample_ratio |
|
|
) |
|
|
vit_embeds = vit_embeds.reshape( |
|
|
vit_embeds.shape[0], -1, vit_embeds.shape[-1] |
|
|
) |
|
|
|
|
|
if self.eagle_model.mlp_checkpoint and vit_embeds.requires_grad: |
|
|
vit_embeds = cp.checkpoint(self.eagle_model.mlp1, vit_embeds) |
|
|
else: |
|
|
vit_embeds = self.eagle_model.mlp1(vit_embeds) |
|
|
|
|
|
|
|
|
input_ids = vl_input["input_ids"] |
|
|
input_embeds = self.embedding_layer(input_ids) |
|
|
|
|
|
|
|
|
if input_embeds.dtype != torch.float16: |
|
|
input_embeds = input_embeds.to(torch.float16) |
|
|
if vit_embeds.dtype != torch.float16: |
|
|
vit_embeds = vit_embeds.to(torch.float16) |
|
|
|
|
|
B, N, C = input_embeds.shape |
|
|
input_embeds = input_embeds.reshape(B * N, C) |
|
|
|
|
|
input_ids_flat = input_ids.reshape(B * N) |
|
|
selected = input_ids_flat == self.image_token_index |
|
|
try: |
|
|
input_embeds[selected] = input_embeds[selected] * 0.0 + vit_embeds.reshape(-1, C) |
|
|
except Exception as e: |
|
|
vit_embeds = vit_embeds.reshape(-1, C) |
|
|
print( |
|
|
f"warning: {e}, input_embeds[selected].shape={input_embeds[selected].shape}, " |
|
|
f"vit_embeds.shape={vit_embeds.shape}" |
|
|
) |
|
|
n_token = selected.sum() |
|
|
input_embeds[selected] = input_embeds[selected] * 0.0 + vit_embeds[:n_token] |
|
|
|
|
|
input_embeds = input_embeds.reshape(B, N, C) |
|
|
|
|
|
self.llm_engine.set_runtime_tensor_shape("inputs_embeds", input_embeds.shape) |
|
|
self.llm_engine.set_runtime_tensor_shape("attention_mask", vl_input["attention_mask"].shape) |
|
|
embeddings = self.llm_engine(input_embeds, vl_input["attention_mask"])["embeddings"] |
|
|
|
|
|
return BatchFeature( |
|
|
data={ |
|
|
"backbone_features": embeddings, |
|
|
"backbone_attention_mask": vl_input["attention_mask"], |
|
|
} |
|
|
) |
|
|
|
|
|
|
|
|
def action_head_tensorrt_forward(self, backbone_output, action_input): |
|
|
|
|
|
if backbone_output.backbone_features.dtype != torch.float16: |
|
|
backbone_output.backbone_features = backbone_output.backbone_features.to(torch.float16) |
|
|
self.vlln_vl_self_attention_engine.set_runtime_tensor_shape( |
|
|
"backbone_features", backbone_output.backbone_features.shape |
|
|
) |
|
|
backbone_output.backbone_features = self.vlln_vl_self_attention_engine( |
|
|
backbone_output.backbone_features |
|
|
)["output"] |
|
|
vl_embs = backbone_output.backbone_features |
|
|
embodiment_id = action_input.embodiment_id |
|
|
batch_size = vl_embs.shape[0] |
|
|
|
|
|
if action_input.state.dtype != torch.float16: |
|
|
action_input.state = action_input.state.to(torch.float16) |
|
|
|
|
|
if embodiment_id.dtype != torch.int64: |
|
|
embodiment_id = embodiment_id.to(torch.int64) |
|
|
|
|
|
if vl_embs.dtype != torch.float16: |
|
|
vl_embs = vl_embs.to(torch.float16) |
|
|
|
|
|
|
|
|
|
|
|
self.state_encoder_engine.set_runtime_tensor_shape("state", action_input.state.shape) |
|
|
self.state_encoder_engine.set_runtime_tensor_shape("embodiment_id", embodiment_id.shape) |
|
|
state_features = self.state_encoder_engine(action_input.state, embodiment_id)["output"] |
|
|
|
|
|
|
|
|
device = vl_embs.device |
|
|
|
|
|
|
|
|
if hasattr(self, "init_actions"): |
|
|
actions = self.init_actions.expand((batch_size, -1, -1)) |
|
|
else: |
|
|
actions = torch.randn( |
|
|
size=(batch_size, self.config.action_horizon, self.config.action_dim), |
|
|
dtype=vl_embs.dtype, |
|
|
device=device, |
|
|
) |
|
|
|
|
|
num_steps = self.num_inference_timesteps |
|
|
dt = 1.0 / num_steps |
|
|
|
|
|
for t in range(num_steps): |
|
|
t_cont = t / float(num_steps) |
|
|
t_discretized = int(t_cont * self.num_timestep_buckets) |
|
|
|
|
|
|
|
|
timesteps_tensor = torch.full(size=(batch_size,), fill_value=t_discretized, device=device) |
|
|
|
|
|
self.action_encoder_engine.set_runtime_tensor_shape("actions", actions.shape) |
|
|
self.action_encoder_engine.set_runtime_tensor_shape( |
|
|
"timesteps_tensor", timesteps_tensor.shape |
|
|
) |
|
|
self.action_encoder_engine.set_runtime_tensor_shape("embodiment_id", embodiment_id.shape) |
|
|
action_features = self.action_encoder_engine(actions, timesteps_tensor, embodiment_id)[ |
|
|
"output" |
|
|
] |
|
|
|
|
|
|
|
|
if self.config.add_pos_embed: |
|
|
pos_ids = torch.arange(action_features.shape[1], dtype=torch.long, device=device) |
|
|
pos_embs = self.position_embedding(pos_ids).unsqueeze(0).to(torch.float16) |
|
|
action_features = action_features + pos_embs |
|
|
|
|
|
|
|
|
future_tokens = self.future_tokens.weight.unsqueeze(0).expand(vl_embs.shape[0], -1, -1) |
|
|
sa_embs = torch.cat((state_features, future_tokens, action_features), dim=1).to( |
|
|
torch.float16 |
|
|
) |
|
|
|
|
|
if vl_embs.dtype != torch.float16: |
|
|
vl_embs = vl_embs.to(torch.float16) |
|
|
|
|
|
self.DiT_engine.set_runtime_tensor_shape("vl_embs", vl_embs.shape) |
|
|
self.DiT_engine.set_runtime_tensor_shape("sa_embs", sa_embs.shape) |
|
|
self.DiT_engine.set_runtime_tensor_shape("timesteps_tensor", timesteps_tensor.shape) |
|
|
model_output = self.DiT_engine(sa_embs, vl_embs, timesteps_tensor)["output"] |
|
|
|
|
|
self.action_decoder_engine.set_runtime_tensor_shape("model_output", model_output.shape) |
|
|
self.action_decoder_engine.set_runtime_tensor_shape("embodiment_id", embodiment_id.shape) |
|
|
pred = self.action_decoder_engine(model_output, embodiment_id)["output"] |
|
|
pred_velocity = pred[:, -self.action_horizon :] |
|
|
|
|
|
|
|
|
actions = actions + dt * pred_velocity |
|
|
return BatchFeature(data={"action_pred": actions}) |
|
|
|
|
|
|
|
|
def setup_tensorrt_engines( |
|
|
policy, trt_engine_path, vit_dtype="fp8", llm_dtype="nvfp4", dit_dtype="fp8" |
|
|
): |
|
|
""" |
|
|
Setup TensorRT engines for GR00T model inference. |
|
|
|
|
|
Args: |
|
|
policy: GR00T policy model instance |
|
|
trt_engine_path: Path to the directory containing TensorRT engine files |
|
|
vit_dtype: ViT model dtype (fp16, fp8) |
|
|
llm_dtype: LLM model dtype (fp16, nvfp4) |
|
|
dit_dtype: DiT model dtype (fp16, fp8) |
|
|
""" |
|
|
policy.model.backbone.num_patches = ( |
|
|
policy.model.backbone.eagle_model.vision_model.vision_model.embeddings.num_patches |
|
|
) |
|
|
|
|
|
if hasattr(policy.model.backbone.eagle_model, "language_model"): |
|
|
policy.model.backbone.embedding_layer = ( |
|
|
policy.model.backbone.eagle_model.language_model.get_input_embeddings() |
|
|
) |
|
|
policy.model.backbone.image_token_index = ( |
|
|
policy.model.backbone.eagle_model.image_token_index |
|
|
) |
|
|
if hasattr(policy.model.backbone.eagle_model, "vision_model"): |
|
|
del policy.model.backbone.eagle_model.vision_model |
|
|
if hasattr(policy.model.backbone.eagle_model, "language_model"): |
|
|
del policy.model.backbone.eagle_model.language_model |
|
|
if hasattr(policy.model.action_head, "vlln"): |
|
|
del policy.model.action_head.vlln |
|
|
if hasattr(policy.model.action_head, "vl_self_attention"): |
|
|
del policy.model.action_head.vl_self_attention |
|
|
if hasattr(policy.model.action_head, "model"): |
|
|
del policy.model.action_head.model |
|
|
if hasattr(policy.model.action_head, "state_encoder"): |
|
|
del policy.model.action_head.state_encoder |
|
|
if hasattr(policy.model.action_head, "action_encoder"): |
|
|
del policy.model.action_head.action_encoder |
|
|
if hasattr(policy.model.action_head, "action_decoder"): |
|
|
del policy.model.action_head.action_decoder |
|
|
torch.cuda.empty_cache() |
|
|
|
|
|
|
|
|
policy.model.backbone.vit_engine = trt.Engine( |
|
|
os.path.join(trt_engine_path, f"vit_{vit_dtype}.engine") |
|
|
) |
|
|
policy.model.backbone.llm_engine = trt.Engine( |
|
|
os.path.join(trt_engine_path, f"llm_{llm_dtype}.engine") |
|
|
) |
|
|
|
|
|
|
|
|
policy.model.action_head.vlln_vl_self_attention_engine = trt.Engine( |
|
|
os.path.join(trt_engine_path, "vlln_vl_self_attention.engine") |
|
|
) |
|
|
policy.model.action_head.action_encoder_engine = trt.Engine( |
|
|
os.path.join(trt_engine_path, "action_encoder.engine") |
|
|
) |
|
|
policy.model.action_head.action_decoder_engine = trt.Engine( |
|
|
os.path.join(trt_engine_path, "action_decoder.engine") |
|
|
) |
|
|
policy.model.action_head.DiT_engine = trt.Engine( |
|
|
os.path.join(trt_engine_path, f"DiT_{dit_dtype}.engine") |
|
|
) |
|
|
policy.model.action_head.state_encoder_engine = trt.Engine( |
|
|
os.path.join(trt_engine_path, "state_encoder.engine") |
|
|
) |
|
|
|
|
|
|
|
|
policy.model.backbone.forward = partial(eagle_tensorrt_forward, policy.model.backbone) |
|
|
policy.model.action_head.get_action = partial( |
|
|
action_head_tensorrt_forward, policy.model.action_head |
|
|
) |
|
|
|