|
|
""" |
|
|
GeoMotionGPT Model |
|
|
|
|
|
This module contains the model implementation for GeoMotionGPT, integrating: |
|
|
1. Motion Tokenizer (DVQ-GSST VQ-VAE) |
|
|
2. Language Model (fine-tuned GPT-2 for motion-to-text) |
|
|
|
|
|
Usage: |
|
|
```python |
|
|
from transformers import AutoModelForCausalLM |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained("zy22b/GeoMotionGPT", trust_remote_code=True) |
|
|
motion_tokenizer = model.motion_tokenizer |
|
|
|
|
|
# Tokenize motion |
|
|
motion_tokens = motion_tokenizer.encode(motion_features) |
|
|
|
|
|
# Generate text |
|
|
text = model.generate_from_motion(motion_tokens) |
|
|
``` |
|
|
""" |
|
|
|
|
|
import torch |
|
|
import torch.nn as nn |
|
|
import torch.nn.functional as F |
|
|
from typing import Optional, Tuple, List, Union |
|
|
from transformers import PreTrainedModel, GPT2LMHeadModel, GPT2Config |
|
|
from transformers.modeling_outputs import CausalLMOutputWithCrossAttentions |
|
|
|
|
|
|
|
|
try: |
|
|
from .configuration_geomotiongpt import GeoMotionGPTConfig |
|
|
except ImportError: |
|
|
from configuration_geomotiongpt import GeoMotionGPTConfig |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class Swish(nn.Module): |
|
|
"""Swish activation function.""" |
|
|
def forward(self, x): |
|
|
return x * torch.sigmoid(x) |
|
|
|
|
|
|
|
|
class ResConv1DBlock(nn.Module): |
|
|
"""Single residual convolution block.""" |
|
|
|
|
|
def __init__(self, n_in, n_state, dilation=1, activation='relu', norm=None): |
|
|
super().__init__() |
|
|
padding = dilation |
|
|
self.norm = norm |
|
|
|
|
|
if norm == "LN": |
|
|
self.norm1 = nn.LayerNorm(n_in) |
|
|
self.norm2 = nn.LayerNorm(n_in) |
|
|
elif norm == "GN": |
|
|
self.norm1 = nn.GroupNorm(num_groups=32, num_channels=n_in, eps=1e-6, affine=True) |
|
|
self.norm2 = nn.GroupNorm(num_groups=32, num_channels=n_in, eps=1e-6, affine=True) |
|
|
elif norm == "BN": |
|
|
self.norm1 = nn.BatchNorm1d(num_features=n_in, eps=1e-6, affine=True) |
|
|
self.norm2 = nn.BatchNorm1d(num_features=n_in, eps=1e-6, affine=True) |
|
|
else: |
|
|
self.norm1 = nn.Identity() |
|
|
self.norm2 = nn.Identity() |
|
|
|
|
|
if activation == "relu": |
|
|
self.activation1 = nn.ReLU() |
|
|
self.activation2 = nn.ReLU() |
|
|
elif activation == "silu": |
|
|
self.activation1 = Swish() |
|
|
self.activation2 = Swish() |
|
|
elif activation == "gelu": |
|
|
self.activation1 = nn.GELU() |
|
|
self.activation2 = nn.GELU() |
|
|
|
|
|
self.conv1 = nn.Conv1d(n_in, n_state, 3, 1, padding, dilation) |
|
|
self.conv2 = nn.Conv1d(n_state, n_in, 1, 1, 0) |
|
|
|
|
|
def forward(self, x): |
|
|
x_orig = x |
|
|
if self.norm == "LN": |
|
|
x = self.norm1(x.transpose(-2, -1)) |
|
|
x = self.activation1(x.transpose(-2, -1)) |
|
|
else: |
|
|
x = self.norm1(x) |
|
|
x = self.activation1(x) |
|
|
x = self.conv1(x) |
|
|
if self.norm == "LN": |
|
|
x = self.norm2(x.transpose(-2, -1)) |
|
|
x = self.activation2(x.transpose(-2, -1)) |
|
|
else: |
|
|
x = self.norm2(x) |
|
|
x = self.activation2(x) |
|
|
x = self.conv2(x) |
|
|
return x + x_orig |
|
|
|
|
|
|
|
|
class Resnet1D(nn.Module): |
|
|
"""1D ResNet block composed of multiple ResConv1DBlocks.""" |
|
|
|
|
|
def __init__(self, n_in, n_depth, dilation_growth_rate=1, |
|
|
reverse_dilation=True, activation='relu', norm=None): |
|
|
super().__init__() |
|
|
blocks = [ |
|
|
ResConv1DBlock(n_in, n_in, dilation=dilation_growth_rate ** depth, |
|
|
activation=activation, norm=norm) |
|
|
for depth in range(n_depth) |
|
|
] |
|
|
if reverse_dilation: |
|
|
blocks = blocks[::-1] |
|
|
self.model = nn.Sequential(*blocks) |
|
|
|
|
|
def forward(self, x): |
|
|
return self.model(x) |
|
|
|
|
|
|
|
|
class MotionEncoder(nn.Module): |
|
|
"""Encoder for motion features with temporal downsampling.""" |
|
|
|
|
|
def __init__(self, input_dim=263, hidden_dim=512, nb_code=512, |
|
|
down_t=3, stride_t=2, depth=3, dilation_growth_rate=3, |
|
|
activation='relu', norm=None): |
|
|
super().__init__() |
|
|
blocks = [] |
|
|
filter_t, pad_t = stride_t * 2, stride_t // 2 |
|
|
blocks.append(nn.Conv1d(input_dim, hidden_dim, 3, 1, 1)) |
|
|
blocks.append(nn.ReLU()) |
|
|
for _ in range(down_t): |
|
|
block = nn.Sequential( |
|
|
nn.Conv1d(hidden_dim, hidden_dim, filter_t, stride_t, pad_t), |
|
|
Resnet1D(hidden_dim, depth, dilation_growth_rate, |
|
|
reverse_dilation=False, activation=activation, norm=norm), |
|
|
) |
|
|
blocks.append(block) |
|
|
blocks.append(nn.Conv1d(hidden_dim, nb_code, 3, 1, 1)) |
|
|
self.model = nn.Sequential(*blocks) |
|
|
|
|
|
def forward(self, x): |
|
|
return self.model(x) |
|
|
|
|
|
|
|
|
class MotionDecoder(nn.Module): |
|
|
"""Decoder for reconstructing motion from quantized features.""" |
|
|
|
|
|
def __init__(self, output_dim=263, hidden_dim=512, code_dim=512, |
|
|
down_t=3, stride_t=2, depth=3, dilation_growth_rate=3, |
|
|
activation='relu', norm=None): |
|
|
super().__init__() |
|
|
blocks = [] |
|
|
blocks.append(nn.Conv1d(code_dim, hidden_dim, 3, 1, 1)) |
|
|
blocks.append(nn.ReLU()) |
|
|
for _ in range(down_t): |
|
|
block = nn.Sequential( |
|
|
Resnet1D(hidden_dim, depth, dilation_growth_rate, |
|
|
reverse_dilation=True, activation=activation, norm=norm), |
|
|
nn.Upsample(scale_factor=2, mode='nearest'), |
|
|
nn.Conv1d(hidden_dim, hidden_dim, 3, 1, 1) |
|
|
) |
|
|
blocks.append(block) |
|
|
blocks.append(nn.Conv1d(hidden_dim, hidden_dim, 3, 1, 1)) |
|
|
blocks.append(nn.ReLU()) |
|
|
blocks.append(nn.Conv1d(hidden_dim, output_dim, 3, 1, 1)) |
|
|
self.model = nn.Sequential(*blocks) |
|
|
|
|
|
def forward(self, x): |
|
|
return self.model(x) |
|
|
|
|
|
|
|
|
class GumbelSoftmaxQuantizer(nn.Module): |
|
|
"""Gumbel-Softmax Straight-Through quantizer for VQ-VAE.""" |
|
|
|
|
|
def __init__(self, nb_code=512, code_dim=512): |
|
|
super().__init__() |
|
|
self.nb_code = nb_code |
|
|
self.code_dim = code_dim |
|
|
self.codebook = nn.Embedding(nb_code, code_dim) |
|
|
nn.init.uniform_(self.codebook.weight, -1.0 / nb_code, 1.0 / nb_code) |
|
|
self.tau = 0.4 |
|
|
|
|
|
def quantize(self, x): |
|
|
"""Quantize encoder output to discrete indices.""" |
|
|
return x.argmax(dim=-1) |
|
|
|
|
|
def dequantize(self, indices): |
|
|
"""Convert indices back to embeddings.""" |
|
|
return self.codebook(indices) |
|
|
|
|
|
def forward(self, x_encoder): |
|
|
"""Forward pass with Gumbel-Softmax sampling.""" |
|
|
N, C, T = x_encoder.shape |
|
|
x = x_encoder.permute(0, 2, 1).contiguous().view(-1, C) |
|
|
|
|
|
|
|
|
y_hard_st = F.gumbel_softmax(x, tau=self.tau, hard=True, dim=-1) |
|
|
x_quantized = torch.matmul(y_hard_st, self.codebook.weight) |
|
|
|
|
|
return x_quantized.view(N, T, -1).permute(0, 2, 1).contiguous() |
|
|
|
|
|
|
|
|
class MotionTokenizer(nn.Module): |
|
|
""" |
|
|
DVQ-GSST Motion Tokenizer. |
|
|
|
|
|
Converts continuous motion features (263-dim HumanML3D format) to discrete tokens. |
|
|
|
|
|
Args: |
|
|
config: GeoMotionGPTConfig containing motion tokenizer parameters |
|
|
|
|
|
Example: |
|
|
```python |
|
|
motion = torch.randn(1, 100, 263) # (batch, time, features) |
|
|
tokens = motion_tokenizer.encode(motion) # (batch, time//8) |
|
|
``` |
|
|
""" |
|
|
|
|
|
def __init__(self, config: GeoMotionGPTConfig): |
|
|
super().__init__() |
|
|
self.config = config |
|
|
|
|
|
self.encoder = MotionEncoder( |
|
|
input_dim=config.motion_input_dim, |
|
|
hidden_dim=config.motion_hidden_dim, |
|
|
nb_code=config.motion_vocab_size, |
|
|
down_t=config.motion_down_t, |
|
|
depth=config.motion_depth, |
|
|
dilation_growth_rate=config.motion_dilation_growth_rate, |
|
|
) |
|
|
|
|
|
self.decoder = MotionDecoder( |
|
|
output_dim=config.motion_input_dim, |
|
|
hidden_dim=config.motion_hidden_dim, |
|
|
code_dim=config.motion_vocab_size, |
|
|
down_t=config.motion_down_t, |
|
|
depth=config.motion_depth, |
|
|
dilation_growth_rate=config.motion_dilation_growth_rate, |
|
|
) |
|
|
|
|
|
self.quantizer = GumbelSoftmaxQuantizer( |
|
|
nb_code=config.motion_vocab_size, |
|
|
code_dim=config.motion_vocab_size, |
|
|
) |
|
|
|
|
|
def encode(self, motion: torch.Tensor) -> torch.Tensor: |
|
|
""" |
|
|
Encode motion features to discrete tokens. |
|
|
|
|
|
Args: |
|
|
motion: Motion features of shape (batch, time, 263) |
|
|
|
|
|
Returns: |
|
|
Token indices of shape (batch, time // downsample_ratio) |
|
|
""" |
|
|
|
|
|
x = motion.permute(0, 2, 1).float() |
|
|
|
|
|
|
|
|
x_enc = self.encoder(x) |
|
|
|
|
|
|
|
|
x_enc = x_enc.permute(0, 2, 1).contiguous() |
|
|
N, T, C = x_enc.shape |
|
|
|
|
|
|
|
|
indices = self.quantizer.quantize(x_enc.view(-1, C)) |
|
|
return indices.view(N, T) |
|
|
|
|
|
def decode(self, tokens: torch.Tensor) -> torch.Tensor: |
|
|
""" |
|
|
Decode tokens back to motion features. |
|
|
|
|
|
Args: |
|
|
tokens: Token indices of shape (batch, time') |
|
|
|
|
|
Returns: |
|
|
Motion features of shape (batch, time, 263) |
|
|
""" |
|
|
|
|
|
x = self.quantizer.dequantize(tokens) |
|
|
|
|
|
|
|
|
x = x.permute(0, 2, 1).contiguous() |
|
|
|
|
|
|
|
|
x_out = self.decoder(x) |
|
|
|
|
|
|
|
|
return x_out.permute(0, 2, 1) |
|
|
|
|
|
def forward(self, motion: torch.Tensor): |
|
|
"""Forward pass for training (encode -> quantize -> decode).""" |
|
|
x = motion.permute(0, 2, 1).float() |
|
|
x_enc = self.encoder(x) |
|
|
x_quant = self.quantizer(x_enc) |
|
|
x_dec = self.decoder(x_quant) |
|
|
return x_dec.permute(0, 2, 1) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class GeoMotionGPTPreTrainedModel(PreTrainedModel): |
|
|
"""Base class for GeoMotionGPT models.""" |
|
|
|
|
|
config_class = GeoMotionGPTConfig |
|
|
base_model_prefix = "geomotiongpt" |
|
|
supports_gradient_checkpointing = True |
|
|
|
|
|
def _init_weights(self, module): |
|
|
"""Initialize weights.""" |
|
|
if isinstance(module, (nn.Linear, nn.Conv1d)): |
|
|
module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) |
|
|
if module.bias is not None: |
|
|
module.bias.data.zero_() |
|
|
elif isinstance(module, nn.Embedding): |
|
|
module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) |
|
|
if module.padding_idx is not None: |
|
|
module.weight.data[module.padding_idx].zero_() |
|
|
elif isinstance(module, nn.LayerNorm): |
|
|
module.bias.data.zero_() |
|
|
module.weight.data.fill_(1.0) |
|
|
|
|
|
|
|
|
class GeoMotionGPTForCausalLM(GeoMotionGPTPreTrainedModel): |
|
|
""" |
|
|
GeoMotionGPT Model for motion-to-text generation. |
|
|
|
|
|
This model combines: |
|
|
1. A VQ-VAE motion tokenizer (DVQ-GSST) for converting motion to discrete tokens |
|
|
2. A fine-tuned GPT-2 model for generating text from motion tokens |
|
|
|
|
|
Example: |
|
|
```python |
|
|
from transformers import AutoModelForCausalLM |
|
|
import torch |
|
|
|
|
|
# Load model |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
"zy22b/GeoMotionGPT", |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
# Access motion tokenizer |
|
|
motion_tokenizer = model.motion_tokenizer |
|
|
|
|
|
# Tokenize motion (batch, time, 263) -> (batch, tokens) |
|
|
motion = torch.randn(1, 100, 263) |
|
|
motion_tokens = motion_tokenizer.encode(motion) |
|
|
|
|
|
# Generate text from motion tokens |
|
|
text = model.generate_text(motion_tokens) |
|
|
``` |
|
|
""" |
|
|
|
|
|
_tied_weights_keys = ["lm_head.weight"] |
|
|
|
|
|
def __init__(self, config: GeoMotionGPTConfig): |
|
|
super().__init__(config) |
|
|
|
|
|
|
|
|
self.motion_tokenizer = MotionTokenizer(config) |
|
|
|
|
|
|
|
|
gpt2_config = GPT2Config( |
|
|
vocab_size=config.vocab_size, |
|
|
n_positions=config.n_positions, |
|
|
n_embd=config.n_embd, |
|
|
n_layer=config.n_layer, |
|
|
n_head=config.n_head, |
|
|
n_inner=config.n_inner, |
|
|
activation_function=config.activation_function, |
|
|
resid_pdrop=config.resid_pdrop, |
|
|
embd_pdrop=config.embd_pdrop, |
|
|
attn_pdrop=config.attn_pdrop, |
|
|
layer_norm_epsilon=config.layer_norm_epsilon, |
|
|
initializer_range=config.initializer_range, |
|
|
bos_token_id=config.bos_token_id, |
|
|
eos_token_id=config.eos_token_id, |
|
|
) |
|
|
|
|
|
|
|
|
self.language_model = GPT2LMHeadModel(gpt2_config) |
|
|
|
|
|
|
|
|
mot_embed_dim = int(config.n_embd // config.n_head * config.mot_factor) * config.n_head |
|
|
self.motion_embed = nn.Embedding( |
|
|
config.motion_vocab_size + 3, |
|
|
mot_embed_dim |
|
|
) |
|
|
self.motion_head = nn.Linear(mot_embed_dim, config.motion_vocab_size + 3, bias=False) |
|
|
|
|
|
|
|
|
self.motion_to_text_proj = nn.Linear(mot_embed_dim, config.n_embd) |
|
|
self.text_to_motion_proj = nn.Linear(config.n_embd, mot_embed_dim) |
|
|
|
|
|
|
|
|
self.post_init() |
|
|
|
|
|
def get_input_embeddings(self): |
|
|
return self.language_model.transformer.wte |
|
|
|
|
|
def set_input_embeddings(self, value): |
|
|
self.language_model.transformer.wte = value |
|
|
|
|
|
def get_output_embeddings(self): |
|
|
return self.language_model.lm_head |
|
|
|
|
|
def set_output_embeddings(self, new_embeddings): |
|
|
self.language_model.lm_head = new_embeddings |
|
|
|
|
|
def encode_motion(self, motion: torch.Tensor) -> torch.Tensor: |
|
|
""" |
|
|
Encode motion features to discrete tokens. |
|
|
|
|
|
Args: |
|
|
motion: Motion features of shape (batch, time, 263) |
|
|
|
|
|
Returns: |
|
|
Token indices of shape (batch, time // 8) |
|
|
""" |
|
|
return self.motion_tokenizer.encode(motion) |
|
|
|
|
|
def forward( |
|
|
self, |
|
|
input_ids: Optional[torch.LongTensor] = None, |
|
|
attention_mask: Optional[torch.FloatTensor] = None, |
|
|
past_key_values: Optional[Tuple[Tuple[torch.Tensor]]] = None, |
|
|
labels: Optional[torch.LongTensor] = None, |
|
|
use_cache: Optional[bool] = None, |
|
|
output_attentions: Optional[bool] = None, |
|
|
output_hidden_states: Optional[bool] = None, |
|
|
return_dict: Optional[bool] = None, |
|
|
**kwargs |
|
|
): |
|
|
""" |
|
|
Forward pass through the language model. |
|
|
|
|
|
For motion-to-text generation, use the `generate_text` method instead. |
|
|
""" |
|
|
return self.language_model( |
|
|
input_ids=input_ids, |
|
|
attention_mask=attention_mask, |
|
|
past_key_values=past_key_values, |
|
|
labels=labels, |
|
|
use_cache=use_cache, |
|
|
output_attentions=output_attentions, |
|
|
output_hidden_states=output_hidden_states, |
|
|
return_dict=return_dict, |
|
|
) |
|
|
|
|
|
def prepare_inputs_for_generation(self, input_ids, past_key_values=None, **kwargs): |
|
|
"""Prepare inputs for text generation.""" |
|
|
return self.language_model.prepare_inputs_for_generation( |
|
|
input_ids, past_key_values=past_key_values, **kwargs |
|
|
) |
|
|
|
|
|
@torch.no_grad() |
|
|
def generate_text( |
|
|
self, |
|
|
motion_tokens: torch.Tensor, |
|
|
max_new_tokens: int = 128, |
|
|
num_beams: int = 4, |
|
|
temperature: float = 0.7, |
|
|
top_p: float = 0.9, |
|
|
do_sample: bool = True, |
|
|
**kwargs |
|
|
) -> List[str]: |
|
|
""" |
|
|
Generate text descriptions from motion tokens. |
|
|
|
|
|
Args: |
|
|
motion_tokens: Motion token indices of shape (batch, seq_len) |
|
|
max_new_tokens: Maximum number of new tokens to generate |
|
|
num_beams: Number of beams for beam search |
|
|
temperature: Sampling temperature |
|
|
top_p: Top-p sampling parameter |
|
|
do_sample: Whether to use sampling |
|
|
|
|
|
Returns: |
|
|
List of generated text strings |
|
|
""" |
|
|
device = motion_tokens.device |
|
|
batch_size = motion_tokens.shape[0] |
|
|
|
|
|
|
|
|
motion_offset = self.config.text_vocab_size |
|
|
input_ids = motion_tokens + motion_offset |
|
|
|
|
|
|
|
|
bos_tokens = torch.full( |
|
|
(batch_size, 1), |
|
|
self.config.bos_token_id, |
|
|
dtype=torch.long, |
|
|
device=device |
|
|
) |
|
|
input_ids = torch.cat([bos_tokens, input_ids], dim=1) |
|
|
|
|
|
|
|
|
outputs = self.language_model.generate( |
|
|
input_ids=input_ids, |
|
|
max_new_tokens=max_new_tokens, |
|
|
num_beams=num_beams, |
|
|
temperature=temperature, |
|
|
top_p=top_p, |
|
|
do_sample=do_sample, |
|
|
pad_token_id=self.config.pad_token_id, |
|
|
eos_token_id=self.config.eos_token_id, |
|
|
**kwargs |
|
|
) |
|
|
|
|
|
|
|
|
generated_ids = outputs[:, input_ids.shape[1]:] |
|
|
|
|
|
|
|
|
|
|
|
return generated_ids |
|
|
|
|
|
|
|
|
|
|
|
GeoMotionGPTConfig.register_for_auto_class() |
|
|
GeoMotionGPTForCausalLM.register_for_auto_class("AutoModelForCausalLM") |
|
|
|