text
stringlengths
0
81
[0] noisy_images_proj.weight (shape: (3584, 64), trainable: True)
[1] cond_proj.weight (shape: (3584, 3584), trainable: True)
[2] t_embedder.mlp.0.weight (shape: (3584, 256), trainable: True)
[3] t_embedder.mlp.2.weight (shape: (3584, 3584), trainable: True)
[4] layers.0.ffn.gate_proj.weight (shape: (10752, 3584), trainable: True)
[5] layers.0.ffn.up_proj.weight (shape: (10752, 3584), trainable: True)
[6] layers.0.ffn.down_proj.weight (shape: (3584, 10752), trainable: True)
[7] layers.0.norm.weight (shape: (3584,), trainable: True)
[8] layers.0.adaLN_modulation.1.weight (shape: (10752, 3584), trainable: True)
[9] layers.1.ffn.gate_proj.weight (shape: (10752, 3584), trainable: True)
[10] layers.1.ffn.up_proj.weight (shape: (10752, 3584), trainable: True)
[11] layers.1.ffn.down_proj.weight (shape: (3584, 10752), trainable: True)
[12] layers.1.norm.weight (shape: (3584,), trainable: True)
[13] layers.1.adaLN_modulation.1.weight (shape: (10752, 3584), trainable: True)
[14] layers.2.ffn.gate_proj.weight (shape: (10752, 3584), trainable: True)
[15] layers.2.ffn.up_proj.weight (shape: (10752, 3584), trainable: True)
[16] layers.2.ffn.down_proj.weight (shape: (3584, 10752), trainable: True)
[17] layers.2.norm.weight (shape: (3584,), trainable: True)
[18] layers.2.adaLN_modulation.1.weight (shape: (10752, 3584), trainable: True)
[19] layers.3.ffn.gate_proj.weight (shape: (10752, 3584), trainable: True)
[20] layers.3.ffn.up_proj.weight (shape: (10752, 3584), trainable: True)
[21] layers.3.ffn.down_proj.weight (shape: (3584, 10752), trainable: True)
[22] layers.3.norm.weight (shape: (3584,), trainable: True)
[23] layers.3.adaLN_modulation.1.weight (shape: (10752, 3584), trainable: True)
[24] final_layer.linear.weight (shape: (64, 3584), trainable: True)
[25] final_layer.adaLN_modulation.1.weight (shape: (7168, 3584), trainable: True)
LICENSE
README.md
pyproject.toml
src/vibevoice/modular/__init__.py
src/vibevoice/modular/configuration_vibevoice.py
src/vibevoice/modular/modeling_vibevoice.py
src/vibevoice/modular/modeling_vibevoice_inference.py
src/vibevoice/modular/modular_vibevoice_diffusion_head.py
src/vibevoice/modular/modular_vibevoice_text_tokenizer.py
src/vibevoice/modular/modular_vibevoice_tokenizer.py
src/vibevoice/modular/streamer.py
src/vibevoice/processor/__init__.py
src/vibevoice/processor/vibevoice_processor.py
src/vibevoice/processor/vibevoice_tokenizer_processor.py
src/vibevoice/schedule/__init__.py
src/vibevoice/schedule/dpm_solver.py
src/vibevoice/schedule/timestep_sampler.py
src/vibevoice/scripts/convert_nnscaler_checkpoint_to_transformers.py
src/vibevoice_finetuning.egg-info/PKG-INFO
src/vibevoice_finetuning.egg-info/SOURCES.txt
src/vibevoice_finetuning.egg-info/dependency_links.txt
src/vibevoice_finetuning.egg-info/requires.txt
src/vibevoice_finetuning.egg-info/top_level.txt
numpy~=1.26.0
resampy==0.4.3
librosa==0.11.0
s3tokenizer
torch
torchaudio
transformers
datasets>=2.18.0
diffusers==0.29.0
resemble-perth==1.0.1
omegaconf==2.3.0
conformer==0.3.2
safetensors==0.5.3
peft>=0.11.0
tensorboard>=2.12
wandb
vibevoice
setuptools==75.2.0
types-setuptools==80.9.0.20250822
requirements-parser==0.9.0
pip==24.1.2
cfgv==3.5.0
torchcodec==0.10.0
pre_commit==4.5.1
transformers==4.51.3
identify==2.6.16
virtualenv==20.36.1
multiprocess==0.70.16
diffusers==0.29.0
conformer==0.3.2
numpy==1.26.4
peft==0.7.1
datasets==2.21.0
resampy==0.4.3
tokenizers==0.21.4
nodeenv==1.10.0
distlib==0.4.0
resemble-perth==1.0.1
safetensors==0.5.3
fsspec==2024.6.1
vibevoice-finetuning==0.1.0
s3tokenizer==0.3.0
dill==0.3.8
click==8.3.1
regex==2025.11.3
joblib==1.5.3
nltk==3.9.2
tqdm==4.67.1
pytools==2025.2.5
pycuda==2025.1.2