gune-amp / README.md
sgune's picture
Upload folder using huggingface_hub
4922b62 verified
---
license: mit
tags:
- audio
- pytorch
- torchscript
- guitar-amp-simulation
- real-time
inference:
framework: pytorch
task: audio-to-audio
inputs:
- name: input
type: float[]
description: "Input waveform or features (e.g. [batch, channels, samples])"
outputs:
- name: output
type: float[]
description: "Output waveform or processed features"
---
## Usage
This is a model I trained to mimic a JCM 800 AMP. It doesn't sound very good, but as a first pass, I'm glad I have it.
![](infer.PNG)
Download [GuneAmp.exe](GuneAmp.exe) and try running your own conversion.
Read my notes [GuneAmpNotes](GuneAmpNotes.pdf)
## Using the TorchScript Model from Hugging Face
If you wish to use the TorchScript version of the model directly, you can download it from Hugging Face and load it using the following Python code.
First, ensure you have the necessary libraries installed:
```bash
pip install torch huggingface_hub
```
Then, use the following Python code to load and use the model:
```python
import torch
from huggingface_hub import hf_hub_download
model_id = 'sgune/gune-amp'
model_filename = 'metal_amp_v2_ts.pt'
model_path = hf_hub_download(repo_id=model_id, filename=model_filename)
#LOAD the model on GPU or CPU
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
print(f"Loading model on device: {device}")
model = torch.jit.load(model_path, map_location=device)
model.eval()
print("Model loaded successfully!")
input_size = 1024
dummy_input = torch.randn(1, input_size, dtype=torch.float32).to(device)
print(f"Running inference with dummy input of shape: {dummy_input.shape}")
with torch.no_grad(): # Disable gradient calculations for inference
output = model(dummy_input)
print("Inference complete!")
print("Example output shape:", output.shape)
print("Example output values:", output)
```
## COMING SOON
`infer.py`, `model.py`, `train.py` and `config.py` deepdives.