MMRM / README.md
rexera's picture
0-shot pipeline test
87224ba
---
title: MMRM
emoji: 🐠
colorFrom: yellow
colorTo: blue
sdk: gradio
sdk_version: 6.5.1
app_file: app.py
pinned: false
license: gpl-3.0
short_description: 'Restoring Ancient Ideograph: A Multimodal Multitask Neural N'
---
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
## Interactive Demo
This Gradio app demonstrates the restoration capabilities of the MMRM model compared to textual and visual baselines on real-world damaged character data.
### Features
- **Real-world Data**: Select from samples in the `data/real` directory.
- **Model Comparison**:
- **Zero-shot Baseline**: Pre-trained GuwenBERT (Works out-of-the-box without training).
- **Textual Baseline**: Fine-tuned RoBERTa.
- **Visual Baseline**: ResNet50.
- **MMRM**: Our proposed Multimodal Multitask Restoring Model.
- **Intermediate Visualization**: Shows the restored image generated by the MMRM capability.
### Running the Demo
1. **Deploy to Hugging Face Spaces**:
- Create a new Space on Hugging Face (SDK: Gradio).
- Upload the contents of this `demo` folder to the Space repository.
- Upload your model checkpoints to the `checkpoints/` folder in the Space.
- `checkpoints/phase2_mmrm_best.pt`
- `checkpoints/phase1_roberta_finetuned.pt`
- `checkpoints/baseline_img.pt`
*Note: Even without checkpoints, the demo will run using the Zero-shot Baseline (downloaded automatically).*
2. **Local Testing**:
- Install requirements: `pip install -r requirements.txt`
- Run: `python app.py` (assuming you are inside the `demo` directory)