MMRM / README.md
rexera's picture
0-shot pipeline test
87224ba

A newer version of the Gradio SDK is available: 6.6.0

Upgrade
metadata
title: MMRM
emoji: 🐠
colorFrom: yellow
colorTo: blue
sdk: gradio
sdk_version: 6.5.1
app_file: app.py
pinned: false
license: gpl-3.0
short_description: 'Restoring Ancient Ideograph: A Multimodal Multitask Neural N'

Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference

Interactive Demo

This Gradio app demonstrates the restoration capabilities of the MMRM model compared to textual and visual baselines on real-world damaged character data.

Features

  • Real-world Data: Select from samples in the data/real directory.
  • Model Comparison:
    • Zero-shot Baseline: Pre-trained GuwenBERT (Works out-of-the-box without training).
    • Textual Baseline: Fine-tuned RoBERTa.
    • Visual Baseline: ResNet50.
    • MMRM: Our proposed Multimodal Multitask Restoring Model.
  • Intermediate Visualization: Shows the restored image generated by the MMRM capability.

Running the Demo

  1. Deploy to Hugging Face Spaces:

    • Create a new Space on Hugging Face (SDK: Gradio).
    • Upload the contents of this demo folder to the Space repository.
    • Upload your model checkpoints to the checkpoints/ folder in the Space.
      • checkpoints/phase2_mmrm_best.pt
      • checkpoints/phase1_roberta_finetuned.pt
      • checkpoints/baseline_img.pt

    Note: Even without checkpoints, the demo will run using the Zero-shot Baseline (downloaded automatically).

  2. Local Testing:

    • Install requirements: pip install -r requirements.txt
    • Run: python app.py (assuming you are inside the demo directory)