File size: 9,385 Bytes
f761cd6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
443d8b1
 
 
 
 
 
 
 
f761cd6
443d8b1
 
 
 
 
 
 
 
 
 
 
 
 
f761cd6
443d8b1
f761cd6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
"""
AI Fitness Coach - Hugging Face Spaces Demo
Fine-tuned persona-based feedback system
"""

import gradio as gr
import torch
import os
import json
import tempfile
import cv2
import numpy as np
from pathlib import Path
from transformers import AutoTokenizer, AutoModelForCausalLM

# Import your existing modules
import sys
PROJECT_ROOT = Path(__file__).parent
sys.path.insert(0, str(PROJECT_ROOT))

from fitness_coach.video_processing import process_video
from fitness_coach.scoring import calculate_overall_score


# Model repository on Hugging Face Hub
MODEL_REPO = "rlogh/fitness-coach-personas"

# Load fine-tuned models
PERSONAS = {
    "Hype Beast πŸ”₯": "persona_hype_beast",
    "Data Scientist πŸ“Š": "persona_data_scientist", 
    "No-Nonsense Pro πŸ’ͺ": "persona_no-nonsense_pro",
    "Mindful Aligner 🧘": "persona_mindful_aligner"
}

models = {}
tokenizers = {}

def load_models():
    """Load all fine-tuned persona models from Hugging Face Hub"""
    from huggingface_hub import hf_hub_download
    import os
    
    for persona_name, model_dir in PERSONAS.items():
        # Download model from HF Hub (cached automatically)
        try:
            print(f"Loading {persona_name} from Hugging Face Hub...")
            model_path = f"{MODEL_REPO}/{model_dir}"
            # Use slow tokenizer to avoid corrupted tokenizer.json
            tokenizers[persona_name] = AutoTokenizer.from_pretrained(model_path, use_fast=False)
            # Try loading with trust_remote_code and low_cpu_mem_usage
            try:
                models[persona_name] = AutoModelForCausalLM.from_pretrained(
                    model_path,
                    torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
                    device_map="auto" if torch.cuda.is_available() else None,
                    trust_remote_code=True,
                    low_cpu_mem_usage=True
                )
                print(f"βœ“ Loaded {persona_name}")
        except Exception as e:
            print(f"βœ— Failed to load {persona_name}: {e}")
            import traceback
            traceback.print_exc()

# Load models on startup
print("Loading fine-tuned models...")
load_models()
print(f"Loaded {len(models)} persona models")


def generate_feedback(persona_name, input_report):
    """Generate persona-specific feedback using fine-tuned model"""
    if persona_name not in models:
        return f"Model for {persona_name} not loaded"
    
    model = models[persona_name]
    tokenizer = tokenizers[persona_name]
    
    # Format prompt as trained
    prompt = f"<|persona|>{persona_name}<|input|>{input_report}<|output|>"
    
    # Generate
    inputs = tokenizer(prompt, return_tensors="pt", truncation=True, max_length=256)
    
    # Use CPU to avoid CUDA token ID issues
    model_device = model.cpu()
    
    with torch.no_grad():
        outputs = model_device.generate(
            **inputs,
            max_new_tokens=500,  # Enough for complete responses
            temperature=0.9,  # Higher temp = more creative
            top_p=0.95,
            top_k=50,
            do_sample=True,
            repetition_penalty=1.2,  # Prevent repetition
            no_repeat_ngram_size=3,  # No 3-word repetitions
            pad_token_id=tokenizer.eos_token_id,
            eos_token_id=tokenizer.eos_token_id
        )
    
    # Decode and extract only the generated part
    full_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
    
    # Extract only the output part after <|output|>
    if "<|output|>" in full_text:
        feedback = full_text.split("<|output|>")[-1].strip()
    else:
        feedback = full_text
    
    return feedback


def analyze_video(video_file, persona_choice, reference_video_choice):
    """Main analysis function"""
    if video_file is None:
        return "Please upload a video", "", ""
    
    try:
        # Process video and get pose analysis
        video_path = video_file
        
        # Download reference data from HF Dataset
        from huggingface_hub import hf_hub_download
        
        # Map exercise to reference
        exercise_map = {
            "Squat": "squat",
            "Pushup": "pushup",
            "Pullup": "pullup"
        }
        exercise_id = exercise_map.get(reference_video_choice, "pushup")
        
        # Download reference keypoints from dataset
        try:
            ref_3d_path = hf_hub_download(
                repo_id="rlogh/fitness-coach-references",
                filename=f"{exercise_id}/keypoints_3D.npz",
                repo_type="dataset"
            )
            ref_results = {'keypoints_3d': np.load(ref_3d_path)['reconstruction']}
        except Exception as e:
            print(f"Warning: Could not load reference: {e}")
            ref_results = None
        
        # Process user video
        print(f"Processing {video_path}...")
        user_results = process_video(video_path)
        
        if not user_results or 'keypoints_3d' not in user_results:
            return "Failed to process video - no pose detected", "", ""
        
        # Calculate scores
        scores = calculate_overall_score(user_results, ref_results)
        
        # Create analysis report
        report = f"""
Exercise Analysis Report

Overall Score: {scores.get('overall_score', 0):.1f}/100

Body Part Breakdown:
- Head/Neck: {scores.get('head_score', 0):.1f}/100
- Shoulders: {scores.get('shoulder_score', 0):.1f}/100  
- Arms: {scores.get('arm_score', 0):.1f}/100
- Torso: {scores.get('torso_score', 0):.1f}/100
- Legs: {scores.get('leg_score', 0):.1f}/100

Key Issues:
{scores.get('issues', 'Form looks good!')}
"""
        
        # Generate persona feedback using fine-tuned model
        feedback = generate_feedback(persona_choice, report)
        
        # Create detailed JSON
        analysis_json = json.dumps({
            "scores": scores,
            "persona": persona_choice,
            "frames_analyzed": len(user_results.get('keypoints_3d', [])),
            "reference_used": reference_video_choice
        }, indent=2)
        
        return report, feedback, analysis_json
        
    except Exception as e:
        import traceback
        error_msg = f"Error analyzing video: {str(e)}\n{traceback.format_exc()}"
        return error_msg, "", ""


# Create Gradio interface
with gr.Blocks(title="AI Fitness Coach", theme=gr.themes.Soft()) as demo:
    gr.Markdown("""
    # πŸ‹οΈ AI Fitness Coach with Fine-Tuned Personas
    
    Upload your exercise video and get personalized feedback from our AI coach!
    
    **Features:**
    - 3D pose estimation and analysis
    - 4 fine-tuned persona models (trained on GPT-4o synthetic data)
    - Real-time scoring and feedback
    """)
    
    with gr.Row():
        with gr.Column():
            video_input = gr.Video(label="Upload Your Exercise Video")
            
            persona_select = gr.Radio(
                choices=list(PERSONAS.keys()),
                value=list(PERSONAS.keys())[0],
                label="Choose Your Coach Persona",
                info="Each persona has a unique coaching style (fine-tuned model)"
            )
            
            reference_select = gr.Radio(
                choices=["Squat", "Pushup", "Pullup"],
                value="Squat",
                label="Exercise Type",
                info="Select reference exercise for comparison"
            )
            
            analyze_btn = gr.Button("Analyze My Form 🎯", variant="primary")
        
        with gr.Column():
            report_output = gr.Textbox(
                label="πŸ“Š Analysis Report",
                lines=12,
                placeholder="Your detailed analysis will appear here..."
            )
            
            feedback_output = gr.Textbox(
                label="πŸ’¬ Persona Feedback (Fine-Tuned Model)",
                lines=10,
                placeholder="Personalized coaching feedback will appear here..."
            )
            
            json_output = gr.JSON(label="πŸ“‹ Detailed Results (JSON)")
    
    # Example videos
    gr.Markdown("### πŸ“Ή Example Videos")
    gr.Examples(
        examples=[
            ["sample_squat.mp4", "Hype Beast πŸ”₯", "Squat"],
            ["sample_pushup.mp4", "Data Scientist πŸ“Š", "Pushup"],
        ],
        inputs=[video_input, persona_select, reference_select],
        label="Try these examples"
    )
    
    # Connect button
    analyze_btn.click(
        fn=analyze_video,
        inputs=[video_input, persona_select, reference_select],
        outputs=[report_output, feedback_output, json_output]
    )
    
    gr.Markdown("""
    ---
    ### πŸ”¬ About the Models
    
    This app uses **4 fine-tuned DistilGPT-2 models**, each trained on persona-specific synthetic data:
    - **Training Data**: 320 examples (80 per persona) generated with GPT-4o
    - **Base Model**: DistilGPT-2
    - **Training**: 1000 steps per persona with FP16 mixed precision
    - **Personas**: Hype Beast, Data Scientist, No-Nonsense Pro, Mindful Aligner
    
    ### πŸš€ Technology Stack
    - **3D Pose Estimation**: PoseFormer (transformer-based)
    - **Video Processing**: MediaPipe + OpenCV
    - **Fine-Tuned Models**: Hugging Face Transformers
    - **Framework**: Gradio + PyTorch
    """)

# Launch
if __name__ == "__main__":
    demo.launch()