Spaces:
Paused
Paused
File size: 8,252 Bytes
e22d944 d08ac50 e22d944 4a7f08c 9173b01 c44caaa 4a7f08c 9523e3c 4a7f08c 03ccf7a 41a4cd6 4a7f08c 6ed3766 4a7f08c c44caaa 4a7f08c 87f3cb8 4a7f08c a422c4e c44caaa 4a7f08c a422c4e 4a7f08c a422c4e | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 | ---
title: "My RecoFM AI Agent Demo"
emoji: "π¬"
colorFrom: pink
colorTo: red
sdk: gradio
sdk_version: 4.31.0
app_file: app.py
license: apache-2.0
tags:
- agent-demo-track
- recommender-system
- gradio
---
# RecoFM: Foundation Recommender model
**[Link to the YouTube demo video](https://youtu.be/JMgwnwxTa_w)**
# Tag: **agent-demo-track**
Recent research showed that combining LLM embeddings with graph convolution layers over them results in remarkable zero-shot performance, surpassing traditional graph-based approaches such as LightGCN, and naturally supporting user intention queries.
In this demo, we demonstrate how to build such a system. To showcase its ability, we've chosen the Amazon Movies 2023 dataset. However, the main beauty of this approach is that it can be applied to any type of recommendation data and delivers strong zero-shot performance **without any additional training**.
## Architecture Overview
The system follows a multi-stage pipeline that processes movie data and user current preferences to generate personalized recommendations:
[//]: # (```mermaid)
[//]: # (graph TD)
[//]: # ( %% ========================)
[//]: # ( %% Data Preparation Section)
[//]: # ( %% ========================)
[//]: # ( A["π½οΈ Movie Titles Dataset"] --> B["Mistral LLM Embedding<br/>(Text Representation)"])
[//]: # ( B --> C["Graph Convolution Layers<br/>(Relationship Learning)"])
[//]: # ( C --> D["π Graph-Enhanced Embeddings"])
[//]: # ( )
[//]: # ( %% ========================)
[//]: # ( %% User Input Section)
[//]: # ( %% ========================)
[//]: # ( F["π¬ User's Liked Movies"] --> G["User Preference Profile<br/>(Average of Graph-Enhanced<br/>Movie Embeddings)"])
[//]: # ( H["π Natural Language Query<br/>(e.g. 'funny sci-fi movies')"] --> I["Query Embedding"])
[//]: # ( G --> J["π§© Combined User Vector"])
[//]: # ( I --> J)
[//]: # ( )
[//]: # ( %% ========================)
[//]: # ( %% System Flow)
[//]: # ( %% ========================)
[//]: # ( D --> K["π Similarity Matching"])
[//]: # ( J --> K)
[//]: # ( K --> L["π Top 100 Candidates"])
[//]: # ( L --> M["π€ AI Agent Ranking<br/>(Context-Aware Filtering)"])
[//]: # ( M --> N["π― Final Recommendations<br/>(Top 10 Movies)"])
[//]: # ( )
[//]: # ( %% ========================)
[//]: # ( %% Visual Grouping)
[//]: # ( %% ========================)
[//]: # ( subgraph "Data Preparation Pipeline")
[//]: # ( A)
[//]: # ( B)
[//]: # ( C)
[//]: # ( D)
[//]: # ( end)
[//]: # ( )
[//]: # ( subgraph "User Understanding")
[//]: # ( F)
[//]: # ( H)
[//]: # ( G)
[//]: # ( I)
[//]: # ( J)
[//]: # ( end)
[//]: # ( )
[//]: # ( subgraph "Two-Stage Retrieval")
[//]: # ( K)
[//]: # ( L)
[//]: # ( end)
[//]: # ( )
[//]: # ( subgraph "Intelligent Ranking")
[//]: # ( M)
[//]: # ( N)
[//]: # ( end)
[//]: # ( )
[//]: # ( %% ========================)
[//]: # ( %% Styling)
[//]: # ( %% ========================)
[//]: # ( style A fill:#e1f5fe,stroke:#039be5)
[//]: # ( style N fill:#c8e6c9,stroke:#2e7d32)
[//]: # ( style K fill:#fff3e0,stroke:#fb8c00)
[//]: # ( style M fill:#fce4ec,stroke:#e91e63)
[//]: # ( )
[//]: # ( classDef data fill:#e1f5fe,stroke:#039be5)
[//]: # ( classDef result fill:#c8e6c9,stroke:#2e7d32)
[//]: # ( classDef process fill:#fff3e0,stroke:#fb8c00)
[//]: # ( classDef ai fill:#fce4ec,stroke:#e91e63)
[//]: # ( )
[//]: # ( class A data)
[//]: # ( class N result)
[//]: # ( class K,L process)
[//]: # ( class M ai)
[//]: # (```)
<!--  -->
<img src="./diagram.svg" alt="Diagram" width="100%">
### Pipeline Stages
The complete preprocessing pipeline code is available in our GitHub repository:
π **[Preprocessing Code](https://github.com/RecoFM/HF-hackathon2025)**
#### 1. Preprocessing Stage
- **Movie Title Processing**: Raw movie titles are converted into semantic embeddings using Mistral AI
- **Graph Convolution Enhancement**: Movie embeddings are enriched through Graph Convolution Layers (GCL) that capture user interaction patterns and movie relationships
#### 2. User Input Processing
- **Movie Selection Encoding**: User-selected movies are converted into collaborative embeddings
- **Natural Language Understanding**: User query is processed through Mistral AI to create a query embedding
- **Preference Fusion**: Movie selections and text queries are combined using a weighted approach (Ξ± parameter)
#### 3. Retrieval Phase
- **Similarity Computation**: Cosine similarity is calculated between a combined user vector and all candidate movie embeddings
- **Candidate Selection**: Top 100 most similar movies are retrieved as candidates
#### 4. Ranking Phase
- **AI Agent Ranking**: An intelligent agent analyzes the top 100 candidates considering user preferences, viewing history, and context
- **Final Selection**: The agent selects and ranks the final top 10 personalized recommendations
## Features
### Dual Embedding Types
- **Pure Language Model (LLM) Embeddings**
Generated for each movie title using Mistral AI.
- **Graph-Enhanced Embeddings (LLM + GCL)**
Combines language understanding with user interaction patterns to enrich the embeddings.
---
### Hybrid Input
- **Movie Selection**
Select movies you've previously enjoyed.
- **Natural Language Query**
Describe the kind of movie you're looking for in natural language.
- **Weight Adjustment (Ξ±)**
Adjust the balance between your movie selections and your text description to personalize the recommendations.
---
### Key Technical Components
- **Mistral AI Integration**: Leverages Mistral's language models for both movie title understanding and user query processing
- **Graph Convolutional Networks**: Enhances embeddings by incorporating collaborative filtering signals through user-movie interaction graphs
- **Hybrid Recommendation Strategy**: Combines content-based (semantic similarity) and collaborative filtering approaches
- **Intelligent Ranking**: Uses an AI agent for context-aware final ranking that goes beyond simple similarity scores
## Requirements
1. Python 3.8+
2. Virtual environment (recommended)
3. Mistral AI API key (get one at https://console.mistral.ai/)
Install the required packages:
```bash
pip install -r requirements.txt
```
## Environment Setup
1. Create a `.env` file in the project root:
```bash
MISTRAL_API_KEY=your_api_key_here
```
2. Ensure you have the necessary data files in the `amazon_movies_2023` directory:
- `title_embeddings.npz`: Movie title embeddings from Mistral AI
- `gcl_embeddings.npz`: Graph-enhanced embeddings
- `title_embeddings_mapping.csv`: Movie metadata mapping
## Usage
1. Activate your virtual environment:
```bash
source venv/bin/activate # On Unix/macOS
```
2. Run the recommender app:
```bash
python movie_recommender_app.py
```
3. Open your browser to the local URL shown in the terminal (typically http://127.0.0.1:7860)
## How It Works
1. **Movie Selection:**
- Search and select up to 5 movies you've enjoyed
- The system uses these as a baseline for your taste
2. **Text Preferences:**
- Describe what you're looking for (e.g., "A thrilling sci-fi movie with deep philosophical themes")
- Your description is converted to embeddings using Mistral AI
3. **Preference Weighting:**
- Use the Ξ± slider to balance between your selected movies and text description
- Ξ± = 0: Only use movie history
- Ξ± = 1: Only use text description
- Values in between combine both signals
4. **Embedding Types:**
- LLM: Pure language model embeddings for semantic understanding
- LLM + GCL: Graph-enhanced embeddings that also consider user interaction patterns
## Data Processing
For information about the dataset processing pipeline, see [DATA_PROCESSING.md](DATA_PROCESSING.md)
## Contributing
Feel free to open issues or submit pull requests with improvements! |