Cultural Exposure & Algorithmic Risk (CEAR) Baseline v1.0
Model Description
The Cultural Exposure & Algorithmic Risk (CEAR) Model is an analytic, rule-based scoring system designed to help users and researchers interpret social media usage in terms of its potential impact on cultural awareness and algorithmic vulnerability.
This version is a V1 Baseline: it is deterministic (theory-driven by fixed rules and weights) and does not rely on supervised machine learning or proprietary user data.
π― Key Outputs
- Cultural Connectedness Score (C-Score): Estimates exposure to viral and trending content, modeled with diminishing returns on time.
- Algorithmic Risk Score (A-Risk): Quantifies vulnerability incurred from concentrated time on high-intensity, opaque algorithmic feeds.
- Platform Diversity Index (D-Index): Measures the concentration/spread of usage across platforms (using $1/\text{HHI}$).
- Cultural Efficiency: Per-platform estimates of C-Score gained per minute spent.
βοΈ Analytic Basis & Scoring Logic
The model is defined by transparent assumptions encoded in the Python code (cear_model.py) and the platform weights (platform_weights.json).
Core Formulas
The key to the C-Score is the Diminishing Returns Function ($f_{DR}$), which prevents the C-Score from increasing linearly with time, acknowledging that the first hour is likely more valuable than the tenth.
The final scores are calculated as:
(Where $W_{C}$ is the Trend Density Weight and $W_{A}$ is the Algorithmic Risk Weight, defined in platform_weights.json.)
π Deployment & Usage (Hugging Face Space)
This repository contains the core logic (cear_model.py) and the application interface (app.py) for a Hugging Face Space.
Model Integration (The Engine)
The core logic can be imported and run in any environment:
import pandas as pd
from cear_model import CEARModel
# Example Input Data
user_data = pd.DataFrame([
{'platform_name': 'TikTok', 'minutes_per_week': 450},
{'platform_name': 'YouTube', 'minutes_per_week': 200},
{'platform_name': 'Reddit', 'minutes_per_week': 50},
])
model = CEARModel()
results = model.calculate_scores(user_data)
# {'C_Score': 3.75, 'A_Risk': 565.0, ...}
# Application Interface (The App - app.py)
The app.py script uses the Gradio library to create an interactive web interface. It handles:
Collecting user input via a table component.
Calling the CEARModel.calculate_scores() method.
Generating a qualitative natural language summary based on the quadrant of the C-Score and A-Risk (e.g., "High C, Low A").
β οΈ Limitations and Ethical Considerations
1. Theoretical, Not Validated: The scores are based on fixed, theoretical assumptions about platform design. They are not calibrated against real-world user survey data or outcomes (e.g., actual cultural literacy, actual regret). Scores are relative estimates only.
2. No Content Analysis: The model only uses time and platform. It cannot distinguish between a productive hour watching educational content and an unproductive hour scrolling low-quality content.
3. Future Work: This deterministic model serves as a foundation. Future versions are intended to use the same input schema to train supervised machine learning models that directly predict outcomes (e.g., predicting user-reported "felt caught up" or "post-scroll regret").
---
## 2. `requirements.txt` (For Deployment)
This file lists the necessary Python packages for the Gradio Space to run your model and interface correctly.
```text
# requirements.txt
# Core Model Dependencies
pandas
numpy
# Gradio Space Dependencies
# Gradio is used to build the simple web application interface (app.py)
gradio