dotoking commited on
Commit
c8a649e
Β·
verified Β·
1 Parent(s): ad95a61

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +83 -99
README.md CHANGED
@@ -1,99 +1,83 @@
1
- ---
2
- title: Cultural Exposure and Algorithmic Risk Model
3
- emoji: "🧭"
4
- colorFrom: "blue"
5
- colorTo: "green"
6
- sdk: gradio # THIS IS THE CRITICAL LINE
7
- app_file: app.py
8
- ---
9
- # Cultural Exposure & Algorithmic Risk (CEAR) Baseline v1.0
10
-
11
- ## Model Description
12
-
13
- The **Cultural Exposure & Algorithmic Risk (CEAR) Model** is an **analytic, rule-based scoring system** designed to help users and researchers interpret social media usage in terms of its potential impact on cultural awareness and algorithmic vulnerability.
14
-
15
- This version is a V1 Baseline: it is **deterministic** (theory-driven by fixed rules and weights) and does not rely on supervised machine learning or proprietary user data.
16
-
17
- ### 🎯 Key Outputs
18
-
19
- 1. **Cultural Connectedness Score (C-Score):** Estimates exposure to viral and trending content, modeled with diminishing returns on time.
20
- 2. **Algorithmic Risk Score (A-Risk):** Quantifies vulnerability incurred from concentrated time on high-intensity, opaque algorithmic feeds.
21
- 3. **Platform Diversity Index (D-Index):** Measures the concentration/spread of usage across platforms (using $1/\text{HHI}$).
22
- 4. **Cultural Efficiency:** Per-platform estimates of C-Score gained per minute spent.
23
-
24
- ## βš™οΈ Analytic Basis & Scoring Logic
25
-
26
- The model is defined by transparent assumptions encoded in the Python code (`cear_model.py`) and the platform weights (`platform_weights.json`).
27
-
28
- ### Core Formulas
29
-
30
- The key to the C-Score is the **Diminishing Returns Function** ($f_{DR}$), which prevents the C-Score from increasing linearly with time, acknowledging that the first hour is likely more valuable than the tenth.
31
-
32
- $$f_{DR}(\text{Min}) = \log_{10}(\text{Min} + 1)$$
33
-
34
- The final scores are calculated as:
35
-
36
- $$C_{Score} = \sum_{i} \left[ W_{C,i} \times f_{DR}(\text{Min}_i) \right]$$
37
-
38
- $$A_{Risk} = \sum_{i} \left[ W_{A,i} \times \text{Min}_i \right]$$
39
-
40
- *(Where $W_{C}$ is the Trend Density Weight and $W_{A}$ is the Algorithmic Risk Weight, defined in `platform_weights.json`.)*
41
-
42
- ## πŸš€ Deployment & Usage (Hugging Face Space)
43
-
44
- This repository contains the core logic (`cear_model.py`) and the application interface (`app.py`) for a Hugging Face Space.
45
-
46
- ### Model Integration (The Engine)
47
-
48
- The core logic can be imported and run in any environment:
49
-
50
- ```python
51
- import pandas as pd
52
- from cear_model import CEARModel
53
-
54
- # Example Input Data
55
- user_data = pd.DataFrame([
56
- {'platform_name': 'TikTok', 'minutes_per_week': 450},
57
- {'platform_name': 'YouTube', 'minutes_per_week': 200},
58
- {'platform_name': 'Reddit', 'minutes_per_week': 50},
59
- ])
60
-
61
- model = CEARModel()
62
- results = model.calculate_scores(user_data)
63
- # {'C_Score': 3.75, 'A_Risk': 565.0, ...}
64
-
65
- # Application Interface (The App - app.py)
66
-
67
- The app.py script uses the Gradio library to create an interactive web interface. It handles:
68
-
69
- Collecting user input via a table component.
70
-
71
- Calling the CEARModel.calculate_scores() method.
72
-
73
- Generating a qualitative natural language summary based on the quadrant of the C-Score and A-Risk (e.g., "High C, Low A").
74
-
75
- ⚠️ Limitations and Ethical Considerations
76
-
77
- 1. Theoretical, Not Validated: The scores are based on fixed, theoretical assumptions about platform design. They are not calibrated against real-world user survey data or outcomes (e.g., actual cultural literacy, actual regret). Scores are relative estimates only.
78
-
79
- 2. No Content Analysis: The model only uses time and platform. It cannot distinguish between a productive hour watching educational content and an unproductive hour scrolling low-quality content.
80
-
81
- 3. Future Work: This deterministic model serves as a foundation. Future versions are intended to use the same input schema to train supervised machine learning models that directly predict outcomes (e.g., predicting user-reported "felt caught up" or "post-scroll regret").
82
-
83
-
84
- ---
85
-
86
- ## 2. `requirements.txt` (For Deployment)
87
-
88
- This file lists the necessary Python packages for the Gradio Space to run your model and interface correctly.
89
-
90
- ```text
91
- # requirements.txt
92
-
93
- # Core Model Dependencies
94
- pandas
95
- numpy
96
-
97
- # Gradio Space Dependencies
98
- # Gradio is used to build the simple web application interface (app.py)
99
- gradio
 
1
+ ---
2
+ title: Cultural Exposure Risk Model
3
+ emoji: 🧭
4
+ colorFrom: blue
5
+ colorTo: green
6
+ sdk: gradio
7
+ sdk_version: 6.0.2
8
+ app_file: app.py
9
+ pinned: false
10
+ ---
11
+ # Cultural Exposure & Algorithmic Risk (CEAR) Baseline v1.0
12
+
13
+ ## Model Description
14
+
15
+ The **Cultural Exposure & Algorithmic Risk (CEAR) Model** is an **analytic, rule-based scoring system** designed to help users and researchers interpret social media usage in terms of its potential impact on cultural awareness and algorithmic vulnerability.
16
+
17
+ This version is a V1 Baseline: it is **deterministic** (theory-driven by fixed rules and weights) and does not rely on supervised machine learning or proprietary user data.
18
+
19
+ ### 🎯 Key Outputs
20
+
21
+ 1. **Cultural Connectedness Score (C-Score):** Estimates exposure to viral and trending content, modeled with diminishing returns on time.
22
+ 2. **Algorithmic Risk Score (A-Risk):** Quantifies vulnerability incurred from concentrated time on high-intensity, opaque algorithmic feeds.
23
+ 3. **Platform Diversity Index (D-Index):** Measures the concentration/spread of usage across platforms (using $1/\text{HHI}$).
24
+ 4. **Cultural Efficiency:** Per-platform estimates of C-Score gained per minute spent.
25
+
26
+ ## βš™οΈ Analytic Basis & Scoring Logic
27
+
28
+ The model is defined by transparent assumptions encoded in the Python code (`cear_model.py`) and the platform weights (`platform_weights.json`).
29
+
30
+ ### Core Formulas
31
+
32
+ The key to the C-Score is the **Diminishing Returns Function** ($f_{DR}$), which prevents the C-Score from increasing linearly with time, acknowledging that the first hour is likely more valuable than the tenth.
33
+
34
+ $$f_{DR}(\text{Min}) = \log_{10}(\text{Min} + 1)$$
35
+
36
+ The final scores are calculated as:
37
+
38
+ $$C_{Score} = \sum_{i} \left[ W_{C,i} \times f_{DR}(\text{Min}_i) \right]$$
39
+
40
+ $$A_{Risk} = \sum_{i} \left[ W_{A,i} \times \text{Min}_i \right]$$
41
+
42
+ *(Where $W_{C}$ is the Trend Density Weight and $W_{A}$ is the Algorithmic Risk Weight, defined in `platform_weights.json`.)*
43
+
44
+ ## πŸš€ Deployment & Usage (Hugging Face Space)
45
+
46
+ This repository contains the core logic (`cear_model.py`) and the application interface (`app.py`) for a Hugging Face Space.
47
+
48
+ ### Model Integration (The Engine)
49
+
50
+ The core logic can be imported and run in any environment:
51
+
52
+ ```python
53
+ import pandas as pd
54
+ from cear_model import CEARModel
55
+
56
+ # Example Input Data
57
+ user_data = pd.DataFrame([
58
+ {'platform_name': 'TikTok', 'minutes_per_week': 450},
59
+ {'platform_name': 'YouTube', 'minutes_per_week': 200},
60
+ {'platform_name': 'Reddit', 'minutes_per_week': 50},
61
+ ])
62
+
63
+ model = CEARModel()
64
+ results = model.calculate_scores(user_data)
65
+ # {'C_Score': 3.75, 'A_Risk': 565.0, ...}
66
+
67
+ # Application Interface (The App - app.py)
68
+
69
+ The app.py script uses the Gradio library to create an interactive web interface. It handles:
70
+
71
+ Collecting user input via a table component.
72
+
73
+ Calling the CEARModel.calculate_scores() method.
74
+
75
+ Generating a qualitative natural language summary based on the quadrant of the C-Score and A-Risk (e.g., "High C, Low A").
76
+
77
+ ⚠️ Limitations and Ethical Considerations
78
+
79
+ 1. Theoretical, Not Validated: The scores are based on fixed, theoretical assumptions about platform design. They are not calibrated against real-world user survey data or outcomes (e.g., actual cultural literacy, actual regret). Scores are relative estimates only.
80
+
81
+ 2. No Content Analysis: The model only uses time and platform. It cannot distinguish between a productive hour watching educational content and an unproductive hour scrolling low-quality content.
82
+
83
+ 3. Future Work: This deterministic model serves as a foundation. Future versions are intended to use the same input schema to train supervised machine learning models that directly predict outcomes (e.g., predicting user-reported "felt caught up" or "post-scroll regret").