dotoking commited on
Commit
b2c5443
·
verified ·
1 Parent(s): 87cdc48

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +98 -90
README.md CHANGED
@@ -1,91 +1,99 @@
1
- # Cultural Exposure & Algorithmic Risk (CEAR) Baseline v1.0
2
-
3
- ## Model Description
4
-
5
- The **Cultural Exposure & Algorithmic Risk (CEAR) Model** is an **analytic, rule-based scoring system** designed to help users and researchers interpret social media usage in terms of its potential impact on cultural awareness and algorithmic vulnerability.
6
-
7
- This version is a V1 Baseline: it is **deterministic** (theory-driven by fixed rules and weights) and does not rely on supervised machine learning or proprietary user data.
8
-
9
- ### 🎯 Key Outputs
10
-
11
- 1. **Cultural Connectedness Score (C-Score):** Estimates exposure to viral and trending content, modeled with diminishing returns on time.
12
- 2. **Algorithmic Risk Score (A-Risk):** Quantifies vulnerability incurred from concentrated time on high-intensity, opaque algorithmic feeds.
13
- 3. **Platform Diversity Index (D-Index):** Measures the concentration/spread of usage across platforms (using $1/\text{HHI}$).
14
- 4. **Cultural Efficiency:** Per-platform estimates of C-Score gained per minute spent.
15
-
16
- ## ⚙️ Analytic Basis & Scoring Logic
17
-
18
- The model is defined by transparent assumptions encoded in the Python code (`cear_model.py`) and the platform weights (`platform_weights.json`).
19
-
20
- ### Core Formulas
21
-
22
- The key to the C-Score is the **Diminishing Returns Function** ($f_{DR}$), which prevents the C-Score from increasing linearly with time, acknowledging that the first hour is likely more valuable than the tenth.
23
-
24
- $$f_{DR}(\text{Min}) = \log_{10}(\text{Min} + 1)$$
25
-
26
- The final scores are calculated as:
27
-
28
- $$C_{Score} = \sum_{i} \left[ W_{C,i} \times f_{DR}(\text{Min}_i) \right]$$
29
-
30
- $$A_{Risk} = \sum_{i} \left[ W_{A,i} \times \text{Min}_i \right]$$
31
-
32
- *(Where $W_{C}$ is the Trend Density Weight and $W_{A}$ is the Algorithmic Risk Weight, defined in `platform_weights.json`.)*
33
-
34
- ## 🚀 Deployment & Usage (Hugging Face Space)
35
-
36
- This repository contains the core logic (`cear_model.py`) and the application interface (`app.py`) for a Hugging Face Space.
37
-
38
- ### Model Integration (The Engine)
39
-
40
- The core logic can be imported and run in any environment:
41
-
42
- ```python
43
- import pandas as pd
44
- from cear_model import CEARModel
45
-
46
- # Example Input Data
47
- user_data = pd.DataFrame([
48
- {'platform_name': 'TikTok', 'minutes_per_week': 450},
49
- {'platform_name': 'YouTube', 'minutes_per_week': 200},
50
- {'platform_name': 'Reddit', 'minutes_per_week': 50},
51
- ])
52
-
53
- model = CEARModel()
54
- results = model.calculate_scores(user_data)
55
- # {'C_Score': 3.75, 'A_Risk': 565.0, ...}
56
-
57
- # Application Interface (The App - app.py)
58
-
59
- The app.py script uses the Gradio library to create an interactive web interface. It handles:
60
-
61
- Collecting user input via a table component.
62
-
63
- Calling the CEARModel.calculate_scores() method.
64
-
65
- Generating a qualitative natural language summary based on the quadrant of the C-Score and A-Risk (e.g., "High C, Low A").
66
-
67
- ⚠️ Limitations and Ethical Considerations
68
-
69
- 1. Theoretical, Not Validated: The scores are based on fixed, theoretical assumptions about platform design. They are not calibrated against real-world user survey data or outcomes (e.g., actual cultural literacy, actual regret). Scores are relative estimates only.
70
-
71
- 2. No Content Analysis: The model only uses time and platform. It cannot distinguish between a productive hour watching educational content and an unproductive hour scrolling low-quality content.
72
-
73
- 3. Future Work: This deterministic model serves as a foundation. Future versions are intended to use the same input schema to train supervised machine learning models that directly predict outcomes (e.g., predicting user-reported "felt caught up" or "post-scroll regret").
74
-
75
-
76
- ---
77
-
78
- ## 2. `requirements.txt` (For Deployment)
79
-
80
- This file lists the necessary Python packages for the Gradio Space to run your model and interface correctly.
81
-
82
- ```text
83
- # requirements.txt
84
-
85
- # Core Model Dependencies
86
- pandas
87
- numpy
88
-
89
- # Gradio Space Dependencies
90
- # Gradio is used to build the simple web application interface (app.py)
 
 
 
 
 
 
 
 
91
  gradio
 
1
+ ---
2
+ title: Cultural Exposure and Algorithmic Risk Model
3
+ emoji: "🧭"
4
+ colorFrom: "blue"
5
+ colorTo: "green"
6
+ sdk: gradio # THIS IS THE CRITICAL LINE
7
+ app_file: app.py
8
+ ---
9
+ # Cultural Exposure & Algorithmic Risk (CEAR) Baseline v1.0
10
+
11
+ ## Model Description
12
+
13
+ The **Cultural Exposure & Algorithmic Risk (CEAR) Model** is an **analytic, rule-based scoring system** designed to help users and researchers interpret social media usage in terms of its potential impact on cultural awareness and algorithmic vulnerability.
14
+
15
+ This version is a V1 Baseline: it is **deterministic** (theory-driven by fixed rules and weights) and does not rely on supervised machine learning or proprietary user data.
16
+
17
+ ### 🎯 Key Outputs
18
+
19
+ 1. **Cultural Connectedness Score (C-Score):** Estimates exposure to viral and trending content, modeled with diminishing returns on time.
20
+ 2. **Algorithmic Risk Score (A-Risk):** Quantifies vulnerability incurred from concentrated time on high-intensity, opaque algorithmic feeds.
21
+ 3. **Platform Diversity Index (D-Index):** Measures the concentration/spread of usage across platforms (using $1/\text{HHI}$).
22
+ 4. **Cultural Efficiency:** Per-platform estimates of C-Score gained per minute spent.
23
+
24
+ ## ⚙️ Analytic Basis & Scoring Logic
25
+
26
+ The model is defined by transparent assumptions encoded in the Python code (`cear_model.py`) and the platform weights (`platform_weights.json`).
27
+
28
+ ### Core Formulas
29
+
30
+ The key to the C-Score is the **Diminishing Returns Function** ($f_{DR}$), which prevents the C-Score from increasing linearly with time, acknowledging that the first hour is likely more valuable than the tenth.
31
+
32
+ $$f_{DR}(\text{Min}) = \log_{10}(\text{Min} + 1)$$
33
+
34
+ The final scores are calculated as:
35
+
36
+ $$C_{Score} = \sum_{i} \left[ W_{C,i} \times f_{DR}(\text{Min}_i) \right]$$
37
+
38
+ $$A_{Risk} = \sum_{i} \left[ W_{A,i} \times \text{Min}_i \right]$$
39
+
40
+ *(Where $W_{C}$ is the Trend Density Weight and $W_{A}$ is the Algorithmic Risk Weight, defined in `platform_weights.json`.)*
41
+
42
+ ## 🚀 Deployment & Usage (Hugging Face Space)
43
+
44
+ This repository contains the core logic (`cear_model.py`) and the application interface (`app.py`) for a Hugging Face Space.
45
+
46
+ ### Model Integration (The Engine)
47
+
48
+ The core logic can be imported and run in any environment:
49
+
50
+ ```python
51
+ import pandas as pd
52
+ from cear_model import CEARModel
53
+
54
+ # Example Input Data
55
+ user_data = pd.DataFrame([
56
+ {'platform_name': 'TikTok', 'minutes_per_week': 450},
57
+ {'platform_name': 'YouTube', 'minutes_per_week': 200},
58
+ {'platform_name': 'Reddit', 'minutes_per_week': 50},
59
+ ])
60
+
61
+ model = CEARModel()
62
+ results = model.calculate_scores(user_data)
63
+ # {'C_Score': 3.75, 'A_Risk': 565.0, ...}
64
+
65
+ # Application Interface (The App - app.py)
66
+
67
+ The app.py script uses the Gradio library to create an interactive web interface. It handles:
68
+
69
+ Collecting user input via a table component.
70
+
71
+ Calling the CEARModel.calculate_scores() method.
72
+
73
+ Generating a qualitative natural language summary based on the quadrant of the C-Score and A-Risk (e.g., "High C, Low A").
74
+
75
+ ⚠️ Limitations and Ethical Considerations
76
+
77
+ 1. Theoretical, Not Validated: The scores are based on fixed, theoretical assumptions about platform design. They are not calibrated against real-world user survey data or outcomes (e.g., actual cultural literacy, actual regret). Scores are relative estimates only.
78
+
79
+ 2. No Content Analysis: The model only uses time and platform. It cannot distinguish between a productive hour watching educational content and an unproductive hour scrolling low-quality content.
80
+
81
+ 3. Future Work: This deterministic model serves as a foundation. Future versions are intended to use the same input schema to train supervised machine learning models that directly predict outcomes (e.g., predicting user-reported "felt caught up" or "post-scroll regret").
82
+
83
+
84
+ ---
85
+
86
+ ## 2. `requirements.txt` (For Deployment)
87
+
88
+ This file lists the necessary Python packages for the Gradio Space to run your model and interface correctly.
89
+
90
+ ```text
91
+ # requirements.txt
92
+
93
+ # Core Model Dependencies
94
+ pandas
95
+ numpy
96
+
97
+ # Gradio Space Dependencies
98
+ # Gradio is used to build the simple web application interface (app.py)
99
  gradio