Update README.md
Browse files
README.md
CHANGED
|
@@ -182,7 +182,7 @@ if embedding_tensor is not None and emotion_mlps:
|
|
| 182 |
else:
|
| 183 |
print("Skipping inference as either embedding tensor is None or no MLP models were loaded.")
|
| 184 |
```
|
| 185 |
-
Performance on EMoNet-FACE HQ Benchmark
|
| 186 |
|
| 187 |
The Empathic-Insight-Face models demonstrate strong performance, achieving near human-expert-level agreement on the EMoNet-FACE HQ benchmark.
|
| 188 |
|
|
@@ -194,8 +194,6 @@ Human Annotators (vs. Humans) ~0.20 - 0.26*
|
|
| 194 |
Empathic-Insight-Face LARGE ~0.18
|
| 195 |
Empathic-Insight-Face SMALL ~0.14
|
| 196 |
Proprietary Models (e.g., HumeFace) ~0.11
|
| 197 |
-
VLMs (Multi-Shot Prompt) Highly Variable
|
| 198 |
-
VLMs (Zero-Shot Prompt) Highly Variable
|
| 199 |
Random Baseline ~0.00
|
| 200 |
|
| 201 |
Human inter-annotator agreement (pairwise κ<sub>w</sub>) varies per annotator; this is an approximate range from Table 6 in the paper.
|
|
@@ -208,14 +206,14 @@ The performance indicates that with focused dataset construction and careful fin
|
|
| 208 |
|
| 209 |
For more detailed benchmark results, including per-emotion performance and comparisons with other models using Spearman's Rho, please refer to the full EMoNet-FACE paper (Figures 3, 4, 9 and Table 6 in particular).
|
| 210 |
|
| 211 |
-
Taxonomy
|
| 212 |
|
| 213 |
The 40 emotion categories are:
|
| 214 |
Affection, Amusement, Anger, Astonishment/Surprise, Awe, Bitterness, Concentration, Confusion, Contemplation, Contempt, Contentment, Disappointment, Disgust, Distress, Doubt, Elation, Embarrassment, Emotional Numbness, Fatigue/Exhaustion, Fear, Helplessness, Hope/Enthusiasm/Optimism, Impatience and Irritability, Infatuation, Interest, Intoxication/Altered States of Consciousness, Jealousy & Envy, Longing, Malevolence/Malice, Pain, Pleasure/Ecstasy, Pride, Relief, Sadness, Sexual Lust, Shame, Sourness, Teasing, Thankfulness/Gratitude, Triumph.
|
| 215 |
|
| 216 |
(See Table 4 in the paper for associated descriptive words for each category).
|
| 217 |
|
| 218 |
-
Limitations
|
| 219 |
|
| 220 |
Synthetic Data: Models are trained on synthetic faces. Generalization to real-world, diverse, in-the-wild images is not guaranteed and requires further investigation.
|
| 221 |
|
|
@@ -225,7 +223,7 @@ Cultural Universality: The 40-category taxonomy, while expert-validated, is one
|
|
| 225 |
|
| 226 |
Subjectivity: Emotion perception is inherently subjective.
|
| 227 |
|
| 228 |
-
Ethical Considerations
|
| 229 |
|
| 230 |
The EMoNet-FACE suite was developed with ethical considerations in mind, including:
|
| 231 |
|
|
@@ -237,7 +235,7 @@ Responsible Use: These models are released for research. Users are urged to cons
|
|
| 237 |
|
| 238 |
Please refer to the "Ethical Considerations" and "Data Integrity, Safety, and Fairness" sections in the EMoNet-FACE paper for a comprehensive discussion.
|
| 239 |
|
| 240 |
-
Citation
|
| 241 |
|
| 242 |
If you use these models or the EMoNet-FACE benchmark in your research, please cite the original paper:
|
| 243 |
|
|
@@ -248,39 +246,3 @@ If you use these models or the EMoNet-FACE benchmark in your research, please ci
|
|
| 248 |
year={2025} % Or actual year of publication
|
| 249 |
% TODO: Add URL/DOI when available
|
| 250 |
}
|
| 251 |
-
IGNORE_WHEN_COPYING_START
|
| 252 |
-
content_copy
|
| 253 |
-
download
|
| 254 |
-
Use code with caution.
|
| 255 |
-
Bibtex
|
| 256 |
-
IGNORE_WHEN_COPYING_END
|
| 257 |
-
|
| 258 |
-
(Please update the year and add URL/DOI once the paper is officially published.)
|
| 259 |
-
|
| 260 |
-
Acknowledgements
|
| 261 |
-
|
| 262 |
-
We thank all the expert annotators for their invaluable contributions to the EMoNet-FACE datasets.
|
| 263 |
-
(Add any other specific acknowledgements if desired)
|
| 264 |
-
|
| 265 |
-
This README was generated based on the EMoNet-FACE paper. For full details, please refer to the publication.
|
| 266 |
-
|
| 267 |
-
**Key changes made:**
|
| 268 |
-
|
| 269 |
-
1. **Author List Updated:** The author list in the introduction and the BibTeX citation has been updated to match the list you provided:
|
| 270 |
-
`Christoph Schuhmann, Robert Kaczmarczyk, Gollam Rabby, Maurice Kraus, Felix Friedrich, Huu Nguyen, Krishna Kalyan, Kourosh Nadi, Kristian Kersting, Sören Auer.`
|
| 271 |
-
2. **`MODEL_DIRECTORY` in Code Example:**
|
| 272 |
-
* A new variable `MODEL_DIRECTORY = Path("./Empathic-Insight-Face-Large")` is introduced.
|
| 273 |
-
* **Crucially, users are instructed to `ADJUST THIS PATH`** to where they have actually downloaded/cloned the Hugging Face repository containing the `.pth` files and the `neutral_stats_cache...` file.
|
| 274 |
-
* The code now uses this `MODEL_DIRECTORY` to load the neutral stats and iterate through the `.pth` files.
|
| 275 |
-
3. **Inference for all 40 Experts (Models) in Code Example:**
|
| 276 |
-
* The code snippet now iterates through all `model_*_best.pth` files found in the `MODEL_DIRECTORY`.
|
| 277 |
-
* It loads each MLP model, performs inference, applies mean subtraction using the corresponding neutral mean, and stores/prints the results for all detected emotion models.
|
| 278 |
-
* Added more robust error handling for file loading.
|
| 279 |
-
* Includes a placeholder for actual image processing, defaulting to a random embedding if an image path is not correctly set up by the user, to ensure the rest of the script can still demonstrate the MLP loading and inference loop.
|
| 280 |
-
|
| 281 |
-
This revised README should be more accurate and provide a more complete and usable code example for users.
|
| 282 |
-
IGNORE_WHEN_COPYING_START
|
| 283 |
-
content_copy
|
| 284 |
-
download
|
| 285 |
-
Use code with caution.
|
| 286 |
-
IGNORE_WHEN_COPYING_END
|
|
|
|
| 182 |
else:
|
| 183 |
print("Skipping inference as either embedding tensor is None or no MLP models were loaded.")
|
| 184 |
```
|
| 185 |
+
## Performance on EMoNet-FACE HQ Benchmark
|
| 186 |
|
| 187 |
The Empathic-Insight-Face models demonstrate strong performance, achieving near human-expert-level agreement on the EMoNet-FACE HQ benchmark.
|
| 188 |
|
|
|
|
| 194 |
Empathic-Insight-Face LARGE ~0.18
|
| 195 |
Empathic-Insight-Face SMALL ~0.14
|
| 196 |
Proprietary Models (e.g., HumeFace) ~0.11
|
|
|
|
|
|
|
| 197 |
Random Baseline ~0.00
|
| 198 |
|
| 199 |
Human inter-annotator agreement (pairwise κ<sub>w</sub>) varies per annotator; this is an approximate range from Table 6 in the paper.
|
|
|
|
| 206 |
|
| 207 |
For more detailed benchmark results, including per-emotion performance and comparisons with other models using Spearman's Rho, please refer to the full EMoNet-FACE paper (Figures 3, 4, 9 and Table 6 in particular).
|
| 208 |
|
| 209 |
+
## Taxonomy
|
| 210 |
|
| 211 |
The 40 emotion categories are:
|
| 212 |
Affection, Amusement, Anger, Astonishment/Surprise, Awe, Bitterness, Concentration, Confusion, Contemplation, Contempt, Contentment, Disappointment, Disgust, Distress, Doubt, Elation, Embarrassment, Emotional Numbness, Fatigue/Exhaustion, Fear, Helplessness, Hope/Enthusiasm/Optimism, Impatience and Irritability, Infatuation, Interest, Intoxication/Altered States of Consciousness, Jealousy & Envy, Longing, Malevolence/Malice, Pain, Pleasure/Ecstasy, Pride, Relief, Sadness, Sexual Lust, Shame, Sourness, Teasing, Thankfulness/Gratitude, Triumph.
|
| 213 |
|
| 214 |
(See Table 4 in the paper for associated descriptive words for each category).
|
| 215 |
|
| 216 |
+
## Limitations
|
| 217 |
|
| 218 |
Synthetic Data: Models are trained on synthetic faces. Generalization to real-world, diverse, in-the-wild images is not guaranteed and requires further investigation.
|
| 219 |
|
|
|
|
| 223 |
|
| 224 |
Subjectivity: Emotion perception is inherently subjective.
|
| 225 |
|
| 226 |
+
## Ethical Considerations
|
| 227 |
|
| 228 |
The EMoNet-FACE suite was developed with ethical considerations in mind, including:
|
| 229 |
|
|
|
|
| 235 |
|
| 236 |
Please refer to the "Ethical Considerations" and "Data Integrity, Safety, and Fairness" sections in the EMoNet-FACE paper for a comprehensive discussion.
|
| 237 |
|
| 238 |
+
## Citation
|
| 239 |
|
| 240 |
If you use these models or the EMoNet-FACE benchmark in your research, please cite the original paper:
|
| 241 |
|
|
|
|
| 246 |
year={2025} % Or actual year of publication
|
| 247 |
% TODO: Add URL/DOI when available
|
| 248 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|