ChristophSchuhmann commited on
Commit
6ca832e
·
verified ·
1 Parent(s): 1352116

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -43
README.md CHANGED
@@ -182,7 +182,7 @@ if embedding_tensor is not None and emotion_mlps:
182
  else:
183
  print("Skipping inference as either embedding tensor is None or no MLP models were loaded.")
184
  ```
185
- Performance on EMoNet-FACE HQ Benchmark
186
 
187
  The Empathic-Insight-Face models demonstrate strong performance, achieving near human-expert-level agreement on the EMoNet-FACE HQ benchmark.
188
 
@@ -194,8 +194,6 @@ Human Annotators (vs. Humans) ~0.20 - 0.26*
194
  Empathic-Insight-Face LARGE ~0.18
195
  Empathic-Insight-Face SMALL ~0.14
196
  Proprietary Models (e.g., HumeFace) ~0.11
197
- VLMs (Multi-Shot Prompt) Highly Variable
198
- VLMs (Zero-Shot Prompt) Highly Variable
199
  Random Baseline ~0.00
200
 
201
  Human inter-annotator agreement (pairwise κ<sub>w</sub>) varies per annotator; this is an approximate range from Table 6 in the paper.
@@ -208,14 +206,14 @@ The performance indicates that with focused dataset construction and careful fin
208
 
209
  For more detailed benchmark results, including per-emotion performance and comparisons with other models using Spearman's Rho, please refer to the full EMoNet-FACE paper (Figures 3, 4, 9 and Table 6 in particular).
210
 
211
- Taxonomy
212
 
213
  The 40 emotion categories are:
214
  Affection, Amusement, Anger, Astonishment/Surprise, Awe, Bitterness, Concentration, Confusion, Contemplation, Contempt, Contentment, Disappointment, Disgust, Distress, Doubt, Elation, Embarrassment, Emotional Numbness, Fatigue/Exhaustion, Fear, Helplessness, Hope/Enthusiasm/Optimism, Impatience and Irritability, Infatuation, Interest, Intoxication/Altered States of Consciousness, Jealousy & Envy, Longing, Malevolence/Malice, Pain, Pleasure/Ecstasy, Pride, Relief, Sadness, Sexual Lust, Shame, Sourness, Teasing, Thankfulness/Gratitude, Triumph.
215
 
216
  (See Table 4 in the paper for associated descriptive words for each category).
217
 
218
- Limitations
219
 
220
  Synthetic Data: Models are trained on synthetic faces. Generalization to real-world, diverse, in-the-wild images is not guaranteed and requires further investigation.
221
 
@@ -225,7 +223,7 @@ Cultural Universality: The 40-category taxonomy, while expert-validated, is one
225
 
226
  Subjectivity: Emotion perception is inherently subjective.
227
 
228
- Ethical Considerations
229
 
230
  The EMoNet-FACE suite was developed with ethical considerations in mind, including:
231
 
@@ -237,7 +235,7 @@ Responsible Use: These models are released for research. Users are urged to cons
237
 
238
  Please refer to the "Ethical Considerations" and "Data Integrity, Safety, and Fairness" sections in the EMoNet-FACE paper for a comprehensive discussion.
239
 
240
- Citation
241
 
242
  If you use these models or the EMoNet-FACE benchmark in your research, please cite the original paper:
243
 
@@ -248,39 +246,3 @@ If you use these models or the EMoNet-FACE benchmark in your research, please ci
248
  year={2025} % Or actual year of publication
249
  % TODO: Add URL/DOI when available
250
  }
251
- IGNORE_WHEN_COPYING_START
252
- content_copy
253
- download
254
- Use code with caution.
255
- Bibtex
256
- IGNORE_WHEN_COPYING_END
257
-
258
- (Please update the year and add URL/DOI once the paper is officially published.)
259
-
260
- Acknowledgements
261
-
262
- We thank all the expert annotators for their invaluable contributions to the EMoNet-FACE datasets.
263
- (Add any other specific acknowledgements if desired)
264
-
265
- This README was generated based on the EMoNet-FACE paper. For full details, please refer to the publication.
266
-
267
- **Key changes made:**
268
-
269
- 1. **Author List Updated:** The author list in the introduction and the BibTeX citation has been updated to match the list you provided:
270
- `Christoph Schuhmann, Robert Kaczmarczyk, Gollam Rabby, Maurice Kraus, Felix Friedrich, Huu Nguyen, Krishna Kalyan, Kourosh Nadi, Kristian Kersting, Sören Auer.`
271
- 2. **`MODEL_DIRECTORY` in Code Example:**
272
- * A new variable `MODEL_DIRECTORY = Path("./Empathic-Insight-Face-Large")` is introduced.
273
- * **Crucially, users are instructed to `ADJUST THIS PATH`** to where they have actually downloaded/cloned the Hugging Face repository containing the `.pth` files and the `neutral_stats_cache...` file.
274
- * The code now uses this `MODEL_DIRECTORY` to load the neutral stats and iterate through the `.pth` files.
275
- 3. **Inference for all 40 Experts (Models) in Code Example:**
276
- * The code snippet now iterates through all `model_*_best.pth` files found in the `MODEL_DIRECTORY`.
277
- * It loads each MLP model, performs inference, applies mean subtraction using the corresponding neutral mean, and stores/prints the results for all detected emotion models.
278
- * Added more robust error handling for file loading.
279
- * Includes a placeholder for actual image processing, defaulting to a random embedding if an image path is not correctly set up by the user, to ensure the rest of the script can still demonstrate the MLP loading and inference loop.
280
-
281
- This revised README should be more accurate and provide a more complete and usable code example for users.
282
- IGNORE_WHEN_COPYING_START
283
- content_copy
284
- download
285
- Use code with caution.
286
- IGNORE_WHEN_COPYING_END
 
182
  else:
183
  print("Skipping inference as either embedding tensor is None or no MLP models were loaded.")
184
  ```
185
+ ## Performance on EMoNet-FACE HQ Benchmark
186
 
187
  The Empathic-Insight-Face models demonstrate strong performance, achieving near human-expert-level agreement on the EMoNet-FACE HQ benchmark.
188
 
 
194
  Empathic-Insight-Face LARGE ~0.18
195
  Empathic-Insight-Face SMALL ~0.14
196
  Proprietary Models (e.g., HumeFace) ~0.11
 
 
197
  Random Baseline ~0.00
198
 
199
  Human inter-annotator agreement (pairwise κ<sub>w</sub>) varies per annotator; this is an approximate range from Table 6 in the paper.
 
206
 
207
  For more detailed benchmark results, including per-emotion performance and comparisons with other models using Spearman's Rho, please refer to the full EMoNet-FACE paper (Figures 3, 4, 9 and Table 6 in particular).
208
 
209
+ ## Taxonomy
210
 
211
  The 40 emotion categories are:
212
  Affection, Amusement, Anger, Astonishment/Surprise, Awe, Bitterness, Concentration, Confusion, Contemplation, Contempt, Contentment, Disappointment, Disgust, Distress, Doubt, Elation, Embarrassment, Emotional Numbness, Fatigue/Exhaustion, Fear, Helplessness, Hope/Enthusiasm/Optimism, Impatience and Irritability, Infatuation, Interest, Intoxication/Altered States of Consciousness, Jealousy & Envy, Longing, Malevolence/Malice, Pain, Pleasure/Ecstasy, Pride, Relief, Sadness, Sexual Lust, Shame, Sourness, Teasing, Thankfulness/Gratitude, Triumph.
213
 
214
  (See Table 4 in the paper for associated descriptive words for each category).
215
 
216
+ ## Limitations
217
 
218
  Synthetic Data: Models are trained on synthetic faces. Generalization to real-world, diverse, in-the-wild images is not guaranteed and requires further investigation.
219
 
 
223
 
224
  Subjectivity: Emotion perception is inherently subjective.
225
 
226
+ ## Ethical Considerations
227
 
228
  The EMoNet-FACE suite was developed with ethical considerations in mind, including:
229
 
 
235
 
236
  Please refer to the "Ethical Considerations" and "Data Integrity, Safety, and Fairness" sections in the EMoNet-FACE paper for a comprehensive discussion.
237
 
238
+ ## Citation
239
 
240
  If you use these models or the EMoNet-FACE benchmark in your research, please cite the original paper:
241
 
 
246
  year={2025} % Or actual year of publication
247
  % TODO: Add URL/DOI when available
248
  }