File size: 1,290 Bytes
011c907
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
---
license: cc-by-4.0
tags:
  - benchmarking
  - llm
  - model-evaluation
  - vision
  - ai
pretty_name: https://OpenMark.ai AI Model Emotion Detection Benchmark
---

# AI Model Emotion Detection Benchmark

Benchmark results from testing 11 AI models on emotion detection from movie stills, conducted on [OpenMark](https://openmark.ai) — a deterministic AI model benchmarking platform.

## Methodology

- **Task:** Identify emotions from 4 movie stills (varying complexity)
- **Models tested:** 11 (GPT-5.2, Gemini 3 Pro, Gemini 3.1 Pro, Claude Opus 4.6, Claude Sonnet 4.6, Grok 4.1 Fast, Llama 4 Maverick, Qwen 3.5, Sonar, Gemini 3 Flash, Mistral Medium)
- **Runs per model:** 3 (for stability measurement)
- **Scoring:** Deterministic, task-specific evaluation
- **Costs:** Real API costs tracked per task

## Key Findings

- GPT-5.2 and Gemini 3 Pro tied at 75% accuracy
- Claude Opus 4.6 ($0.025/task) scored identically to Llama 4 Maverick ($0.002/task) — 12x price difference
- Half the models showed ±1.000 stability variance (changed answers across runs)

## Source

Data generated using [OpenMark](https://openmark.ai). Full analysis: [I benchmarked 10 ai models on reading human emotions](https://dev.to/openmarkai/i-benchmarked-10-ai-models-on-reading-human-emotions-3m0b)