tonalityprint-v1 / CODEBOOK.md
tonalityprint's picture
Upload 12 files
b5651d4 verified
# CODEBOOK - TonalityPrint Voice Dataset v1.0
## Overview
This codebook provides definitions for variables, file naming conventions, and data structures in the TonalityPrint Voice Dataset v1.0.
**Dataset Information**:
- **Total Files**: 144 audio files + 144 JSON + 144 CSV + 1 combined CSV
- **DOI**: https://doi.org/10.5281/zenodo.17913895
- **License**: CC BY-NC 4.0
- **Contact**: ronda@TonalityPrint.com
**Quick Navigation**:
- [File Naming Convention](#file-naming-convention)
- [CSV Variables](#csv-variables-23-columns)
- [Tonality Indices](#tonality-indices-0-100-scale)
- [Intention Categories](#intention-categories)
- [Modifier Codes](#modifier-codes-24-optional-sub-modifiers)
- [Segment-Level Data](#segment-level-data-structure)
---
## File Naming Convention
### Audio Files (.wav)
**Structure**:
```
[Version]_[Batch]_[Utterance]_[Type]_[Intention]_[Modifier]_[Ambivalence]_[Speaker].wav
```
**Examples**:
1. **Single** (Primary Intent only):
`TPV1_B1_UTT1_S_Att_SP-Ronda.wav`
2. **Compound** (Primary Intent + Sub-modifier):
`TPV1_B1_UTT1_S_Reci_affi_SP-Ronda.wav`
3. **Complex** (Primary Intent + Sub-modifier + Ambivalence):
`TPV1_B1_UTT1_S_Reci_affi_ambivalex_SP-Ronda.wav`
### Component Definitions
| Component | Description | Valid Values | Example |
|-----------|-------------|--------------|---------|
| **Version** | Dataset version | `TPV1` | TPV1 |
| **Batch** | Batch number (1-6) | `B1`, `B2`, `B3`, `B4`, `B5`, `B6` | B1 |
| **Utterance** | Utterance ID (1-18) | `UTT1` through `UTT18` | UTT1 |
| **Type** | Statement/Question | `S` (Statement), `Q` (Question) | S |
| **Intention** | Primary tonal intent | `Att`, `Trus`, `Reci`, `Emre`, `Cogen`, `Baseneutral` | Att |
| **Modifier** | Optional sub-modifier | See [Modifier Codes](#modifier-codes-24-optional-sub-modifiers) | affi, calm |
| **Ambivalence** | Ambivalence marker | `ambivalex` (or omitted) | ambivalex |
| **Speaker** | Speaker identifier | `SP-Ronda` | SP-Ronda |
---
## CSV Variables (23 Columns)
### Complete Variable List
The combined CSV file (`ALL_TONALITY_DATA_COMBINED.csv`) and individual CSV files contain these 23 variables:
| # | Variable Name | Type | Description |
|---|--------------|------|-------------|
| 1 | `Version` | String | Dataset version identifier |
| 2 | `Batch_Number` | String | Batch identifier (B1-B6) |
| 3 | `Utterance_Number` | String | Utterance identifier (UTT1-UTT18) |
| 4 | `Utterance_Type` | String | S (Statement) or Q (Question) |
| 5 | `File_Name` | String | Complete audio filename |
| 6 | `Primary_Intention` | String | Primary tonal intent category |
| 7 | `Sub_Modifier` | String | Optional sub-modifier (or empty) |
| 8 | `Ambivalex` | String | Ambivalence marker (or empty) |
| 9 | `Speaker` | String | Speaker name |
| 10 | `Utterance_Text` | String | Transcribed utterance text |
| 11 | `Trust_Index` | Integer | Trust tonality score (0-100) |
| 12 | `Reciprocity_Index` | Integer | Reciprocity score (0-100) |
| 13 | `Empathy_Resonance_Index` | Integer | Empathy resonance score (0-100) |
| 14 | `Cognitive_Energy_Index` | Integer | Cognitive energy score (0-100) |
| 15 | `Attention_Index` | Integer | Attention score (0-100) |
| 16 | `Notes` | String | Annotation notes and observations |
| 17 | `Duration` | Time | Utterance duration (MM:SS format) |
| 18 | `Date_Recorded` | Date | Recording date (YYYY-MM-DD) |
| 19 | `Source` | String | Data source description |
| 20 | `Segments` | JSON String | Time-aligned segment data |
| 21 | `Start_Time` | Time | Utterance start time (MM:SS) |
| 22 | `End_Time` | Time | Utterance end time (MM:SS) |
| 23 | `Timestamp` | DateTime | ISO 8601 timestamp |
---
## Variable Definitions (Detailed)
### Metadata Variables
#### 1. Version
- **Type**: String
- **Description**: Dataset version identifier
- **Values**: `"TPV1"` (TonalityPrint Version 1)
- **Example**: `TPV1`
#### 2. Batch_Number
- **Type**: String
- **Description**: Recording batch identifier
- **Values**: `B1`, `B2`, `B3`, `B4`, `B5`, `B6`
- **Total Batches**: 6
- **Utterances per Batch**: 18
- **Example**: `B1`
#### 3. Utterance_Number
- **Type**: String
- **Description**: Unique utterance identifier within each batch
- **Values**: `UTT1`, `UTT2`, ..., `UTT18`
- **Example**: `UTT1`
#### 4. Utterance_Type
- **Type**: String (Categorical)
- **Description**: Syntactic type of the utterance
- **Values**:
- `S` = Statement (declarative sentence)
- `Q` = Question (interrogative sentence)
- **Distribution**: ~83% Statements, ~17% Questions
- **Example**: `S`
#### 5. File_Name
- **Type**: String
- **Description**: Complete audio filename with extension
- **Format**: `TPV1_[Batch]_[Utterance]_[Type]_[Intention]_[Modifier]_[Ambivalence]_SP-Ronda.wav`
- **Example**: `TPV1_B1_UTT1_S_Att_SP-Ronda.wav`
#### 6. Primary_Intention
- **Type**: String (Categorical)
- **Description**: Primary functional tonal intent category
- **Values**:
- `Attention` (directing focus and engagement)
- `Trust` (conveying reliability and credibility)
- `Reciprocity` (expressing mutual exchange)
- `Empathy Resonance` (demonstrating empathetic connection)
- `Cognitive Energy` (showing mental engagement)
- `Baseline Neutral` (neutral control sample)
- **Note**: Full word used in CSV (e.g., "Attention"), abbreviated in filename (e.g., "Att")
- **Example**: `Attention`
#### 7. Sub_Modifier
- **Type**: String (Optional)
- **Description**: Optional sub-modifier providing nuanced tonality descriptor
- **Values**: See [Modifier Codes](#modifier-codes-24-optional-sub-modifiers) table
- **Missing Data**: Empty string if not applicable
- **Example**: `affi` (Affirming), empty string `""`
#### 8. Ambivalex
- **Type**: String (Optional)
- **Description**: Cross-modifier Ambivalence marker indicating mixed or transitional tonality
- **Values**:
- `ambivalex` = Ambivalence present
- Empty string = No ambivalence
- **Definition**: Two or more contradictory/competing sub-modifier layers present simultaneously
- **Example**: `ambivalex`, empty string `""`
#### 9. Speaker
- **Type**: String
- **Description**: Speaker identifier
- **Values**: `Ronda`
- **Note**: Single-speaker dataset (all 144 files same speaker)
- **Example**: `Ronda`
#### 10. Utterance_Text
- **Type**: String
- **Description**: Verbatim transcription of spoken utterance
- **Encoding**: UTF-8
- **Max Length**: ~200 characters
- **Example**: `"I want to make sure I understand what you need"`
---
### Tonality Indices (0-100 Scale)
All five tonality indices are measured on a continuous 0-100 scale where higher values indicate stronger presence of the measured tonal quality.
#### 11. Trust_Index
- **Type**: Integer
- **Range**: 0-100
- **Description**: Quantified measure of trust tonality (perceived safety, authenticity, credibility)
- **Interpretation**:
- **Low (0-33)**: Uncertain, hesitant tonality
- **Moderate (34-66)**: Moderately reliable tonality
- **High (67-100)**: Highly trustworthy tonality
- **Example**: `75`
#### 12. Reciprocity_Index
- **Type**: Integer
- **Range**: 0-100
- **Description**: Quantified measure of reciprocal/collaborative tonality (inviting response, conversational balance)
- **Interpretation**:
- **Low (0-33)**: Unilateral communication
- **Moderate (34-66)**: Somewhat collaborative
- **High (67-100)**: Highly collaborative, balanced
- **Example**: `93`
#### 13. Empathy_Resonance_Index
- **Type**: Integer
- **Range**: 0-100
- **Description**: Quantified measure of empathetic tonality (emotional attunement, mirroring listener state)
- **Interpretation**:
- **Low (0-33)**: Detached, impersonal
- **Moderate (34-66)**: Moderately attuned
- **High (67-100)**: Highly empathetic, warm
- **Example**: `76`
#### 14. Cognitive_Energy_Index
- **Type**: Integer
- **Range**: 0-100
- **Description**: Quantified measure of cognitive engagement and mental energy (activation, momentum, pacing)
- **Interpretation**:
- **Low (0-33)**: Low engagement, slow pacing
- **Moderate (34-66)**: Moderate engagement
- **High (67-100)**: High mental energy, dynamic
- **Known Issue**: Shows systematic elevation across corpus (see Notes)
- **Example**: `96`
#### 15. Attention_Index
- **Type**: Integer
- **Range**: 0-100
- **Description**: Quantified measure of attentional focus (directing perceptual priority, maintaining engagement)
- **Interpretation**:
- **Low (0-33)**: Unfocused, diffuse attention
- **Moderate (34-66)**: Moderately engaged
- **High (67-100)**: Highly focused, commanding attention
- **Example**: `80`
**Scoring Methodology**: All indices were scored by expert practitioner trained in "Tonality as Attention" framework based on perceptual assessment and acoustic analysis.
---
### Additional Variables
#### 16. Notes
- **Type**: String (Free text)
- **Description**: Annotation notes, quality observations, and systematic bias documentation
- **Common Note**: "Cognitive Energy (CE) seemingly exhibits systemic leaks/dominance, possibly due to speaker ecological style, lexical content and /or practitioner bias. Intentionally retained for transparency."
- **Missing Data**: Empty string if no notes
- **Example**: `"Cognitive Energy (CE) seemingly exhibits systemic leaks/dominance..."`
#### 17. Duration
- **Type**: Time (MM:SS format)
- **Description**: Total duration of audio utterance
- **Format**: `M:SS` or `MM:SS`
- **Range**: ~3-6 seconds per utterance
- **Total Duration**: ~10 minutes (all 144 files)
- **Example**: `0:04` (4 seconds)
#### 18. Date_Recorded
- **Type**: Date (YYYY-MM-DD)
- **Description**: Date the audio was recorded
- **Date Range**: December 19, 2025 - January 23, 2026
- **Example**: `2026-01-20`
#### 19. Source
- **Type**: String
- **Description**: Data source and annotation method
- **Values**: `"Recording - Expert Practitioner Annotator"`
- **Note**: All annotations performed by single expert practitioner
- **Example**: `Recording - Expert Practitioner Annotator`
#### 20. Segments
- **Type**: JSON Array (stored as string in CSV)
- **Description**: Time-aligned segment-level tonality data with millisecond precision
- **Structure**: Array of objects with `startTime`, `endTime`, and five tonality indices
- **See**: [Segment-Level Data Structure](#segment-level-data-structure) section
- **Example**: `[{"startTime":0,"endTime":4284.083333333333,"trust":75,"reciprocity":93,"empathy":76,"cognitive":96,"attention":80}]`
#### 21. Start_Time
- **Type**: Time (MM:SS format)
- **Description**: Utterance start time (typically 0:00)
- **Example**: `0:00`
#### 22. End_Time
- **Type**: Time (MM:SS format)
- **Description**: Utterance end time (matches Duration)
- **Example**: `0:04`
#### 23. Timestamp
- **Type**: DateTime (ISO 8601 format)
- **Description**: Precise timestamp of annotation creation
- **Format**: `YYYY-MM-DDTHH:MM:SS.sssZ`
- **Timezone**: UTC (Z suffix)
- **Example**: `2026-01-20T16:45:24.342Z`
---
## Intention Categories
### Primary Functional Tonal Intent States (6 Categories)
| Category | Code (Filename) | Full Name (CSV) | Description |
|----------|----------------|-----------------|-------------|
| **Attention** | `Att` | `Attention` | Directing focus, capturing and maintaining listener engagement |
| **Trust** | `Trus` | `Trust` | Conveying trustworthiness, reliability, credibility, and authenticity |
| **Reciprocity** | `Reci` | `Reciprocity` | Expressing mutual exchange, collaborative communication, inviting response |
| **Empathy Resonance** | `Emre` | `Empathy Resonance` | Demonstrating empathetic connection, emotional attunement, warmth |
| **Cognitive Energy** | `Cogen` | `Cognitive Energy` | Showing mental engagement, cognitive processing, activation, momentum |
| **Baseline Neutral** | `Baseneutral` | `Baseline Neutral` | Neutral control sample, default prosody for comparative analysis |
**Capitalization Rules**:
- First letter capitalized in filenames: `Att`, `Cogen`
- Full words in CSV: `Attention`, `Cognitive Energy`
- Baseline: `Baseneutral` (one word, capital B)
---
## Modifier Codes (24 Optional Sub-Modifiers)
### 1. Trust Modifiers (5)
| Code | Full Name | Description |
|------|-----------|-------------|
| `auth` | Authoritative | Commanding, expert tone |
| `calm` | Calm | Soothing, measured tone |
| `conf` | Confident | Self-assured, certain tone |
| `rest` | Formal/Respectful | Professional, courteous tone |
| `reas` | Reassuring | Comforting, supportive tone |
### 2. Attention Modifiers (5)
| Code | Full Name | Description |
|------|-----------|-------------|
| `cert` | Certainty | Confident, definite tone |
| `clar` | Clarity | Clear, precise communication |
| `curi` | Curious | Inquisitive, interested tone |
| `focu` | Focused | Concentrated, directed attention |
| `urge` | Urgent/Pressure | Time-sensitive, pressing tone |
### 3. Reciprocity Modifiers (5)
| Code | Full Name | Description |
|------|-----------|-------------|
| `affi` | Affirming | Validating, confirming tone |
| `colla` | Collaborative | Cooperative, team-oriented tone |
| `enga` | Engaged | Active, participatory tone |
| `open` | Open | Receptive, non-defensive tone |
| `refl` | Reflective | Thoughtful, contemplative tone |
### 4. Empathy Resonance Modifiers (5)
| Code | Full Name | Description |
|------|-----------|-------------|
| `casu` | Casual | Informal, relaxed tone |
| `comp` | Compassion | Kind, caring tone |
| `corr` | Corrective (softened) | Gentle correction or guidance |
| `symp` | Sympathetic | Understanding, supportive tone |
| `warm` | Warm | Friendly, approachable tone |
### 5. Cognitive Energy Modifiers (4)
| Code | Full Name | Description |
|------|-----------|-------------|
| `ana` | Analytical | Logical, reasoning-oriented tone |
| `dyna` | Dynamic | Energetic, active tone |
| `enth` | Enthusiastic | Excited, passionate tone |
| `skep` | Skeptical | Questioning, doubtful tone |
### Cross-Intent Modifier (1)
| Code | Full Name | Description |
|------|-----------|-------------|
| `ambivalex` | Ambivalence | Mixed, transitional, or competing tonal cues present simultaneously |
**Capitalization Rule**: All modifier codes are lowercase in filenames: `affi`, `warm`, `ana`, `ambivalex`
---
## Segment-Level Data Structure
### JSON Structure in "Segments" Field
Each utterance includes time-aligned segment-level tonality data stored as a JSON array string in the CSV.
**Structure**:
```json
[
{
"startTime": <milliseconds>,
"endTime": <milliseconds>,
"trust": <0-100>,
"reciprocity": <0-100>,
"empathy": <0-100>,
"cognitive": <0-100>,
"attention": <0-100>
}
]
```
**Real Example**:
```json
[{
"startTime": 0,
"endTime": 4284.083333333333,
"trust": 75,
"reciprocity": 93,
"empathy": 76,
"cognitive": 96,
"attention": 80
}]
```
### Segment Field Definitions
| Field | Type | Unit | Description |
|-------|------|------|-------------|
| `startTime` | Float | Milliseconds | Segment start time from utterance beginning |
| `endTime` | Float | Milliseconds | Segment end time from utterance beginning |
| `trust` | Integer | 0-100 | Trust tonality score for this segment |
| `reciprocity` | Integer | 0-100 | Reciprocity score for this segment |
| `empathy` | Integer | 0-100 | Empathy resonance score for this segment |
| `cognitive` | Integer | 0-100 | Cognitive energy score for this segment |
| `attention` | Integer | 0-100 | Attention score for this segment |
**Notes**:
- Most utterances contain a single segment (entire utterance)
- Times in milliseconds with decimal precision
- Segment scores may differ from utterance-level indices in multi-segment utterances
- To convert milliseconds to seconds: `seconds = milliseconds / 1000`
---
## Missing Data Codes
### How Missing Data is Represented
| Field Type | Missing Data Representation |
|-----------|----------------------------|
| String fields (Sub_Modifier, Ambivalex, Notes) | Empty string `""` |
| Numeric fields | No missing data (all utterances fully annotated) |
| Segments | No missing data (all utterances have segment data) |
**Important**:
- There is **NO use of** `-999`, `NULL`, `NA`, or other special missing data codes
- Empty string `""` indicates "not applicable" for optional fields
- All tonality indices are complete (no missing values)
---
## Statistical Summary
### Dataset Overview
| Statistic | Value |
|-----------|-------|
| Total Utterances | 144 |
| Total Batches | 6 |
| Utterances per Batch | 18 |
| Single Speaker | Yes (Ronda) |
| Language | English (American) |
| Recording Period | Dec 19, 2025 - Jan 23, 2026 |
| Total Duration | ~10 minutes |
### Audio Specifications
| Specification | Value |
|---------------|-------|
| Sample Rate | 48,000 Hz |
| Bit Depth | 16-bit |
| Channels | Mono (1) |
| Format | WAV (uncompressed PCM) |
| Duration Range | 3-6 seconds per file |
### Index Distributions
*Note: Actual statistical summaries (mean, SD, min, max) should be calculated from the complete dataset.*
**Expected Patterns**:
- Cognitive_Energy_Index: Known systematic elevation (typically 90-100)
- Other indices: Expected to vary by Primary_Intention category
- See METHODOLOGY.md for quality control discussion
---
## Known Issues & Limitations
### Cognitive Energy Systematic Bias
**Issue**: Cognitive_Energy_Index shows systematic elevation across most utterances, regardless of Primary_Intention category.
**Possible Causes** (as noted in dataset documentation):
1. Speaker's ecological style (natural high-energy delivery)
2. Lexical content effects
3. Practitioner bias in scoring
**Resolution**: Intentionally retained for transparency and to reflect ecological reality of speech production. Researchers should account for this bias in analyses.
**Impact**:
- Trust and Empathy Resonance indices most affected
- Suggests need for speaker-specific normalization in some applications
- Does not invalidate other tonality measures
### Single-Speaker Limitation
- All 144 files from same speaker (Ronda)
- Findings may not generalize to other speakers
- Multi-speaker extension needed for broader applicability
### Controlled Environment
- Professional studio recordings
- May not reflect naturalistic speech conditions
- Scripted content (not spontaneous speech)
---
## Usage Notes
### Loading Data in Python
```python
import pandas as pd
import json
# Load combined CSV
df = pd.read_csv('ALL_TONALITY_DATA_COMBINED.csv')
# Parse Segments JSON
df['Segments_Parsed'] = df['Segments'].apply(json.loads)
# Access first segment's trust score
first_segment_trust = df['Segments_Parsed'].iloc[0][0]['trust']
```
### Loading Data in R
```r
library(readr)
library(jsonlite)
# Load CSV
data <- read_csv('ALL_TONALITY_DATA_COMBINED.csv')
# Parse Segments JSON
data$Segments_Parsed <- lapply(data$Segments, fromJSON)
# Access segment data
first_segment <- data$Segments_Parsed[[1]][[1]]
```
### Filtering by Intention
```python
# Get all Attention utterances
attention_data = df[df['Primary_Intention'] == 'Attention']
# Get all utterances with ambivalence
ambivalent_data = df[df['Ambivalex'] == 'ambivalex']
# Get Trust utterances with calm modifier
trust_calm = df[
(df['Primary_Intention'] == 'Trust') &
(df['Sub_Modifier'] == 'calm')
]
```
---
## Citation
When using this dataset, please cite:
```bibtex
@dataset{polhill_2026_tonalityprint,
author = {Polhill, Ronda},
title = {TonalityPrint: A Contrast-Structured Voice Dataset for Exploring Functional Tonal Intent, Ambivalence, and Inference-Time Prosodic Alignment v1.0},
year = 2026,
publisher = {Zenodo},
version = {1.0.0},
doi = {10.5281/zenodo.17913895},
url = {https://doi.org/10.5281/zenodo.17913895}
}
```
---
## Contact
**Dataset Curator**: Ronda Polhill
**Email**: ronda@TonalityPrint.com
**DOI**: https://doi.org/10.5281/zenodo.17913895
For questions about:
- Variable definitions → This codebook
- Annotation methodology → METHODOLOGY.md
- Dataset usage → DATACARD.md
- Technical issues → ronda@TonalityPrint.com
---
**Version**: 1.0.0
**Last Updated**: January 24, 2026
**License**: CC BY-NC 4.0