npy listlengths 75 751 | ref.npy listlengths 76 750 ⌀ | json dict | __key__ stringlengths 27 44 | __url__ stringclasses 9
values |
|---|---|---|---|---|
[[2.279296875,0.09014892578125,-0.437744140625,-0.779296875,-0.82275390625,-0.5615234375,-0.17333984(...TRUNCATED) | [[-0.0911865234375,0.27587890625,0.4677734375,0.35888671875,-0.876953125,-0.01708984375,-0.8359375,-(...TRUNCATED) | {"__key__":"638569_00126632","annotation_scores":{"Affection":1.291,"Age":0.8398,"Amusement":2.2109,(...TRUNCATED) | podcastt_000001_638569_00126632 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[0.441650390625,-0.02886962890625,0.548828125,-0.227294921875,0.98046875,-1.220703125,-0.8979492187(...TRUNCATED) | [[1.6435546875,-0.86328125,-0.54248046875,0.327880859375,-0.724609375,-0.51123046875,0.369140625,-0.(...TRUNCATED) | {"__key__":"594432_00245696","annotation_scores":{"Affection":0.0096,"Age":0.8379,"Amusement":2.5684(...TRUNCATED) | podcastt_000002_594432_00245696 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[0.0635986328125,-0.634765625,-0.50439453125,-1.1123046875,-0.08233642578125,0.1146240234375,-0.317(...TRUNCATED) | [[0.13232421875,-0.36083984375,0.26416015625,0.8193359375,-0.1304931640625,0.343017578125,0.6796875,(...TRUNCATED) | {"__key__":"106569_00005704","annotation_scores":{"Affection":1.0625,"Age":0.7832,"Amusement":1.4287(...TRUNCATED) | podcastt_000010_106569_00005704 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[-1.044921875,1.31640625,1.3515625,0.444580078125,-0.210205078125,-0.1419677734375,0.213134765625,0(...TRUNCATED) | [[-1.087890625,0.6923828125,0.62744140625,-0.2193603515625,0.482421875,-0.66845703125,-0.30981445312(...TRUNCATED) | {"__key__":"319224_00197619","annotation_scores":{"Affection":0.0073,"Age":0.8467,"Amusement":0.7275(...TRUNCATED) | podcastt_000010_319224_00197619 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[-0.2086181640625,0.2022705078125,0.407470703125,-0.135009765625,0.1680908203125,-1.2568359375,-1.0(...TRUNCATED) | [[0.9208984375,-0.4228515625,-0.060028076171875,-0.64697265625,0.79443359375,0.1695556640625,0.97460(...TRUNCATED) | {"__key__":"112090_00112392","annotation_scores":{"Affection":0.0073,"Age":0.8525,"Amusement":0.7271(...TRUNCATED) | podcastt_000012_112090_00112392 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[0.04119873046875,0.4638671875,0.3359375,-1.14453125,-0.343994140625,-0.40478515625,-0.689453125,-0(...TRUNCATED) | [[-0.189697265625,-1.4375,-0.4140625,-0.70654296875,1.16015625,0.5244140625,-0.273193359375,0.345947(...TRUNCATED) | {"__key__":"106569_00044104","annotation_scores":{"Affection":0.0057,"Age":0.8389,"Amusement":1.0811(...TRUNCATED) | podcastt_000012_106569_00044104 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[-0.490478515625,1.6591796875,-1.6708984375,-0.309814453125,-1.080078125,1.49609375,-1.9677734375,1(...TRUNCATED) | [[0.49951171875,-0.4853515625,0.4716796875,0.07952880859375,-0.349609375,-0.275634765625,0.374267578(...TRUNCATED) | {"__key__":"900574_00056591","annotation_scores":{"Affection":0.0073,"Age":0.8496,"Amusement":1.2402(...TRUNCATED) | podcastt_000013_900574_00056591 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[-1.40625,-1.0390625,0.119140625,0.447998046875,0.9404296875,-1.2294921875,0.6826171875,-0.78320312(...TRUNCATED) | [[1.2734375,0.47265625,-2.177734375,-0.1771240234375,0.329833984375,0.07000732421875,0.75341796875,0(...TRUNCATED) | {"__key__":"809809_00014080","annotation_scores":{"Affection":1.3574,"Age":0.8486,"Amusement":1.1416(...TRUNCATED) | podcastt_000013_809809_00014080 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[0.329345703125,-0.24609375,0.0113525390625,-0.47021484375,-1.111328125,-0.221435546875,-2.19921875(...TRUNCATED) | [[-0.1220703125,0.4716796875,1.43359375,-0.281005859375,0.1973876953125,-0.92236328125,-0.4482421875(...TRUNCATED) | {"__key__":"330085_00187696","annotation_scores":{"Affection":0.0059,"Age":0.8516,"Amusement":0.9644(...TRUNCATED) | podcastt_000013_330085_00187696 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
[[-0.0076904296875,0.476806640625,-1.580078125,-1.2939453125,0.428466796875,-1.1376953125,-0.4516601(...TRUNCATED) | [[-0.004619598388671875,-1.2119140625,-0.2744140625,0.282958984375,-0.578125,1.458984375,0.263183593(...TRUNCATED) | {"__key__":"833995_00183052","annotation_scores":{"Affection":1.2344,"Age":0.8418,"Amusement":1.04,"(...TRUNCATED) | podcastt_000020_833995_00183052 | "hf://datasets/TTS-AGI/emotion-attribute-conditioning-dacvae@a17ca2046e864f12632c011451b57a0c0c2b981(...TRUNCATED) |
YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
Echo TTS - Emotion & Attribute Conditioning Dataset (DAC-VAE Latents)
Pre-bucketed speech dataset with DAC-VAE latent representations organized by 40 emotion categories and 13 vocal/audio attributes. Built for conditioning fine-tuning of Echo TTS and similar DiT-based TTS models.
Overview
- Total emotion samples: 163,271 (across 40 emotions, 10K cap per emotion)
- Total attribute samples: ~785K (across 13 attributes x 7 buckets, 10K cap per bucket)
- Format: WebDataset
.tarfiles (312 total) - Latent format: DAC-VAE 128-dim at 25 fps (variable length)
- Source: Filtered from 71.2M samples across 3 datasets
Source Datasets
- TTS-AGI/podcast-tokenized-bg3.5-enj5 (481 tars)
- TTS-AGI/podcast-tokenized-bg2.5-enj4.5 (6,496 tars)
- TTS-AGI/emolia-hq-tokenized (10,491 tars)
Annotation Source
All emotion and attribute scores from LAION Empathic-Insight-Voice-Plus via emotion-annotations.
Data Format
Each sample in the tar files contains 3 files:
| File | Shape | Description |
|---|---|---|
{key}.npy |
(frames, 128) float32 |
DAC-VAE latent (target speech, 25 fps) |
{key}.ref.npy |
(ref_frames, 128) float32 |
DAC-VAE latent (speaker reference) |
{key}.json |
- | Metadata with text, scores, speaker info |
JSON Metadata Fields
| Field | Type | Description |
|---|---|---|
text |
string | Transcript text |
annotation_scores |
dict (59 keys) | All emotion + attribute scores |
caption |
string | Natural language description of the audio |
target_duration |
float | Duration in seconds |
context_duration |
float | Speaker reference duration |
latent_frames |
int | Number of latent frames in .npy |
ref_latent_frames |
int | Number of frames in ref.npy |
latent_dim |
int | Always 128 |
speaker |
string | Speaker ID |
language |
string | Language code |
episode_id |
int | Source episode identifier |
Loading Example
import webdataset as wds
import numpy as np
import json
dataset = wds.WebDataset("emotions/Amusement/000000.tar")
for sample in dataset:
key = sample["__key__"]
latent = np.frombuffer(sample["npy"], dtype=np.float32).reshape(-1, 128)
ref_latent = np.frombuffer(sample["ref.npy"], dtype=np.float32).reshape(-1, 128)
metadata = json.loads(sample["json"])
text = metadata["text"]
scores = metadata["annotation_scores"]
print(f"{key}: {latent.shape}, text='{text[:60]}...', amusement={scores.get('Amusement', 0):.2f}")
Directory Structure
emotions/
{emotion_name}/ # 40 emotion directories
000000.tar # WebDataset tar (up to 4096 samples each)
000001.tar
...
attributes/
{attribute_name}/ # 13 attribute directories
bucket_{i}_{lo}_to_{hi}/ # 7 range buckets per attribute
000000.tar
...
Emotion Buckets (40 emotions)
Samples are routed to the emotion with the highest score, if that score >= 2.5. Each sample appears in at most one emotion bucket.
| Emotion | Samples | Emotion | Samples | |
|---|---|---|---|---|
| Interest | 10,000 | Pain | 1,220 | |
| Hope_Enthusiasm_Optimism | 10,000 | Shame | 1,215 | |
| Thankfulness_Gratitude | 10,000 | Infatuation | 934 | |
| Amusement | 10,000 | Malevolence_Malice | 874 | |
| Concentration | 10,000 | Doubt | 507 | |
| Impatience_and_Irritability | 10,000 | Disgust | 452 | |
| Contemplation | 10,000 | Awe | 421 | |
| Anger | 10,000 | Contentment | 376 | |
| Astonishment_Surprise | 10,000 | Contempt | 348 | |
| Affection | 10,000 | Teasing | 282 | |
| Confusion | 9,756 | Jealousy_&_Envy | 203 | |
| Intoxication_Altered_States | 9,375 | Helplessness | 194 | |
| Distress | 6,182 | Embarrassment | 96 | |
| Fatigue_Exhaustion | 5,672 | Bitterness | 74 | |
| Longing | 4,452 | Pleasure_Ecstasy | 45 | |
| Relief | 3,822 | Sourness | 27 | |
| Sadness | 3,634 | |||
| Elation | 3,042 | |||
| Triumph | 2,001 | |||
| Sexual_Lust | 1,967 | |||
| Emotional_Numbness | 1,721 | |||
| Fear | 1,676 | |||
| Pride | 1,374 | |||
| Disappointment | 1,329 |
Total: 163,271 samples across 40 emotions
Attribute Buckets (13 attributes x 7 levels)
Each attribute is divided into 7 linear buckets. Samples can appear in multiple attribute buckets (one per attribute).
Age (0.0 - 6.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | 0.0 - 0.9 | 10,000 |
| 1 | 0.9 - 1.7 | 10,000 |
| 2 | 1.7 - 2.6 | 10,000 |
| 3 | 2.6 - 3.4 | 10,000 |
| 4 | 3.4 - 4.3 | 10,000 |
| 5 | 4.3 - 5.1 | 1,068 |
| 6 | 5.1 - 6.0 | 3 |
Arousal (0.0 - 4.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | 0.0 - 0.6 | 10,000 |
| 1 | 0.6 - 1.1 | 10,000 |
| 2 | 1.1 - 1.7 | 10,000 |
| 3 | 1.7 - 2.3 | 10,000 |
| 4 | 2.3 - 2.9 | 10,000 |
| 5 | 2.9 - 3.4 | 10,000 |
| 6 | 3.4 - 4.0 | 5,402 |
Confident vs. Hesitant (0.0 - 4.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | 0.0 - 0.6 | 10,000 |
| 1 | 0.6 - 1.1 | 10,000 |
| 2 | 1.1 - 1.7 | 10,000 |
| 3 | 1.7 - 2.3 | 10,000 |
| 4 | 2.3 - 2.9 | 10,000 |
| 5 | 2.9 - 3.4 | 10,000 |
| 6 | 3.4 - 4.0 | 1,376 |
Gender (-2.0 - 2.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | -2.0 - -1.4 | 10,000 |
| 1 | -1.4 - -0.9 | 10,000 |
| 2 | -0.9 - -0.3 | 10,000 |
| 3 | -0.3 - 0.3 | 10,000 |
| 4 | 0.3 - 0.9 | 10,000 |
| 5 | 0.9 - 1.4 | 10,000 |
| 6 | 1.4 - 2.0 | 10,000 |
Monotone vs. Expressive (0.0 - 4.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | 0.0 - 0.6 | 4,928 |
| 1 | 0.6 - 1.1 | 10,000 |
| 2 | 1.1 - 1.7 | 10,000 |
| 3 | 1.7 - 2.3 | 10,000 |
| 4 | 2.3 - 2.9 | 10,000 |
| 5 | 2.9 - 3.4 | 10,000 |
| 6 | 3.4 - 4.0 | 10,000 |
Serious vs. Humorous (0.0 - 4.0)
All 7 buckets at 10,000 samples.
Soft vs. Harsh (-2.0 - 2.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | -2.0 - -1.4 | 349 |
| 1-5 | ... | 10,000 each |
| 6 | 1.4 - 2.0 | 1,176 |
Valence (-3.0 - 3.0)
All 7 buckets at 10,000 samples.
Warm vs. Cold (-2.0 - 2.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | -2.0 - -1.4 | 420 |
| 1-6 | ... | 10,000 each |
Duration (1.0 - 30.0 seconds)
All 7 buckets at 10,000 samples.
Talking Speed (5.0 - 25.0 CPS)
All 7 buckets at 10,000 samples.
High-Pitched vs. Low-Pitched (0.0 - 4.0)
| Bucket | Range | Samples |
|---|---|---|
| 0 | 0.0 - 0.6 | 6 |
| 1-5 | ... | 10,000 each |
| 6 | 3.4 - 4.0 | 10 |
Submissive vs. Dominant (-3.0 - 3.0)
| Bucket | Range | Samples |
|---|---|---|
| 2 | -1.3 - -0.4 | 162 |
| 3 | -0.4 - 0.4 | 10,000 |
| 4 | 0.4 - 1.3 | 10,000 |
| 5 | 1.3 - 2.1 | 10,000 |
| 6 | 2.1 - 3.0 | 86 |
59 Annotation Score Dimensions
The annotation_scores field in each JSON contains all 59 dimensions from Empathic-Insight-Voice-Plus:
Emotions (40): Affection, Amusement, Anger, Astonishment_Surprise, Awe, Bitterness, Concentration, Confusion, Contemplation, Contempt, Contentment, Disappointment, Disgust, Distress, Doubt, Elation, Embarrassment, Emotional_Numbness, Fatigue_Exhaustion, Fear, Helplessness, Hope_Enthusiasm_Optimism, Impatience_and_Irritability, Infatuation, Interest, Intoxication_Altered_States_of_Consciousness, Jealousy_&_Envy, Longing, Malevolence_Malice, Pain, Pleasure_Ecstasy, Pride, Relief, Sadness, Sexual_Lust, Shame, Sourness, Teasing, Thankfulness_Gratitude, Triumph
Vocal Attributes (15): Age, Arousal, Authenticity, Background_Noise, Confident_vs._Hesitant, Gender, High-Pitched_vs._Low-Pitched, Monotone_vs._Expressive, Recording_Quality, Serious_vs._Humorous, Soft_vs._Harsh, Submissive_vs._Dominant, Valence, Vulnerable_vs._Emotionally_Detached, Warm_vs._Cold
Audio Quality (4): score_background_quality, score_content_enjoyment, score_overall_quality, score_speech_quality
Bucketing Strategy
Emotion routing: Each sample is assigned to the emotion with the highest score, if that score >= 2.5. Samples below threshold or with no strong emotion are excluded. Each sample appears in at most 1 emotion bucket.
Attribute routing: 7 linearly-spaced buckets per attribute. Each sample is placed into all matching attribute buckets (one per attribute dimension). Bucket size capped at 10,000 samples.
Pipeline Details
- Samples scanned: 71,242,881
- Processing time: ~15 hours
- Datasets processed: 17,468 tar files across 3 HuggingFace datasets
Citation
@misc{echo-conditioning-data,
title={Echo TTS Emotion & Attribute Conditioning Dataset},
url={https://huggingface.co/datasets/TTS-AGI/emotion-attribute-conditioning-dacvae},
note={Bucketed DAC-VAE latent speech data with 59-dimensional emotion/attribute annotations},
}
Scores from LAION Empathic-Insight-Voice-Plus and emotion-annotations.
- Downloads last month
- 9,230