--- dataset_info: features: - name: dialog_id dtype: string - name: turns list: - name: bigram_overlap_prev dtype: float64 - name: context_embedding list: float64 - name: intent_label dtype: string - name: is_user dtype: int64 - name: length_bucket dtype: string - name: nb_response_candidates list: string - name: readability dtype: float64 - name: readability_score dtype: float64 - name: role_embedding list: int64 - name: sentiment_polarity dtype: float64 - name: speaker dtype: string - name: text dtype: string splits: - name: train num_bytes: 515339977 num_examples: 13215 download_size: 458215847 dataset_size: 515339977 configs: - config_name: default data_files: - split: train path: data/train-* --- ## Taskmaster-1 Enriched Dialog Dataset (Combined) ## Overview This dataset is a combined, enriched version of the self_dialog and woz_dialog splits from the Taskmaster-1 dataset. It consists of multi-turn, human-human and human-simulated conversations with systematic enhancements for machine learning workflows—especially dialog modeling, generation, and fine-grained evaluation. All conversations are structured in a JSON format with consistent schema and include added semantic, linguistic, and behavioral annotations. ## Enrichments Included 1. Role Embedding Each turn includes a binary role embedding: [1, 0] for USER [0, 1] for ASSISTANT This makes it easier for sequence models to learn speaker turns without relying on string labels. Use case: Improves model performance in transformer-based dialog agents by allowing role-aware generation and classification. 2. Response Candidates Each user turn is enriched with nb_response_candidates — 2 to 4 plausible assistant responses sampled from the dataset. These are not ground truth but plausible continuations. Use case: Ideal for retrieval-based dialog training or negative sampling in response ranking tasks. 3. Readability Score Computed using Flesch-Kincaid metrics and other NLP readability formulas. Stored as readability (0–100 scale, higher = easier). Use case: Enables analysis of language complexity and training adaptive LLMs for education, accessibility, or voice interfaces. 4. Readability Grade Score Stored as readability_score on a U.S. grade level (lower = easier to read). Especially relevant for UX tuning. Use case: Allows controlling reading level in generation tasks or selecting user-appropriate training samples. 5. Context Embedding Each turn is augmented with a context_embedding vector (384-dim, Sentence-BERT). Represents the semantic context of the turn. Use case: Enables plug-and-play use with FAISS-based semantic search, response reranking, and memory-augmented generation. 6. Speaker Role Flags An is_user flag is included for each turn (1 = user, 0 = assistant). Use case: Simplifies filtering, evaluation, or role-specific metric computation. 7. Utterance Length Bucketing Each turn is labeled as: short (<= 5 tokens) medium (6–15 tokens) long (> 15 tokens) Use case: Enables sampling, curriculum learning, or model analysis across turn complexity. 8. Bigram Overlap with Previous Turn Computed as bigram_overlap_prev (float between 0 and 1). Measures lexical repetition with the preceding utterance. Use case: Useful for: Dialogue coherence metrics Detecting stagnation or repetition in generated responses Analyzing repair-based utterances 9. Sentiment Polarity Computed using a sentiment analyzer. Stored as sentiment_polarity: Ranges from –1 (strongly negative) to +1 (strongly positive) Use case: Enables emotion-aware generation, tone control, or training sentiment-conditioned agents. 10. Format Summary Each conversation has: dialog_id: Unique identifier turns: List of enriched utterances Each turn includes: { "speaker": "USER", "text": "I’d like to book a table for 2", "role_embedding": [1, 0], "intent_label": "request", "nb_response_candidates": [...], "readability_score": 4.5, "context_embedding": [...], "readability": 85.6, "is_user": 1, "length_bucket": "medium", "bigram_overlap_prev": 0.2, "sentiment_polarity": 0.1 } ## Suggested Use Cases Fine-tuning LLMs for goal-oriented dialog Training dialog state trackers and response rankers Evaluating model outputs with context-aware metrics Curriculum learning based on length or readability Emotion- and intent-conditioned dialog modeling Semantic retrieval and reranking systems ## Citation @inproceedings{48484, title = {Taskmaster-1: Toward a Realistic and Diverse Dialog Dataset}, author = {Bill Byrne and Karthik Krishnamoorthi and Chinnadhurai Sankar and Arvind Neelakantan and Daniel Duckworth and Semih Yavuz and Ben Goodrich and Amit Dubey and Kyu-Young Kim and Andy Cedilnik}, year = {2019} } ## Taskmaster-1: Towards a Realistic Goal-Oriented Dialogue Dataset (Google-Research-Datasets) ## Original base dataset: @patil-suraj (Original contributor) ## Enrichments and combined version by: GenAIDevTOProd (Adithya) ## License: Same as Taskmaster-1 (if public domain or open license)