|
|
--- |
|
|
license: mit |
|
|
task_categories: |
|
|
- feature-extraction |
|
|
- sentence-similarity |
|
|
- text-retrieval |
|
|
dataset_info: |
|
|
features: |
|
|
- name: text |
|
|
dtype: string |
|
|
- name: length_category |
|
|
dtype: string |
|
|
- name: source |
|
|
dtype: string |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 373887763 |
|
|
num_examples: 8000 |
|
|
download_size: 210093799 |
|
|
dataset_size: 373887763 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/train-* |
|
|
tags: |
|
|
- embedding |
|
|
- benchmark |
|
|
- long-context |
|
|
- deepinfra |
|
|
- rag |
|
|
- wikitext |
|
|
language: |
|
|
- en |
|
|
|
|
|
size_categories: |
|
|
- 1K<n<10K |
|
|
--- |
|
|
|
|
|
|
|
|
# Variable Length Embedding Benchmark (VLEB) |
|
|
|
|
|
## Dataset Summary |
|
|
**VLEB (Variable Length Embedding Benchmark)** is a specialized dataset designed to evaluate the performance, latency, and stability of **embedding models and rerankers** across a wide spectrum of context lengths. |
|
|
|
|
|
Unlike standard datasets that focus on short passages, VLEB provides a balanced distribution of text ranging from standard RAG chunks to maximum-context documents (up to 32k tokens). It is constructed from `wikitext-103-raw-v1` using a **smart-clipping strategy** that preserves semantic integrity without splitting words. |
|
|
|
|
|
This benchmark is essential for: |
|
|
- **Length Generalization:** Testing if models maintain semantic understanding as context grows. |
|
|
- **RAG Profiling:** Measuring encoding latency and memory usage at different bins. |
|
|
- **"Lost-in-the-Middle" Analysis:** Evaluating retrieval degradation in long-context windows. |
|
|
|
|
|
## Data Structure |
|
|
|
|
|
The dataset consists of **8,000 samples**, strictly balanced across 4 length categories (2,000 samples per bin). Token counts are calculated using the `Qwen/Qwen2.5-7B-Instruct` tokenizer. |
|
|
|
|
|
| Category | Token Range (Qwen) | Typical Use Case | |
|
|
| :--- | :--- | :--- | |
|
|
| **Short** | 512 - 2,048 | Standard RAG chunks, abstracts, news snippets. | |
|
|
| **Medium** | 2,048 - 8,192 | Full articles, technical reports, single-file code. | |
|
|
| **Long** | 8,192 - 16,384 | Multiple papers, book chapters, long legal contracts. | |
|
|
| **Very Long** | 16,384 - 32,000 | Entire books, massive documentation, stress testing context limits. | |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
# Load the full dataset |
|
|
dataset = load_dataset("ovuruska/variable-length-embedding-bench") |
|
|
|
|
|
# Filter for specific length requirements |
|
|
short_contexts = dataset.filter(lambda x: x['length_category'] == 'Short') |
|
|
very_long_contexts = dataset.filter(lambda x: x['length_category'] == 'Very Long') |
|
|
|
|
|
print(f"Sample Text ({very_long_contexts[0]['length_category']}):") |
|
|
print(very_long_contexts[0]['text'][:200] + "...") |
|
|
|
|
|
``` |
|
|
|
|
|
## Construction Methodology |
|
|
|
|
|
1. **Source:** The dataset is derived from the `wikitext-103-raw-v1` corpus. |
|
|
2. **Stream Buffering:** The raw text was processed as a continuous stream rather than isolated lines. |
|
|
3. **Smart Clipping:** A buffer system accumulated tokens until a target length (randomly selected within bin ranges) was met. The text was then clipped at the exact token boundary and decoded back to string, ensuring **no words are split** and the text remains natural. |
|
|
4. **Validation:** All samples were re-tokenized to ensure they strictly fall within their assigned bin limits. |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use this dataset for benchmarking, please cite: |
|
|
|
|
|
```bibtex |
|
|
@misc{vleb_2026, |
|
|
author = {DeepInfra Engineering Team}, |
|
|
title = {Variable Length Embedding Benchmark (VLEB)}, |
|
|
year = {2026}, |
|
|
publisher = {Hugging Face}, |
|
|
howpublished = {\url{[https://huggingface.co/datasets/DeepInfra/variable-length-embedding-benchmark](https://huggingface.co/datasets/DeepInfra/variable-length-embedding-benchmark)}} |
|
|
} |
|
|
|
|
|
``` |
|
|
|
|
|
|
|
|
|
|
|
|