Dataset Preview
Duplicate
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    FileNotFoundError
Message:      tencent/VCB-Bench@39c3387c65423c4218329f3d68846d4ce41ce416/vcb_bench/pretrain_evaluation/story_continue/audio/context_sc_common_000001.wav (repository not found)
Traceback:    Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 409, in hf_raise_for_status
                  response.raise_for_status()
                File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 1026, in raise_for_status
                  raise HTTPError(http_error_msg, response=self)
              requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: https://huggingface.co/api/datasets/tencent/VCB-Bench/revision/39c3387c65423c4218329f3d68846d4ce41ce416
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 125, in _repo_and_revision_exist
                  self._api.repo_info(
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
                  return fn(*args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_api.py", line 2816, in repo_info
                  return method(
                         ^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
                  return fn(*args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_api.py", line 2674, in dataset_info
                  hf_raise_for_status(r)
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 459, in hf_raise_for_status
                  raise _format(RepositoryNotFoundError, message, response) from e
              huggingface_hub.errors.RepositoryNotFoundError: 401 Client Error. (Request ID: Root=1-694bd4b2-6b1889ea2fb28d284edf027a;0dccb3a0-4f1b-4a48-85d5-2a081774f1dd)
              
              Repository Not Found for url: https://huggingface.co/api/datasets/tencent/VCB-Bench/revision/39c3387c65423c4218329f3d68846d4ce41ce416.
              Please make sure you specified the correct `repo_id` and `repo_type`.
              If you are trying to access a private or gated repo, make sure you are authenticated. For more details, see https://huggingface.co/docs/huggingface_hub/authentication
              Invalid username or password.
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1595, in _prepare_split_single
                  num_examples, num_bytes = writer.finalize()
                                            ^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 728, in finalize
                  self.write_examples_on_file()
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 581, in write_examples_on_file
                  self.write_batch(batch_examples=batch_examples)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 701, in write_batch
                  self.write_table(pa_table, writer_batch_size)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 716, in write_table
                  pa_table = embed_table_storage(pa_table)
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2249, in embed_table_storage
                  embed_array_storage(table[name], feature, token_per_repo_id=token_per_repo_id)
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1795, in wrapper
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2124, in embed_array_storage
                  return feature.embed_storage(array, token_per_repo_id=token_per_repo_id)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/features/audio.py", line 291, in embed_storage
                  (path_to_bytes(x["path"]) if x["bytes"] is None else x["bytes"]) if x is not None else None
                   ^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/utils/py_utils.py", line 310, in wrapper
                  return func(value) if value is not None else None
                         ^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/features/audio.py", line 286, in path_to_bytes
                  with xopen(path, "rb", download_config=download_config) as f:
                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 935, in xopen
                  file_obj = fsspec.open(file, mode=mode, *args, **kwargs).open()
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/fsspec/core.py", line 135, in open
                  return self.__enter__()
                         ^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/fsspec/core.py", line 103, in __enter__
                  f = self.fs.open(self.path, mode=mode)
                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "<string>", line 3, in open
                File "/usr/local/lib/python3.12/unittest/mock.py", line 1139, in __call__
                  return self._mock_call(*args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/unittest/mock.py", line 1143, in _mock_call
                  return self._execute_mock_call(*args, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/unittest/mock.py", line 1204, in _execute_mock_call
                  result = effect(*args, **kwargs)
                           ^^^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 770, in wrapped
                  f = fs_open(self, urlpath, mode, *args, **kwargs)
                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1293, in open
                  f = self._open(
                      ^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 275, in _open
                  return HfFileSystemFile(self, path, mode=mode, revision=revision, block_size=block_size, **kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 938, in __init__
                  self.resolved_path = fs.resolve_path(path, revision=revision)
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 200, in resolve_path
                  _raise_file_not_found(path, err)
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1138, in _raise_file_not_found
                  raise FileNotFoundError(msg) from err
              FileNotFoundError: tencent/VCB-Bench@39c3387c65423c4218329f3d68846d4ce41ce416/vcb_bench/pretrain_evaluation/story_continue/audio/context_sc_common_000001.wav (repository not found)
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1334, in compute_config_parquet_and_info_response
                  parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet(
                                                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 911, in stream_convert_to_parquet
                  builder._prepare_split(
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1447, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1604, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

audio
audio
End of preview.
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/datasets-cards)

VCB-Bench: An Evaluation Benchmark for Audio-Grounded Large Language Model Conversational Agents

arXiv GitHub Hugging Face

Introduction

Voice Chat Bot Bench (VCB Bench) is a high-quality Chinese benchmark built entirely on real human speech. It evaluates large audio language models (LALMs) along three complementary dimensions:
(1) Instruction following: Text Instruction Following (TIF), Speech Instruction Following (SIF), English Text Instruction Following (TIF-En), English Speech Instruction Following (SIF-En) and Multi-turn Dialog (MTD);
(2) Knowledge: General Knowledge (GK), Mathematical Logic (ML), Discourse Comprehension (DC) and Story Continuation (SC).
(3) Robustness: Speaker Variations (SV), Environmental Variations (EV), and Content Variations (CV).

Getting Started

Installation:

git clone https://github.com/Tencent/VCB-Bench.git
cd VCB-Bench
pip install -r requirements.txt

Note: To evaluate Qwen3-omni, please replace it with the environment it requires.

Download Dataset:

Download the dataset from Hugging Face and place the 'vcb_bench' into 'data/downloaded_datasets'.

Evaluation:

This code is adapted from Kimi-Audio-Evalkit, where you can find more details about the evaluation commands.

(1) Inference + Evaluation:

python run_audio.py --model {model_name} --data {data_name}

For example:

CUDA_VISIBLE_DEVICES=1 python run_audio.py --model Qwen2.5-Omni-7B --data general_knowledge

(2) Only Inference:

python run_audio.py --model {model_name} --data {data_name} --skip-eval

For example:

CUDA_VISIBLE_DEVICES=4,5,6,7 python run_audio.py --model  StepAudio  --data continuation_en  creation_en  empathy_en  recommendation_en  rewriting_en  safety_en  simulation_en emotional_control_en  language_control_en  non_verbal_vocalization_en  pacing_control_en  style_control_en  volume_control_en --skip-eval 

(3) Only Evaluation:

python run_audio.py --model {model_name} --data {data_name} --reeval

For example:

CUDA_VISIBLE_DEVICES=2 nohup python run_audio.py --model  Mimo-Audio --data continuation  creation  empathy --reeval

(4) Inference + ASR + Evaluation:

python run_audio.py --model {model_name} --data {data_name} --wasr

For example:

CUDA_VISIBLE_DEVICES=3 python run_audio.py --model  StepAudio2 --data rewriting  safety  simulation  continuation_en  --wasr 

Format Result:

python sumup_eval.py --model {model_name}
python sumup_eval.py --model {model_name} --export_excel --output_file my_results.xlsx

Supported Datasets and Models

(1) Locate the dataset you need to evaluate from the Data Name column in the Datasets table, and populate the {data_name} parameter in the evaluation command accordingly.
(2) Each dataset in the SV, EV, and CV sections has a corresponding comparison dataset named "{data_name}_cmp", following the specified naming convention.
(3) Identify the model you intend to evaluate from the Model Name column in the Models table, and insert the appropriate {model_name} into the evaluation command.

Datasets:

Data Type Data Name Detail
TIF continuation -
creation -
empathy -
recommendation -
rewriting -
safety -
simulation -
TIF-En continuation_en -
creation_en -
empathy_en -
recommendation_en -
rewriting_en -
safety_en -
simulation_en -
SIF emotional_control -
language_control -
non_verbal_vocalization -
pacing_control -
style_control -
volume_control -
SIF-En emotional_control_en -
language_control_en -
non_verbal_vocalization_en -
pacing_control_en -
style_control_en -
volume_control_en -
MTD progression -
backtracking -
transition -
GK general_knowledge mathematics, geography, politics, chemistry, biology, law, physics, history, medicine, economics, sports, culture
ML basic_math -
math -
logical_reasoning analysis, induction, analogy, logic
DC discourse_comprehension inference, induction, analysis
SV age child, elder
accent tianjin, beijing, dongbei, sichuan
volume down, up
speed -
EV non_vocal_noise echo, outdoors, far_field
vocal_noise TV_playback, background_chat, vocal_music, voice_announcement
unstable_signal -
CV casual_talk -
mispronunciation -
grammatical_error -
topic_shift -
code_switching -

Models:

Model Type Model Name
Chat Model Qwen2-Audio-7B-Instruct
Qwen2.5-Omni-7B
Baichuan-Audio-Chat
GLM4-Voice
Kimi-Audio
Mimo-Audio
StepAudio
StepAudio2
GPT4O-Audio
Qwen3-Omni-Instruct
Pretrain Model Qwen2-Audio-7B
Baichuan-Audio
Kimi-Audio-Base
StepAudio2-Base

Acknowledge

We borrow some code from Kimi-Audio-Evalkit, GLM-4-Voice, Baichuan-Audio, Kimi-Audio, Mimo-Audio, Step-Audio2, and StepAudio.

Citation

@misc{hu2025vcbbenchevaluationbenchmark,
      title={VCB Bench: An Evaluation Benchmark for Audio-Grounded Large Language Model Conversational Agents}, 
      author={Jiliang Hu and Wenfu Wang and Zuchao Li and Chenxing Li and Yiyang Zhao and Hanzhao Li and Liqiang Zhang and Meng Yu and Dong Yu},
      year={2025},
      eprint={2510.11098},
      archivePrefix={arXiv},
      primaryClass={cs.SD},
      url={https://arxiv.org/abs/2510.11098}, 
}
Downloads last month
-