The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    HfHubHTTPError
Message:      429 Client Error: Too Many Requests for url: https://huggingface.co/datasets/Tnaot/large-dataset-audio/resolve/7e5eb9de649f4a96cb2537cc82fb16f325f65465/audio/audio_00469.mp3 (Request ID: Root=1-693aa9a6-6a1924020f1fd1e1784ad4bc;1032a286-70ba-4501-8d91-49daa7000c89)

maximum queue size reached
Traceback:    Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 409, in hf_raise_for_status
                  response.raise_for_status()
                File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 1026, in raise_for_status
                  raise HTTPError(http_error_msg, response=self)
              requests.exceptions.HTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co/datasets/Tnaot/large-dataset-audio/resolve/7e5eb9de649f4a96cb2537cc82fb16f325f65465/audio/audio_00469.mp3
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1586, in _prepare_split_single
                  writer.write(example, key)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 623, in write
                  self.write_examples_on_file()
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 581, in write_examples_on_file
                  self.write_batch(batch_examples=batch_examples)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 701, in write_batch
                  self.write_table(pa_table, writer_batch_size)
                File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 716, in write_table
                  pa_table = embed_table_storage(pa_table)
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2249, in embed_table_storage
                  embed_array_storage(table[name], feature, token_per_repo_id=token_per_repo_id)
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1795, in wrapper
                  return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2124, in embed_array_storage
                  return feature.embed_storage(array, token_per_repo_id=token_per_repo_id)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/features/audio.py", line 291, in embed_storage
                  (path_to_bytes(x["path"]) if x["bytes"] is None else x["bytes"]) if x is not None else None
                   ^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/utils/py_utils.py", line 310, in wrapper
                  return func(value) if value is not None else None
                         ^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/features/audio.py", line 287, in path_to_bytes
                  return f.read()
                         ^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 813, in read_with_retries
                  out = read(*args, **kwargs)
                        ^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 728, in track_read
                  out = f_read(*args, **kwargs)
                        ^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1012, in read
                  out = f.read()
                        ^^^^^^^^
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 728, in track_read
                  out = f_read(*args, **kwargs)
                        ^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1078, in read
                  hf_raise_for_status(self.response)
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 482, in hf_raise_for_status
                  raise _format(HfHubHTTPError, str(e), response) from e
              huggingface_hub.errors.HfHubHTTPError: 429 Client Error: Too Many Requests for url: https://huggingface.co/datasets/Tnaot/large-dataset-audio/resolve/7e5eb9de649f4a96cb2537cc82fb16f325f65465/audio/audio_00469.mp3 (Request ID: Root=1-693aa9a6-6a1924020f1fd1e1784ad4bc;1032a286-70ba-4501-8d91-49daa7000c89)
              
              maximum queue size reached
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1334, in compute_config_parquet_and_info_response
                  parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet(
                                                                        ^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 911, in stream_convert_to_parquet
                  builder._prepare_split(
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1447, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1604, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

audio
audio
End of preview.