Dataset Viewer
Duplicate
The dataset viewer is not available for this split.
The info cannot be fetched for the config 'default' of the dataset.
Error code:   InfoError
Exception:    SplitsNotFoundError
Message:      The split names could not be parsed from the dataset config.
Traceback:    Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 779, in _error_catcher
                  yield
                File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 925, in _raw_read
                  raise IncompleteRead(self._fp_bytes_read, self.length_remaining)
              urllib3.exceptions.IncompleteRead: IncompleteRead(3549783 bytes read, 1693641 more expected)
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 820, in generate
                  yield from self.raw.stream(chunk_size, decode_content=True)
                File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 1091, in stream
                  data = self.read(amt=amt, decode_content=decode_content)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 1008, in read
                  data = self._raw_read(amt)
                         ^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 903, in _raw_read
                  with self._error_catcher():
                       ^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/contextlib.py", line 158, in __exit__
                  self.gen.throw(value)
                File "/usr/local/lib/python3.12/site-packages/urllib3/response.py", line 803, in _error_catcher
                  raise ProtocolError(arg, e) from e
              urllib3.exceptions.ProtocolError: ('Connection broken: IncompleteRead(3549783 bytes read, 1693641 more expected)', IncompleteRead(3549783 bytes read, 1693641 more expected))
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 286, in get_dataset_config_info
                  for split_generator in builder._split_generators(
                                         ^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/hdf5/hdf5.py", line 61, in _split_generators
                  self.info.features = _recursive_infer_features(h5)
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/hdf5/hdf5.py", line 220, in _recursive_infer_features
                  features = _recursive_infer_features(dset)
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/hdf5/hdf5.py", line 220, in _recursive_infer_features
                  features = _recursive_infer_features(dset)
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/hdf5/hdf5.py", line 218, in _recursive_infer_features
                  for path, dset in h5_obj.items():
                                    ^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/h5py/_hl/base.py", line 436, in __iter__
                  for key in self._mapping:
                             ^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/h5py/_hl/group.py", line 502, in __iter__
                  for x in self.id.__iter__():
                           ^^^^^^^^^^^^^^^^^^
                File "h5py/h5g.pyx", line 479, in h5py.h5g.GroupID.__iter__
                File "h5py/h5g.pyx", line 480, in h5py.h5g.GroupID.__iter__
                File "h5py/h5g.pyx", line 112, in h5py.h5g.GroupIter.__init__
                File "h5py/_objects.pyx", line 56, in h5py._objects.with_phil.wrapper
                File "h5py/_objects.pyx", line 57, in h5py._objects.with_phil.wrapper
                File "h5py/h5g.pyx", line 346, in h5py.h5g.GroupID.get_num_objs
                File "h5py/h5fd.pyx", line 162, in h5py.h5fd.H5FD_fileobj_read
                File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1856, in readinto
                  data = self.read(out.nbytes)
                         ^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 844, in read_with_retries
                  out = read(*args, **kwargs)
                        ^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1015, in read
                  return super().read(length)
                         ^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1846, in read
                  out = self.cache._fetch(self.loc, self.loc + length)
                        ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/fsspec/caching.py", line 189, in _fetch
                  self.cache = self.fetcher(start, end)  # new block replaces old
                               ^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 969, in _fetch_range
                  r = http_backoff(
                      ^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 310, in http_backoff
                  response = session.request(method=method, url=url, **kwargs)
                             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 589, in request
                  resp = self.send(prep, **send_kwargs)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 724, in send
                  history = [resp for resp in gen]
                                              ^^^
                File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 265, in resolve_redirects
                  resp = self.send(
                         ^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/requests/sessions.py", line 746, in send
                  r.content
                File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 902, in content
                  self._content = b"".join(self.iter_content(CONTENT_CHUNK_SIZE)) or b""
                                  ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 822, in generate
                  raise ChunkedEncodingError(e)
              requests.exceptions.ChunkedEncodingError: ('Connection broken: IncompleteRead(3549783 bytes read, 1693641 more expected)', IncompleteRead(3549783 bytes read, 1693641 more expected))
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 223, in compute_first_rows_from_streaming_response
                  info = get_dataset_config_info(path=dataset, config_name=config, token=hf_token)
                         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
                File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 291, in get_dataset_config_info
                  raise SplitsNotFoundError("The split names could not be parsed from the dataset config.") from err
              datasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

No dataset card yet

Downloads last month
32