Search is not available for this dataset
The dataset viewer is not available for this split.
The info cannot be fetched for the config 'default' of the dataset.
Error code: InfoError
Exception: SplitsNotFoundError
Message: The split names could not be parsed from the dataset config.
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 286, in get_dataset_config_info
for split_generator in builder._split_generators(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/parquet/parquet.py", line 118, in _split_generators
self.info.features = datasets.Features.from_arrow_schema(pq.read_schema(f))
^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/pyarrow/parquet/core.py", line 2392, in read_schema
file = ParquetFile(
^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/pyarrow/parquet/core.py", line 328, in __init__
self.reader.open(
File "pyarrow/_parquet.pyx", line 1656, in pyarrow._parquet.ParquetReader.open
File "pyarrow/error.pxi", line 92, in pyarrow.lib.check_status
pyarrow.lib.ArrowInvalid: Parquet magic bytes not found in footer. Either the file is corrupted or this is not a parquet file.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/split/first_rows.py", line 223, in compute_first_rows_from_streaming_response
info = get_dataset_config_info(path=dataset, config_name=config, token=hf_token)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 291, in get_dataset_config_info
raise SplitsNotFoundError("The split names could not be parsed from the dataset config.") from err
datasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
This dataset was created using LeRobot.
Dataset Structure
{
"codebase_version": "v3.0",
"robot_type": "omy_f3m",
"total_episodes": 15,
"total_frames": 9001,
"total_tasks": 1,
"chunks_size": 1000,
"data_files_size_in_mb": 100,
"video_files_size_in_mb": 200,
"fps": 30,
"splits": {
"train": "0:15"
},
"data_path": "data/chunk-{chunk_index:03d}/file-{file_index:03d}.parquet",
"video_path": "videos/{video_key}/chunk-{chunk_index:03d}/file-{file_index:03d}.mp4",
"features": {
"timestamp": {
"dtype": "float32",
"shape": [
1
],
"names": null
},
"frame_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"episode_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"task_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"observation.images.cam_wrist": {
"dtype": "video",
"names": [
"height",
"width",
"channels"
],
"shape": [
480,
848,
3
],
"info": {
"video.height": 480,
"video.width": 848,
"video.channels": 3,
"video.codec": "libx264",
"video.pix_fmt": "yuv420p"
}
},
"observation.images.cam_top": {
"dtype": "video",
"names": [
"height",
"width",
"channels"
],
"shape": [
720,
1280,
3
],
"info": {
"video.height": 720,
"video.width": 1280,
"video.channels": 3,
"video.codec": "libx264",
"video.pix_fmt": "yuv420p"
}
},
"observation.state": {
"dtype": "float32",
"names": [
"joint1",
"joint2",
"joint3",
"joint4",
"joint5",
"joint6",
"rh_r1_joint"
],
"shape": [
7
]
},
"action": {
"dtype": "float32",
"names": [
"joint1",
"joint2",
"joint3",
"joint4",
"joint5",
"joint6",
"rh_r1_joint"
],
"shape": [
7
]
}
},
"total_videos": 30
}
Citation
BibTeX:
[More Information Needed]
- Downloads last month
- 70