mirror of
https://github.com/huggingface/lerobot.git
synced 2026-05-14 16:19:45 +00:00
Compare commits
10 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| bcc13f1d90 | |||
| 76f25f6afd | |||
| ce23681d4b | |||
| e195f8d287 | |||
| bbcffc4999 | |||
| 20333abc72 | |||
| 00a4e6bfb3 | |||
| a19bd6e84d | |||
| 550866a3c5 | |||
| 3ec4e4ce37 |
@@ -19,6 +19,7 @@ import shutil
|
|||||||
import tempfile
|
import tempfile
|
||||||
from collections.abc import Callable
|
from collections.abc import Callable
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
from typing import Any
|
||||||
|
|
||||||
import datasets
|
import datasets
|
||||||
import numpy as np
|
import numpy as np
|
||||||
@@ -31,6 +32,8 @@ import torch
|
|||||||
import torch.utils
|
import torch.utils
|
||||||
from huggingface_hub import HfApi, snapshot_download
|
from huggingface_hub import HfApi, snapshot_download
|
||||||
from huggingface_hub.errors import RevisionNotFoundError
|
from huggingface_hub.errors import RevisionNotFoundError
|
||||||
|
from PIL import Image as PILImage
|
||||||
|
from torchvision import transforms
|
||||||
|
|
||||||
from lerobot.datasets.compute_stats import aggregate_stats, compute_episode_stats
|
from lerobot.datasets.compute_stats import aggregate_stats, compute_episode_stats
|
||||||
from lerobot.datasets.image_writer import AsyncImageWriter, write_image
|
from lerobot.datasets.image_writer import AsyncImageWriter, write_image
|
||||||
@@ -50,11 +53,9 @@ from lerobot.datasets.utils import (
|
|||||||
get_file_size_in_mb,
|
get_file_size_in_mb,
|
||||||
get_hf_features_from_features,
|
get_hf_features_from_features,
|
||||||
get_safe_version,
|
get_safe_version,
|
||||||
hf_transform_to_torch,
|
|
||||||
is_valid_version,
|
is_valid_version,
|
||||||
load_episodes,
|
load_episodes,
|
||||||
load_info,
|
load_info,
|
||||||
load_nested_dataset,
|
|
||||||
load_stats,
|
load_stats,
|
||||||
load_tasks,
|
load_tasks,
|
||||||
update_chunk_file_indices,
|
update_chunk_file_indices,
|
||||||
@@ -79,6 +80,51 @@ from lerobot.utils.constants import HF_LEROBOT_HOME
|
|||||||
CODEBASE_VERSION = "v3.0"
|
CODEBASE_VERSION = "v3.0"
|
||||||
|
|
||||||
|
|
||||||
|
def hf_transform_to_torch(items_dict: dict[str, list[Any]]) -> dict[str, list[torch.Tensor | str]]:
|
||||||
|
"""
|
||||||
|
Converts a batch from a Hugging Face dataset to torch tensors.
|
||||||
|
"""
|
||||||
|
|
||||||
|
# Create a single ToTensor transform instance to reuse
|
||||||
|
to_tensor = transforms.ToTensor()
|
||||||
|
|
||||||
|
for key in items_dict:
|
||||||
|
items_list = items_dict[key]
|
||||||
|
|
||||||
|
# Check if the list is non-empty
|
||||||
|
if not items_list:
|
||||||
|
continue
|
||||||
|
|
||||||
|
first_item = items_list[0]
|
||||||
|
|
||||||
|
if isinstance(first_item, PILImage.Image):
|
||||||
|
# This is the (slow) CPU-bound part.
|
||||||
|
# We convert every image in the batch list to a tensor.
|
||||||
|
items_dict[key] = [to_tensor(img) for img in items_list]
|
||||||
|
|
||||||
|
elif isinstance(first_item, (str, bytes)):
|
||||||
|
# List of strings (e.g., 'task'), do nothing
|
||||||
|
pass
|
||||||
|
|
||||||
|
elif first_item is None:
|
||||||
|
# List of Nones, do nothing
|
||||||
|
pass
|
||||||
|
|
||||||
|
else:
|
||||||
|
# List of other things (int, float, list, np.ndarray)
|
||||||
|
try:
|
||||||
|
# Convert each item in the list to a tensor
|
||||||
|
items_dict[key] = [torch.tensor(item) for item in items_list]
|
||||||
|
except Exception as e:
|
||||||
|
# This catch is what was missing from the original v3.0 code
|
||||||
|
print(
|
||||||
|
f"Error converting batch['{key}'] to tensor. First item: {first_item}, Type: {type(first_item)}"
|
||||||
|
)
|
||||||
|
raise e
|
||||||
|
|
||||||
|
return items_dict
|
||||||
|
|
||||||
|
|
||||||
class LeRobotDatasetMetadata:
|
class LeRobotDatasetMetadata:
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
@@ -693,6 +739,9 @@ class LeRobotDataset(torch.utils.data.Dataset):
|
|||||||
self.repo_id, self.root, self.revision, force_cache_sync=force_cache_sync
|
self.repo_id, self.root, self.revision, force_cache_sync=force_cache_sync
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# Pre-load episodes metadata into memory to avoid file I/O in __getitem__
|
||||||
|
self.episodes_metadata_list = [ep for ep in self.meta.episodes]
|
||||||
|
|
||||||
# Track dataset state for efficient incremental writing
|
# Track dataset state for efficient incremental writing
|
||||||
self._lazy_loading = False
|
self._lazy_loading = False
|
||||||
self._recorded_frames = self.meta.total_frames
|
self._recorded_frames = self.meta.total_frames
|
||||||
@@ -829,8 +878,36 @@ class LeRobotDataset(torch.utils.data.Dataset):
|
|||||||
|
|
||||||
def load_hf_dataset(self) -> datasets.Dataset:
|
def load_hf_dataset(self) -> datasets.Dataset:
|
||||||
"""hf_dataset contains all the observations, states, actions, rewards, etc."""
|
"""hf_dataset contains all the observations, states, actions, rewards, etc."""
|
||||||
|
|
||||||
features = get_hf_features_from_features(self.features)
|
features = get_hf_features_from_features(self.features)
|
||||||
hf_dataset = load_nested_dataset(self.root / "data", features=features)
|
|
||||||
|
if self.episodes is not None:
|
||||||
|
# Path for episode-specific loading (e.g., visualization)
|
||||||
|
fpaths = set()
|
||||||
|
for ep_idx in self.episodes:
|
||||||
|
ep_meta = self.episodes_metadata_list[ep_idx]
|
||||||
|
chunk_idx = ep_meta["data/chunk_index"]
|
||||||
|
file_idx = ep_meta["data/file_index"]
|
||||||
|
fpath_str = self.meta.data_path.format(chunk_index=chunk_idx, file_index=file_idx)
|
||||||
|
fpaths.add(str(self.root / fpath_str))
|
||||||
|
|
||||||
|
data_files = sorted(list(fpaths))
|
||||||
|
|
||||||
|
hf_dataset = datasets.load_dataset(
|
||||||
|
"parquet", data_files=data_files, features=features, split="train"
|
||||||
|
)
|
||||||
|
|
||||||
|
requested_episodes_set = set(self.episodes)
|
||||||
|
hf_dataset = hf_dataset.filter(
|
||||||
|
lambda x: x["episode_index"] in requested_episodes_set, batched=True, batch_size=1000
|
||||||
|
)
|
||||||
|
|
||||||
|
else:
|
||||||
|
# THIS IS THE FAST PATH FOR TRAINING (self.episodes is None)
|
||||||
|
# Use `data_dir` to trigger the v2.1-style efficient cache.
|
||||||
|
data_dir = str(self.root / "data")
|
||||||
|
hf_dataset = datasets.load_dataset("parquet", data_dir=data_dir, features=features, split="train")
|
||||||
|
|
||||||
hf_dataset.set_transform(hf_transform_to_torch)
|
hf_dataset.set_transform(hf_transform_to_torch)
|
||||||
return hf_dataset
|
return hf_dataset
|
||||||
|
|
||||||
@@ -909,7 +986,7 @@ class LeRobotDataset(torch.utils.data.Dataset):
|
|||||||
return get_hf_features_from_features(self.features)
|
return get_hf_features_from_features(self.features)
|
||||||
|
|
||||||
def _get_query_indices(self, idx: int, ep_idx: int) -> tuple[dict[str, list[int | bool]]]:
|
def _get_query_indices(self, idx: int, ep_idx: int) -> tuple[dict[str, list[int | bool]]]:
|
||||||
ep = self.meta.episodes[ep_idx]
|
ep = self.episodes_metadata_list[ep_idx]
|
||||||
ep_start = ep["dataset_from_index"]
|
ep_start = ep["dataset_from_index"]
|
||||||
ep_end = ep["dataset_to_index"]
|
ep_end = ep["dataset_to_index"]
|
||||||
query_indices = {
|
query_indices = {
|
||||||
@@ -952,7 +1029,7 @@ class LeRobotDataset(torch.utils.data.Dataset):
|
|||||||
Segmentation Fault. This probably happens because a memory reference to the video loader is created in
|
Segmentation Fault. This probably happens because a memory reference to the video loader is created in
|
||||||
the main process and a subprocess fails to access it.
|
the main process and a subprocess fails to access it.
|
||||||
"""
|
"""
|
||||||
ep = self.meta.episodes[ep_idx]
|
ep = self.episodes_metadata_list[ep_idx]
|
||||||
item = {}
|
item = {}
|
||||||
for vid_key, query_ts in query_timestamps.items():
|
for vid_key, query_ts in query_timestamps.items():
|
||||||
# Episodes are stored sequentially on a single mp4 to reduce the number of files.
|
# Episodes are stored sequentially on a single mp4 to reduce the number of files.
|
||||||
@@ -983,29 +1060,72 @@ class LeRobotDataset(torch.utils.data.Dataset):
|
|||||||
def __getitem__(self, idx) -> dict:
|
def __getitem__(self, idx) -> dict:
|
||||||
# Ensure dataset is loaded when we actually need to read from it
|
# Ensure dataset is loaded when we actually need to read from it
|
||||||
self._ensure_hf_dataset_loaded()
|
self._ensure_hf_dataset_loaded()
|
||||||
item = self.hf_dataset[idx]
|
|
||||||
ep_idx = item["episode_index"].item()
|
|
||||||
|
|
||||||
|
# 1. Get query indices if deltas are needed
|
||||||
query_indices = None
|
query_indices = None
|
||||||
|
padding = {}
|
||||||
if self.delta_indices is not None:
|
if self.delta_indices is not None:
|
||||||
query_indices, padding = self._get_query_indices(idx, ep_idx)
|
# We need the episode index *first* to get boundaries.
|
||||||
query_result = self._query_hf_dataset(query_indices)
|
# This is a small read for just one item.
|
||||||
item = {**item, **padding}
|
ep_idx_only = self.hf_dataset[idx : idx + 1]["episode_index"][0].item()
|
||||||
for key, val in query_result.items():
|
query_indices, padding = self._get_query_indices(idx, ep_idx_only)
|
||||||
item[key] = val
|
|
||||||
|
|
||||||
|
# 2. Fetch all data (including images)
|
||||||
|
if query_indices is not None:
|
||||||
|
# --- Delta path ---
|
||||||
|
# Fetch all keys (state, action, AND images) for all deltas
|
||||||
|
item_batch = self.hf_dataset[query_indices["index"]]
|
||||||
|
|
||||||
|
# hf_transform_to_torch (item-level) has already run,
|
||||||
|
# so all values are tensors. We stack them.
|
||||||
|
item = {}
|
||||||
|
for key in item_batch:
|
||||||
|
item[key] = torch.stack(item_batch[key])
|
||||||
|
|
||||||
|
item.update(padding)
|
||||||
|
|
||||||
|
# Use the "current" item's index/timestamp/ep_idx
|
||||||
|
# (assuming 'index' is the key for the main query)
|
||||||
|
current_idx_in_batch = query_indices["index"].index(idx)
|
||||||
|
item["index"] = item["index"][current_idx_in_batch]
|
||||||
|
item["timestamp"] = item["timestamp"][current_idx_in_batch]
|
||||||
|
item["episode_index"] = item["episode_index"][current_idx_in_batch]
|
||||||
|
item["task_index"] = item["task_index"][current_idx_in_batch]
|
||||||
|
|
||||||
|
ep_idx = item["episode_index"].item()
|
||||||
|
|
||||||
|
else:
|
||||||
|
# --- Single-frame path ---
|
||||||
|
item = self.hf_dataset[idx]
|
||||||
|
ep_idx = item["episode_index"].item()
|
||||||
|
|
||||||
|
# 3. Handle videos (which are always separate)
|
||||||
if len(self.meta.video_keys) > 0:
|
if len(self.meta.video_keys) > 0:
|
||||||
current_ts = item["timestamp"].item()
|
current_ts = (
|
||||||
query_timestamps = self._get_query_timestamps(current_ts, query_indices)
|
item["timestamp"].item()
|
||||||
|
if query_indices is None
|
||||||
|
else item["timestamp"][current_idx_in_batch].item()
|
||||||
|
)
|
||||||
|
|
||||||
|
video_query_indices = query_indices
|
||||||
|
if video_query_indices is None:
|
||||||
|
# If no deltas, create a dummy query_indices for _get_query_timestamps
|
||||||
|
video_query_indices = {key: [idx] for key in self.meta.video_keys}
|
||||||
|
|
||||||
|
query_timestamps = self._get_query_timestamps(current_ts, video_query_indices)
|
||||||
video_frames = self._query_videos(query_timestamps, ep_idx)
|
video_frames = self._query_videos(query_timestamps, ep_idx)
|
||||||
|
|
||||||
|
# video_frames are already stacked tensors (B, C, H, W) or (C, H, W)
|
||||||
item = {**video_frames, **item}
|
item = {**video_frames, **item}
|
||||||
|
|
||||||
|
# 4. Apply image transforms
|
||||||
if self.image_transforms is not None:
|
if self.image_transforms is not None:
|
||||||
image_keys = self.meta.camera_keys
|
image_keys = self.meta.camera_keys
|
||||||
for cam in image_keys:
|
for cam in image_keys:
|
||||||
item[cam] = self.image_transforms(item[cam])
|
if cam in item: # videos or images
|
||||||
|
item[cam] = self.image_transforms(item[cam])
|
||||||
|
|
||||||
# Add task as a string
|
# 5. Add task string
|
||||||
task_idx = item["task_index"].item()
|
task_idx = item["task_index"].item()
|
||||||
item["task"] = self.meta.tasks.iloc[task_idx].name
|
item["task"] = self.meta.tasks.iloc[task_idx].name
|
||||||
return item
|
return item
|
||||||
|
|||||||
@@ -35,7 +35,6 @@ from datasets.table import embed_table_storage
|
|||||||
from huggingface_hub import DatasetCard, DatasetCardData, HfApi
|
from huggingface_hub import DatasetCard, DatasetCardData, HfApi
|
||||||
from huggingface_hub.errors import RevisionNotFoundError
|
from huggingface_hub.errors import RevisionNotFoundError
|
||||||
from PIL import Image as PILImage
|
from PIL import Image as PILImage
|
||||||
from torchvision import transforms
|
|
||||||
|
|
||||||
from lerobot.configs.types import FeatureType, PolicyFeature
|
from lerobot.configs.types import FeatureType, PolicyFeature
|
||||||
from lerobot.datasets.backward_compatibility import (
|
from lerobot.datasets.backward_compatibility import (
|
||||||
@@ -116,10 +115,15 @@ def load_nested_dataset(pq_dir: Path, features: datasets.Features | None = None)
|
|||||||
if len(paths) == 0:
|
if len(paths) == 0:
|
||||||
raise FileNotFoundError(f"Provided directory does not contain any parquet file: {pq_dir}")
|
raise FileNotFoundError(f"Provided directory does not contain any parquet file: {pq_dir}")
|
||||||
|
|
||||||
|
# Convert Path objects to a list of strings
|
||||||
|
file_paths = [str(path) for path in paths]
|
||||||
|
|
||||||
|
# Use datasets.load_dataset to force creation of an efficient cache
|
||||||
|
# This pre-decodes the images and avoids the on-the-fly bottleneck.
|
||||||
# TODO(rcadene): set num_proc to accelerate conversion to pyarrow
|
# TODO(rcadene): set num_proc to accelerate conversion to pyarrow
|
||||||
with SuppressProgressBars():
|
with SuppressProgressBars():
|
||||||
datasets = Dataset.from_parquet([str(path) for path in paths], features=features)
|
dataset = datasets.load_dataset("parquet", data_files=file_paths, features=features, split="train")
|
||||||
return datasets
|
return dataset
|
||||||
|
|
||||||
|
|
||||||
def get_parquet_num_frames(parquet_path: str | Path) -> int:
|
def get_parquet_num_frames(parquet_path: str | Path) -> int:
|
||||||
@@ -394,33 +398,6 @@ def load_image_as_numpy(
|
|||||||
return img_array
|
return img_array
|
||||||
|
|
||||||
|
|
||||||
def hf_transform_to_torch(items_dict: dict[str, list[Any]]) -> dict[str, list[torch.Tensor | str]]:
|
|
||||||
"""Convert a batch from a Hugging Face dataset to torch tensors.
|
|
||||||
|
|
||||||
This transform function converts items from Hugging Face dataset format (pyarrow)
|
|
||||||
to torch tensors. Importantly, images are converted from PIL objects (H, W, C, uint8)
|
|
||||||
to a torch image representation (C, H, W, float32) in the range [0, 1]. Other
|
|
||||||
types are converted to torch.tensor.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
items_dict (dict): A dictionary representing a batch of data from a
|
|
||||||
Hugging Face dataset.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
dict: The batch with items converted to torch tensors.
|
|
||||||
"""
|
|
||||||
for key in items_dict:
|
|
||||||
first_item = items_dict[key][0]
|
|
||||||
if isinstance(first_item, PILImage.Image):
|
|
||||||
to_tensor = transforms.ToTensor()
|
|
||||||
items_dict[key] = [to_tensor(img) for img in items_dict[key]]
|
|
||||||
elif first_item is None:
|
|
||||||
pass
|
|
||||||
else:
|
|
||||||
items_dict[key] = [x if isinstance(x, str) else torch.tensor(x) for x in items_dict[key]]
|
|
||||||
return items_dict
|
|
||||||
|
|
||||||
|
|
||||||
def is_valid_version(version: str) -> bool:
|
def is_valid_version(version: str) -> bool:
|
||||||
"""Check if a string is a valid PEP 440 version.
|
"""Check if a string is a valid PEP 440 version.
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user