From 093a85f946a626d05af841b7f84449b686e74de9 Mon Sep 17 00:00:00 2001 From: Michel Aractingi Date: Fri, 12 Dec 2025 16:36:18 +0000 Subject: [PATCH] nit --- src/lerobot/optim/optimizers.py | 2 +- src/lerobot/policies/smolvla/configuration_smolvla.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/src/lerobot/optim/optimizers.py b/src/lerobot/optim/optimizers.py index dde118f39..66d819ac6 100644 --- a/src/lerobot/optim/optimizers.py +++ b/src/lerobot/optim/optimizers.py @@ -81,7 +81,7 @@ class AdamWConfig(OptimizerConfig): eps: float = 1e-8 weight_decay: float = 1e-2 grad_clip_norm: float = 10.0 - fused: bool = False + fused: bool = False def build(self, params: dict) -> torch.optim.Optimizer: kwargs = asdict(self) diff --git a/src/lerobot/policies/smolvla/configuration_smolvla.py b/src/lerobot/policies/smolvla/configuration_smolvla.py index eea7299a1..d27f1a5a5 100644 --- a/src/lerobot/policies/smolvla/configuration_smolvla.py +++ b/src/lerobot/policies/smolvla/configuration_smolvla.py @@ -79,7 +79,7 @@ class SmolVLAConfig(PreTrainedConfig): optimizer_eps: float = 1e-8 optimizer_weight_decay: float = 1e-10 optimizer_grad_clip_norm: float = 10 - optimizer_fused: bool = False + optimizer_fused: bool = False scheduler_warmup_steps: int = 1_000 scheduler_decay_steps: int = 30_000