From 9b211a45d6c237ddd1c249bf91318a681a8ac326 Mon Sep 17 00:00:00 2001 From: Pepijn Date: Mon, 30 Mar 2026 15:31:18 +0200 Subject: [PATCH] fix: disable thinking mode in Qwen35VL single-episode fallback path The single-episode `segment_skills` method was missing `enable_thinking=False` in `apply_chat_template`, causing the model to output reasoning traces instead of JSON when the batch path fails and falls back to per-episode processing. Made-with: Cursor --- .../data_processing/data_annotations/vlm_annotations.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/src/lerobot/data_processing/data_annotations/vlm_annotations.py b/src/lerobot/data_processing/data_annotations/vlm_annotations.py index 5ec03ffc3..237c8af51 100644 --- a/src/lerobot/data_processing/data_annotations/vlm_annotations.py +++ b/src/lerobot/data_processing/data_annotations/vlm_annotations.py @@ -495,7 +495,9 @@ class Qwen35VL(BaseVLM): }, ] - text = self.processor.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) + text = self.processor.apply_chat_template( + messages, tokenize=False, add_generation_prompt=True, enable_thinking=False + ) image_inputs, video_inputs = self.process_vision_info(messages) inputs = self.processor( text=[text],