From ce6a26deeb042c35549ed1eb050370613defac2d Mon Sep 17 00:00:00 2001 From: Francesco Capuano <74058581+fracapuano@users.noreply.github.com> Date: Sat, 14 Jun 2025 19:25:50 +0200 Subject: [PATCH] Fixing `PI0` Policy (#1297) --- lerobot/common/policies/pi0/paligemma_with_expert.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/lerobot/common/policies/pi0/paligemma_with_expert.py b/lerobot/common/policies/pi0/paligemma_with_expert.py index 49c844c7b..fb5077fb2 100644 --- a/lerobot/common/policies/pi0/paligemma_with_expert.py +++ b/lerobot/common/policies/pi0/paligemma_with_expert.py @@ -223,7 +223,7 @@ class PaliGemmaWithExpertModel(PreTrainedModel): return self.paligemma.model.get_image_features(image) def embed_language_tokens(self, tokens: torch.Tensor): - return self.paligemma.language_model.model.embed_tokens(tokens) + return self.paligemma.language_model.embed_tokens(tokens) # TODO: break down this huge forward into modules or functions def forward( @@ -235,7 +235,7 @@ class PaliGemmaWithExpertModel(PreTrainedModel): use_cache: Optional[bool] = None, fill_kv_cache: Optional[bool] = None, ): - models = [self.paligemma.language_model.model, self.gemma_expert.model] + models = [self.paligemma.language_model, self.gemma_expert.model] for hidden_states in inputs_embeds: # TODO this is very inefficient