From ae22e7f52aefb01bb5670e6c70eae8150cead226 Mon Sep 17 00:00:00 2001 From: Mustafa Eyceoz Date: Wed, 11 Feb 2026 19:53:31 +0000 Subject: [PATCH] fix _no_split_modules subscript error for transformers v5 --- src/instructlab/training/accelerator.py | 2 +- tests/unit/test_accelerator.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/src/instructlab/training/accelerator.py b/src/instructlab/training/accelerator.py index 2d88df2c..4bfc6603 100644 --- a/src/instructlab/training/accelerator.py +++ b/src/instructlab/training/accelerator.py @@ -135,7 +135,7 @@ def __getattr__(self, name): def get_fsdp_config(self): is_lora = self.model.lora_config is not None - block_name = self.model._no_split_modules[0] + block_name = next(iter(self.model._no_split_modules)) wrap_policy = None if is_lora > 0: diff --git a/tests/unit/test_accelerator.py b/tests/unit/test_accelerator.py index c208580d..21117537 100644 --- a/tests/unit/test_accelerator.py +++ b/tests/unit/test_accelerator.py @@ -18,7 +18,7 @@ def mock_model(): model = MagicMock(spec=Model) model.model = MagicMock() model.lora_config = None - model._no_split_modules = ["LlamaDecoderLayer"] + model._no_split_modules = {"LlamaDecoderLayer"} # Add children method to model model.children = MagicMock(return_value=[]) model.model.children = MagicMock(return_value=[])