From 0f66983c71b8f3edc7cf326f5bdfea38b672f281 Mon Sep 17 00:00:00 2001 From: tastelikefeet Date: Mon, 2 Mar 2026 13:59:45 +0800 Subject: [PATCH] fix TEGroupedLinear --- src/twinkle/model/megatron/tuners/lora.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/src/twinkle/model/megatron/tuners/lora.py b/src/twinkle/model/megatron/tuners/lora.py index 60c1e7d7..b80c906d 100644 --- a/src/twinkle/model/megatron/tuners/lora.py +++ b/src/twinkle/model/megatron/tuners/lora.py @@ -136,8 +136,6 @@ def update_layer(self, adapter_name: str, r: int, *, lora_alpha: int, lora_dropo 'config': self.config, 'is_expert': self.is_expert, } - if exists('megatron_core>=0.13'): - kwargs['tp_group'] = self.base_layer.tp_group if isinstance(self.base_layer, TopKRouter): # Router layer - no parallelism needed