From ce7d33a9323f3b3ec500a0afff4129e612941527 Mon Sep 17 00:00:00 2001 From: Dillon Laird Date: Thu, 30 Nov 2023 10:44:46 -0800 Subject: [PATCH] include last attention layer in feature output --- models/fastvit.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/models/fastvit.py b/models/fastvit.py index 6f0363e..6db422f 100644 --- a/models/fastvit.py +++ b/models/fastvit.py @@ -828,7 +828,7 @@ def __init__( # For segmentation and detection, extract intermediate output if self.fork_feat: # add a norm layer for each output - self.out_indices = [0, 2, 4, 6] + self.out_indices = [0, 2, 4, 7] for i_emb, i_layer in enumerate(self.out_indices): if i_emb == 0 and os.environ.get("FORK_LAST3", None): """For RetinaNet, `start_level=1`. The first norm layer will not used.