From 1713c8cb37073ee9e5276fa8d55f314cc11906b2 Mon Sep 17 00:00:00 2001 From: Aryan Date: Mon, 29 Jul 2024 15:35:00 +0200 Subject: [PATCH] remove ada_norm_continuous from norm3 list --- src/diffusers/models/attention.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/diffusers/models/attention.py b/src/diffusers/models/attention.py index f6969470c36e..6c708cd1d09a 100644 --- a/src/diffusers/models/attention.py +++ b/src/diffusers/models/attention.py @@ -376,7 +376,7 @@ def __init__( "layer_norm", ) - elif norm_type in ["ada_norm_zero", "ada_norm", "layer_norm", "ada_norm_continuous"]: + elif norm_type in ["ada_norm_zero", "ada_norm", "layer_norm"]: self.norm3 = nn.LayerNorm(dim, norm_eps, norm_elementwise_affine) elif norm_type == "layer_norm_i2vgen": self.norm3 = None