diff --git a/src/diffusers/models/attention.py b/src/diffusers/models/attention.py index b204770e6d37..74a9d9efc6bd 100644 --- a/src/diffusers/models/attention.py +++ b/src/diffusers/models/attention.py @@ -376,7 +376,7 @@ def __init__( "layer_norm", ) - elif norm_type in ["ada_norm_zero", "ada_norm", "layer_norm", "ada_norm_continuous"]: + elif norm_type in ["ada_norm_zero", "ada_norm", "layer_norm"]: self.norm3 = nn.LayerNorm(dim, norm_eps, norm_elementwise_affine) elif norm_type == "layer_norm_i2vgen": self.norm3 = None