Skip to content

Commit 26686ad

Browse files
committed
fix ci
1 parent 49eada3 commit 26686ad

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

paddlenlp/experimental/transformers/fused_transformer_layers.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -356,7 +356,6 @@ def __init__(self, config: FusedMultiTransformerConfig):
356356
is_bias=True,
357357
dtype=self._norm_weight_dtype,
358358
)
359-
self.init_weight_shape(config)
360359

361360
qkv_weight = self.create_parameter(
362361
shape=self.qkv_weight_shape,
@@ -530,6 +529,9 @@ def __init__(self, config: FusedMultiTransformerConfig):
530529

531530
self.linear = fused_linear
532531

532+
def __post_init__(self):
533+
self.init_weight_shape(self.config)
534+
533535
def get_attr(self, attrs, idx):
534536
if isinstance(attrs, (list, tuple)):
535537
assert (

0 commit comments

Comments
 (0)