Skip to content

Commit 858a686

Browse files
committed
Add taskflow loading finetune model
1 parent 70cf0d8 commit 858a686

File tree

1 file changed

+68
-4
lines changed

1 file changed

+68
-4
lines changed

paddlenlp/taskflow/feature_extraction.py

Lines changed: 68 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -67,13 +67,77 @@ class MultimodalFeatureExtractionTask(Task):
6767
kwargs (dict, optional): Additional keyword arguments passed along to the specific task.
6868
"""
6969

70+
resource_files_names = {
71+
"model_state": "model_state.pdparams",
72+
"config": "config.json",
73+
"vocab_file": "vocab.txt",
74+
"preprocessor_config": "preprocessor_config.json",
75+
"special_tokens_map": "special_tokens_map.json",
76+
"tokenizer_config": "tokenizer_config.json",
77+
}
78+
resource_files_urls = {
79+
"PaddlePaddle/ernie_vil-2.0-base-zh": {
80+
"model_state": [
81+
"https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/model_state.pdparams",
82+
"38d8c8e01f74ba881e87d9a3f669e5ae",
83+
],
84+
"config": [
85+
"https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/config.json",
86+
"caf929b450d5638e8df2a95c936519e7",
87+
],
88+
"vocab_file": [
89+
"https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/vocab.txt",
90+
"1c1c1f4fd93c5bed3b4eebec4de976a8",
91+
],
92+
"preprocessor_config": [
93+
"https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/preprocessor_config.json",
94+
"9a2e8da9f41896fedb86756b79355ee2",
95+
],
96+
"special_tokens_map": [
97+
"https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/special_tokens_map.json",
98+
"8b3fb1023167bb4ab9d70708eb05f6ec",
99+
],
100+
"tokenizer_config": [
101+
"https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/tokenizer_config.json",
102+
"da5385c23c8f522d33fc3aac829e4375",
103+
],
104+
},
105+
"OFA-Sys/chinese-clip-vit-base-patch16": {
106+
"model_state": [
107+
"https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/model_state.pdparams",
108+
"d594c94833b8cfeffc4f986712b3ef79",
109+
],
110+
"config": [
111+
"https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/config.json",
112+
"3611b5c34ad69dcf91e3c1d03b01a93a",
113+
],
114+
"vocab_file": [
115+
"https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/vocab.txt",
116+
"3b5b76c4aef48ecf8cb3abaafe960f09",
117+
],
118+
"preprocessor_config": [
119+
"https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/preprocessor_config.json",
120+
"ba1fb66c75b18b3c9580ea5120e01ced",
121+
],
122+
"special_tokens_map": [
123+
"https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/special_tokens_map.json",
124+
"8b3fb1023167bb4ab9d70708eb05f6ec",
125+
],
126+
"tokenizer_config": [
127+
"https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/tokenizer_config.json",
128+
"573ba0466e15cdb5bd423ff7010735ce",
129+
],
130+
},
131+
}
132+
70133
def __init__(self, task, model, batch_size=1, _static_mode=True, **kwargs):
71134
super().__init__(task=task, model=model, **kwargs)
72135
self._seed = None
73136
# we do not use batch
74137
self.mode = "text"
75138
self._batch_size = batch_size
76-
self._construct_tokenizer(model_name=model)
139+
self._check_task_files()
140+
self._construct_tokenizer()
77141
self._static_mode = _static_mode
78142
self._config_map = {}
79143
self.predictor_map = {}
@@ -90,14 +154,14 @@ def _construct_model(self, model):
90154
"""
91155
Construct the inference model for the predictor.
92156
"""
93-
self._model = AutoModel.from_pretrained(model)
157+
self._model = AutoModel.from_pretrained(self._task_path)
94158
self._model.eval()
95159

96-
def _construct_tokenizer(self, model_name):
160+
def _construct_tokenizer(self):
97161
"""
98162
Construct the tokenizer for the predictor.
99163
"""
100-
self._processor = AutoProcessor.from_pretrained(model_name)
164+
self._processor = AutoProcessor.from_pretrained(self.model)
101165

102166
def _batchify(self, data, batch_size):
103167
"""

0 commit comments

Comments
 (0)