@@ -67,13 +67,77 @@ class MultimodalFeatureExtractionTask(Task):
67
67
kwargs (dict, optional): Additional keyword arguments passed along to the specific task.
68
68
"""
69
69
70
+ resource_files_names = {
71
+ "model_state" : "model_state.pdparams" ,
72
+ "config" : "config.json" ,
73
+ "vocab_file" : "vocab.txt" ,
74
+ "preprocessor_config" : "preprocessor_config.json" ,
75
+ "special_tokens_map" : "special_tokens_map.json" ,
76
+ "tokenizer_config" : "tokenizer_config.json" ,
77
+ }
78
+ resource_files_urls = {
79
+ "PaddlePaddle/ernie_vil-2.0-base-zh" : {
80
+ "model_state" : [
81
+ "https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/model_state.pdparams" ,
82
+ "38d8c8e01f74ba881e87d9a3f669e5ae" ,
83
+ ],
84
+ "config" : [
85
+ "https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/config.json" ,
86
+ "caf929b450d5638e8df2a95c936519e7" ,
87
+ ],
88
+ "vocab_file" : [
89
+ "https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/vocab.txt" ,
90
+ "1c1c1f4fd93c5bed3b4eebec4de976a8" ,
91
+ ],
92
+ "preprocessor_config" : [
93
+ "https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/preprocessor_config.json" ,
94
+ "9a2e8da9f41896fedb86756b79355ee2" ,
95
+ ],
96
+ "special_tokens_map" : [
97
+ "https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/special_tokens_map.json" ,
98
+ "8b3fb1023167bb4ab9d70708eb05f6ec" ,
99
+ ],
100
+ "tokenizer_config" : [
101
+ "https://paddlenlp.bj.bcebos.com/models/community/PaddlePaddle/ernie_vil-2.0-base-zh/tokenizer_config.json" ,
102
+ "da5385c23c8f522d33fc3aac829e4375" ,
103
+ ],
104
+ },
105
+ "OFA-Sys/chinese-clip-vit-base-patch16" : {
106
+ "model_state" : [
107
+ "https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/model_state.pdparams" ,
108
+ "d594c94833b8cfeffc4f986712b3ef79" ,
109
+ ],
110
+ "config" : [
111
+ "https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/config.json" ,
112
+ "3611b5c34ad69dcf91e3c1d03b01a93a" ,
113
+ ],
114
+ "vocab_file" : [
115
+ "https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/vocab.txt" ,
116
+ "3b5b76c4aef48ecf8cb3abaafe960f09" ,
117
+ ],
118
+ "preprocessor_config" : [
119
+ "https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/preprocessor_config.json" ,
120
+ "ba1fb66c75b18b3c9580ea5120e01ced" ,
121
+ ],
122
+ "special_tokens_map" : [
123
+ "https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/special_tokens_map.json" ,
124
+ "8b3fb1023167bb4ab9d70708eb05f6ec" ,
125
+ ],
126
+ "tokenizer_config" : [
127
+ "https://paddlenlp.bj.bcebos.com/models/community/OFA-Sys/chinese-clip-vit-base-patch16/tokenizer_config.json" ,
128
+ "573ba0466e15cdb5bd423ff7010735ce" ,
129
+ ],
130
+ },
131
+ }
132
+
70
133
def __init__ (self , task , model , batch_size = 1 , _static_mode = True , ** kwargs ):
71
134
super ().__init__ (task = task , model = model , ** kwargs )
72
135
self ._seed = None
73
136
# we do not use batch
74
137
self .mode = "text"
75
138
self ._batch_size = batch_size
76
- self ._construct_tokenizer (model_name = model )
139
+ self ._check_task_files ()
140
+ self ._construct_tokenizer ()
77
141
self ._static_mode = _static_mode
78
142
self ._config_map = {}
79
143
self .predictor_map = {}
@@ -90,14 +154,14 @@ def _construct_model(self, model):
90
154
"""
91
155
Construct the inference model for the predictor.
92
156
"""
93
- self ._model = AutoModel .from_pretrained (model )
157
+ self ._model = AutoModel .from_pretrained (self . _task_path )
94
158
self ._model .eval ()
95
159
96
- def _construct_tokenizer (self , model_name ):
160
+ def _construct_tokenizer (self ):
97
161
"""
98
162
Construct the tokenizer for the predictor.
99
163
"""
100
- self ._processor = AutoProcessor .from_pretrained (model_name )
164
+ self ._processor = AutoProcessor .from_pretrained (self . model )
101
165
102
166
def _batchify (self , data , batch_size ):
103
167
"""
0 commit comments