@@ -144,7 +144,7 @@ def load_dataset(self, data_args, model_args, training_args):
144144 data_args .dataset_name ,
145145 data_args .dataset_config_name ,
146146 cache_dir = model_args .cache_dir ,
147- use_auth_token = True if model_args .use_auth_token else None ,
147+ token = model_args .token ,
148148 streaming = data_args .streaming ,
149149 )
150150
@@ -154,15 +154,15 @@ def load_dataset(self, data_args, model_args, training_args):
154154 data_args .dataset_config_name ,
155155 split = f"train[:{ data_args .validation_split_percentage } %]" ,
156156 cache_dir = model_args .cache_dir ,
157- use_auth_token = True if model_args .use_auth_token else None ,
157+ token = model_args .token ,
158158 streaming = data_args .streaming ,
159159 )
160160 raw_datasets ["train" ] = load_dataset (
161161 data_args .dataset_name ,
162162 data_args .dataset_config_name ,
163163 split = f"train[{ data_args .validation_split_percentage } %:]" ,
164164 cache_dir = model_args .cache_dir ,
165- use_auth_token = True if model_args .use_auth_token else None ,
165+ token = model_args .token ,
166166 streaming = data_args .streaming ,
167167 )
168168 else :
@@ -184,7 +184,7 @@ def load_dataset(self, data_args, model_args, training_args):
184184 extension ,
185185 data_files = data_files ,
186186 cache_dir = model_args .cache_dir ,
187- use_auth_token = True if model_args .use_auth_token else None ,
187+ token = model_args .token ,
188188 ** dataset_args ,
189189 )
190190
@@ -195,15 +195,15 @@ def load_dataset(self, data_args, model_args, training_args):
195195 data_files = data_files ,
196196 split = f"train[:{ data_args .validation_split_percentage } %]" ,
197197 cache_dir = model_args .cache_dir ,
198- use_auth_token = True if model_args .use_auth_token else None ,
198+ token = model_args .token ,
199199 ** dataset_args ,
200200 )
201201 raw_datasets ["train" ] = load_dataset (
202202 extension ,
203203 data_files = data_files ,
204204 split = f"train[{ data_args .validation_split_percentage } %:]" ,
205205 cache_dir = model_args .cache_dir ,
206- use_auth_token = True if model_args .use_auth_token else None ,
206+ token = model_args .token ,
207207 ** dataset_args ,
208208 )
209209 return raw_datasets
@@ -212,7 +212,7 @@ def load_model_config(self, model_args):
212212 config_kwargs = {
213213 "cache_dir" : model_args .cache_dir ,
214214 "revision" : model_args .model_revision ,
215- "use_auth_token " : True if model_args .use_auth_token else None ,
215+ "token " : model_args .token ,
216216 "trust_remote_code" : True if model_args .trust_remote_code else None ,
217217 }
218218 if model_args .config_name :
@@ -230,7 +230,7 @@ def load_tokenizer(self, model_args):
230230 "cache_dir" : model_args .cache_dir ,
231231 "use_fast" : model_args .use_fast_tokenizer ,
232232 "revision" : model_args .model_revision ,
233- "use_auth_token " : True if model_args .use_auth_token else None ,
233+ "token " : model_args .token ,
234234 "trust_remote_code" : model_args .trust_remote_code ,
235235 }
236236 if model_args .tokenizer_name :
@@ -378,7 +378,7 @@ def finetune_clm(self, model_args, data_args, training_args, finetune_args, conf
378378 device_map = self .device_map ,
379379 quantization_config = self .bitsandbytes_quant_config ,
380380 revision = model_args .model_revision ,
381- use_auth_token = True if model_args .use_auth_token else None ,
381+ token = model_args .token ,
382382 trust_remote_code = True if model_args .trust_remote_code else None ,
383383 torch_dtype = model_dtype ,
384384 low_cpu_mem_usage = low_cpu_mem_usage ,
@@ -807,7 +807,7 @@ def preprocess_logits_for_metrics(logits, labels):
807807 device_map = self .device_map ,
808808 quantization_config = self .bitsandbytes_quant_config ,
809809 revision = model_args .model_revision ,
810- use_auth_token = True if model_args .use_auth_token else None ,
810+ token = model_args .token ,
811811 torch_dtype = model_dtype ,
812812 load_in_4bit = self .load_in_4bit ,
813813 load_in_8bit = self .load_in_8bit ,
0 commit comments