Skip to content
This repository was archived by the owner on Oct 25, 2024. It is now read-only.

Commit c02dd7b

Browse files
authored
upgrade huggingface use_auth_token to token (#943)
1 parent f9df4c2 commit c02dd7b

File tree

2 files changed

+21
-12
lines changed

2 files changed

+21
-12
lines changed

intel_extension_for_transformers/llm/finetuning/finetuning.py

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -144,7 +144,7 @@ def load_dataset(self, data_args, model_args, training_args):
144144
data_args.dataset_name,
145145
data_args.dataset_config_name,
146146
cache_dir=model_args.cache_dir,
147-
use_auth_token=True if model_args.use_auth_token else None,
147+
token=model_args.token,
148148
streaming=data_args.streaming,
149149
)
150150

@@ -154,15 +154,15 @@ def load_dataset(self, data_args, model_args, training_args):
154154
data_args.dataset_config_name,
155155
split=f"train[:{data_args.validation_split_percentage}%]",
156156
cache_dir=model_args.cache_dir,
157-
use_auth_token=True if model_args.use_auth_token else None,
157+
token=model_args.token,
158158
streaming=data_args.streaming,
159159
)
160160
raw_datasets["train"] = load_dataset(
161161
data_args.dataset_name,
162162
data_args.dataset_config_name,
163163
split=f"train[{data_args.validation_split_percentage}%:]",
164164
cache_dir=model_args.cache_dir,
165-
use_auth_token=True if model_args.use_auth_token else None,
165+
token=model_args.token,
166166
streaming=data_args.streaming,
167167
)
168168
else:
@@ -184,7 +184,7 @@ def load_dataset(self, data_args, model_args, training_args):
184184
extension,
185185
data_files=data_files,
186186
cache_dir=model_args.cache_dir,
187-
use_auth_token=True if model_args.use_auth_token else None,
187+
token=model_args.token,
188188
**dataset_args,
189189
)
190190

@@ -195,15 +195,15 @@ def load_dataset(self, data_args, model_args, training_args):
195195
data_files=data_files,
196196
split=f"train[:{data_args.validation_split_percentage}%]",
197197
cache_dir=model_args.cache_dir,
198-
use_auth_token=True if model_args.use_auth_token else None,
198+
token=model_args.token,
199199
**dataset_args,
200200
)
201201
raw_datasets["train"] = load_dataset(
202202
extension,
203203
data_files=data_files,
204204
split=f"train[{data_args.validation_split_percentage}%:]",
205205
cache_dir=model_args.cache_dir,
206-
use_auth_token=True if model_args.use_auth_token else None,
206+
token=model_args.token,
207207
**dataset_args,
208208
)
209209
return raw_datasets
@@ -212,7 +212,7 @@ def load_model_config(self, model_args):
212212
config_kwargs = {
213213
"cache_dir": model_args.cache_dir,
214214
"revision": model_args.model_revision,
215-
"use_auth_token": True if model_args.use_auth_token else None,
215+
"token": model_args.token,
216216
"trust_remote_code": True if model_args.trust_remote_code else None,
217217
}
218218
if model_args.config_name:
@@ -230,7 +230,7 @@ def load_tokenizer(self, model_args):
230230
"cache_dir": model_args.cache_dir,
231231
"use_fast": model_args.use_fast_tokenizer,
232232
"revision": model_args.model_revision,
233-
"use_auth_token": True if model_args.use_auth_token else None,
233+
"token": model_args.token,
234234
"trust_remote_code": model_args.trust_remote_code,
235235
}
236236
if model_args.tokenizer_name:
@@ -378,7 +378,7 @@ def finetune_clm(self, model_args, data_args, training_args, finetune_args, conf
378378
device_map=self.device_map,
379379
quantization_config=self.bitsandbytes_quant_config,
380380
revision=model_args.model_revision,
381-
use_auth_token=True if model_args.use_auth_token else None,
381+
token=model_args.token,
382382
trust_remote_code=True if model_args.trust_remote_code else None,
383383
torch_dtype=model_dtype,
384384
low_cpu_mem_usage=low_cpu_mem_usage,
@@ -807,7 +807,7 @@ def preprocess_logits_for_metrics(logits, labels):
807807
device_map=self.device_map,
808808
quantization_config=self.bitsandbytes_quant_config,
809809
revision=model_args.model_revision,
810-
use_auth_token=True if model_args.use_auth_token else None,
810+
token=model_args.token,
811811
torch_dtype=model_dtype,
812812
load_in_4bit=self.load_in_4bit,
813813
load_in_8bit=self.load_in_8bit,

intel_extension_for_transformers/neural_chat/config.py

Lines changed: 11 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -92,11 +92,20 @@ class ModelArguments:
9292
"help": "The specific model version to use (can be a branch name, tag name or commit id)."
9393
},
9494
)
95+
token: str = field(
96+
default=None,
97+
metadata={
98+
"help": (
99+
"The token to use as HTTP bearer authorization for remote files. If not specified, will use the token "
100+
"generated when running `huggingface-cli login` (stored in `~/.huggingface`)."
101+
)
102+
},
103+
)
95104
use_auth_token: bool = field(
96105
default=False,
97106
metadata={
98-
"help": "Will use the token generated when running `transformers-cli login` (necessary to use this script "
99-
"with private models)."
107+
"help": "The `use_auth_token` argument is deprecated and will be removed in v4.34."
108+
"Please use `token` instead."
100109
},
101110
)
102111
trust_remote_code: bool = field(

0 commit comments

Comments
 (0)