Skip to content
This repository was archived by the owner on Oct 25, 2024. It is now read-only.

Conversation

@lkk12014402
Copy link
Contributor

Type of Change

API changed for supporting full finetuning

Description

add if branch for supporting full parameters finetuning

@lkk12014402
Copy link
Contributor Author

lkk12014402 commented Nov 30, 2023

adding ut WIP

@hshen14 hshen14 changed the title support full parameters finetuning [NeuralChat] support full parameters finetuning Dec 3, 2023
@lkk12014402
Copy link
Contributor Author

lkk12014402 commented Dec 5, 2023

add ut done.

finish offline tests:

  1. lora/full_finetuning on gaudi/cpu/gpu
  2. qlora on gpu
@hshen14 hshen14 merged commit 2b5411f into main Dec 6, 2023
@hshen14 hshen14 deleted the support_full_finetune branch December 6, 2023 03:18
delock pushed a commit to delock/intel-extension-for-transformers that referenced this pull request Dec 16, 2023
* fix init_vector * adjust spmm_amx_bf16 ut * make cpplint happy
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

4 participants