Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Wrong LoRA tuning example #185

Open
sofyc opened this issue Sep 12, 2024 · 0 comments
Open

Wrong LoRA tuning example #185

sofyc opened this issue Sep 12, 2024 · 0 comments
Assignees

Comments

@sofyc
Copy link

sofyc commented Sep 12, 2024

Please provide us with the following information:

This issue is for a: (mark with an x)

- [x] bug report -> please search issues before submitting
- [ ] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)

https://github.com/microsoft/Phi-3CookBook/blob/main/code/04.Finetuning/Phi-3-finetune-lora-python.ipynb
https://github.com/microsoft/Phi-3CookBook/blob/main/code/04.Finetuning/Phi-3-finetune-qlora-python.ipynb

The example here in these example scripts appears to be misleading, as it in peft_config, they show

target_modules = ['k_proj', 'q_proj', 'v_proj', 'o_proj', 'gate_proj', 'down_proj', 'up_proj'],

which is incorrect based on the model's architecture.
366854892-e031d0bf-a871-40d7-b35f-8a04b56979e8

It should be
target_modules = ['qkv_proj', 'o_proj', 'gate_up_proj', 'down_proj'],

Tasks

Preview Give feedback
No tasks being tracked yet.
@sofyc sofyc changed the title Wong LoRA tuning example Wrong LoRA tuning example Sep 12, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants