Skip to content

Commit

Permalink
Update alpaca-lora to use transformers main branch
Browse files Browse the repository at this point in the history
  • Loading branch information
oceanplexian authored and tloen committed Mar 16, 2023
1 parent c3d9470 commit 1862976
Show file tree
Hide file tree
Showing 5 changed files with 13 additions and 16 deletions.
7 changes: 2 additions & 5 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,16 +16,13 @@ Without hyperparameter tuning or validation-based checkpointing, the LoRA model

### Setup

Until Jason Phang's [LLaMA implementation](https://github.com/huggingface/transformers/pull/21955)
is merged, users will need to replace their local `transformers` package.

1. Install dependencies (**install zphang's transformers fork**)
1. Install dependencies

```
pip install -q datasets loralib sentencepiece accelerate
pip uninstall transformers
pip install -q git+https://github.com/zphang/transformers@c3dc391
pip install -q git+https://github.com/huggingface/transformers.git
pip install -q git+https://github.com/huggingface/peft.git
```
Expand Down
6 changes: 3 additions & 3 deletions export_state_dict_checkpoint.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,11 +3,11 @@

import torch
from peft import PeftModel, LoraConfig
from transformers import LLaMATokenizer, LLaMAForCausalLM
from transformers import LlamaTokenizer, LlamaForCausalLM

tokenizer = LLaMATokenizer.from_pretrained("decapoda-research/llama-7b-hf")
tokenizer = LlamaTokenizer.from_pretrained("decapoda-research/llama-7b-hf")

base_model = LLaMAForCausalLM.from_pretrained(
base_model = LlamaForCausalLM.from_pretrained(
"decapoda-research/llama-7b-hf",
load_in_8bit=False,
torch_dtype=torch.float16,
Expand Down
6 changes: 3 additions & 3 deletions finetune.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
import bitsandbytes as bnb
from datasets import load_dataset
import transformers
from transformers import AutoTokenizer, AutoConfig, LLaMAForCausalLM, LLaMATokenizer
from transformers import AutoTokenizer, AutoConfig, LlamaForCausalLM, LlamaTokenizer
from peft import prepare_model_for_int8_training, LoraConfig, get_peft_model


Expand All @@ -21,12 +21,12 @@
LORA_ALPHA = 16
LORA_DROPOUT = 0.05

model = LLaMAForCausalLM.from_pretrained(
model = LlamaForCausalLM.from_pretrained(
"decapoda-research/llama-7b-hf",
load_in_8bit=True,
device_map="auto",
)
tokenizer = LLaMATokenizer.from_pretrained(
tokenizer = LlamaTokenizer.from_pretrained(
"decapoda-research/llama-7b-hf", add_eos_token=True
)

Expand Down
6 changes: 3 additions & 3 deletions generate.py
Original file line number Diff line number Diff line change
@@ -1,10 +1,10 @@
import torch
from peft import PeftModel
from transformers import LLaMATokenizer, LLaMAForCausalLM, GenerationConfig
from transformers import LlamaTokenizer, LlamaForCausalLM, GenerationConfig

tokenizer = LLaMATokenizer.from_pretrained("decapoda-research/llama-7b-hf")
tokenizer = LlamaTokenizer.from_pretrained("decapoda-research/llama-7b-hf")

model = LLaMAForCausalLM.from_pretrained(
model = LlamaForCausalLM.from_pretrained(
"decapoda-research/llama-7b-hf",
load_in_8bit=True,
torch_dtype=torch.float16,
Expand Down
4 changes: 2 additions & 2 deletions lengths.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -19,10 +19,10 @@
],
"source": [
"from datasets import load_dataset\n",
"from transformers import LLaMATokenizer\n",
"from transformers import LlamaTokenizer\n",
"\n",
"\n",
"tokenizer = LLaMATokenizer.from_pretrained(\"decapoda-research/llama-7b-hf\", add_eos_token=True)\n",
"tokenizer = LlamaTokenizer.from_pretrained(\"decapoda-research/llama-7b-hf\", add_eos_token=True)\n",
"tokenizer.pad_token = tokenizer.eos_token\n",
"tokenizer.pad_token_id = tokenizer.eos_token_id\n",
"\n",
Expand Down

0 comments on commit 1862976

Please sign in to comment.