Skip to content

Commit

Permalink
Update transformers cache on Jenkins (NVIDIA#7854)
Browse files Browse the repository at this point in the history
* update transformers cache

Signed-off-by: eharper <[email protected]>

* update

Signed-off-by: eharper <[email protected]>

* add cd

Signed-off-by: eharper <[email protected]>

---------

Signed-off-by: eharper <[email protected]>
Signed-off-by: Piotr Żelasko <[email protected]>
  • Loading branch information
ericharper authored and pzelasko committed Jan 3, 2024
1 parent d0fde23 commit 7d694d1
Showing 1 changed file with 31 additions and 31 deletions.
62 changes: 31 additions & 31 deletions Jenkinsfile
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ pipeline {
agent {
docker {
image 'nvcr.io/nvidia/pytorch:23.09-py3'
args '--device=/dev/nvidia0 --gpus all --user 0:128 -v /home/TestData:/home/TestData -v $HOME/.cache:/root/.cache --shm-size=8g --env TRANSFORMERS_OFFLINE=1 --env HYDRA_FULL_ERROR=1'
args '--device=/dev/nvidia0 --gpus all --user 0:128 -v /home/TestData:/home/TestData -v $HOME/.cache:/root/.cache --shm-size=8g --env TRANSFORMERS_OFFLINE=0 --env HYDRA_FULL_ERROR=1'
}
}
options {
Expand Down Expand Up @@ -801,7 +801,7 @@ pipeline {
// TODO: pleasefixme @redoctopus
// stage('ByT5G2P training, evaluation and inference') {
// steps {
// sh 'TRANSFORMERS_OFFLINE=1 && cd examples/tts/g2p && \
// sh 'cd examples/tts/g2p && \
// TIME=`date +"%Y-%m-%d-%T"` && OUTPUT_DIR_T5=output_byt5_${TIME} && \
// python g2p_train_and_evaluate.py \
// train_manifest=/home/TestData/g2p/g2p.json \
Expand All @@ -818,7 +818,7 @@ pipeline {
// python g2p_inference.py \
// pretrained_model=${OUTPUT_DIR_T5}/T5G2P/test/checkpoints/T5G2P.nemo \
// manifest_filepath=/home/TestData/g2p/g2p.json \
// phoneme_field=text && TRANSFORMERS_OFFLINE=1'
// phoneme_field=text'
// }
// }
stage('HeteronymClassificationModel training, evaluation and inference') {
Expand Down Expand Up @@ -945,7 +945,7 @@ pipeline {
parallel {
stage('Dialogue: Intent and slot classification using GPT') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
model.dataset.data_dir=/home/TestData/nlp/sgd_small \
model.language_model.lm_checkpoint=/home/TestData/nlp/gpt2/pytorch_model.bin\
Expand All @@ -972,7 +972,7 @@ pipeline {
}
stage('Intent and slot classification using SGDQA') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
model.dataset.data_dir=/home/TestData/nlp/sgd_small \
model.dataset.dialogues_example_dir=sgd_gen_bert_outputs \
Expand All @@ -995,7 +995,7 @@ pipeline {
}
stage('Intent and slot classification using IntentSlotClassificationModel') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
model.dataset.data_dir=/home/TestData/nlp/processed_assistant \
model.dataset.dialogues_example_dir=sgd_gen_bert_intent_classification_outputs \
Expand All @@ -1012,12 +1012,12 @@ pipeline {
model.language_model.pretrained_model_name=bert-base-uncased \
trainer.accelerator=gpu \
exp_manager=null && \
rm -rf sgd_gen_bert_intent_classification_outputs && TRANSFORMERS_OFFLINE=1'
rm -rf sgd_gen_bert_intent_classification_outputs'
}
}
stage('Intent classification using ZeroShotIntentModel') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
do_training=False \
model.dataset.data_dir=/home/TestData/nlp/drive_thru_revised \
Expand All @@ -1037,12 +1037,12 @@ pipeline {
model.language_model.pretrained_model_name=bert-base-uncased \
trainer.accelerator=gpu \
exp_manager=null && \
rm -rf sgd_gen_zero_shot_intent_classification_outputs && TRANSFORMERS_OFFLINE=1'
rm -rf sgd_gen_zero_shot_intent_classification_outputs'
}
}
stage('Design Intent classification using ZeroShotIntentModel') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
do_training=False \
model.dataset.data_dir=/home/TestData/nlp/design_dataset \
Expand All @@ -1063,12 +1063,12 @@ pipeline {
model.language_model.pretrained_model_name=bert-base-uncased \
trainer.accelerator=gpu \
exp_manager=null && \
rm -rf design_zero_shot_intent_classification_outputs && TRANSFORMERS_OFFLINE=1'
rm -rf design_zero_shot_intent_classification_outputs'
}
}
stage('Design Intent classification using ZeroShotIntentModel BART Classifier') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
do_training=False \
model.dataset.data_dir=/home/TestData/nlp/design_dataset \
Expand All @@ -1082,12 +1082,12 @@ pipeline {
model.language_model.pretrained_model_name=bert-base-uncased \
trainer.accelerator=gpu \
exp_manager=null && \
rm -rf design_zero_shot_intent_classification_bart_outputs && TRANSFORMERS_OFFLINE=1'
rm -rf design_zero_shot_intent_classification_bart_outputs'
}
}
stage('Design Intent classification using DialogueNearestNeighbourModel') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
do_training=False \
model.dataset.data_dir=/home/TestData/nlp/design_dataset \
Expand All @@ -1100,7 +1100,7 @@ pipeline {
model.language_model.pretrained_model_name=sentence-transformers/all-MiniLM-L6-v2 \
trainer.accelerator=gpu \
exp_manager=null && \
rm -rf design_dialogue_nearest_neighbour_classification_outputs && TRANSFORMERS_OFFLINE=1'
rm -rf design_dialogue_nearest_neighbour_classification_outputs'
}
}
}
Expand All @@ -1116,7 +1116,7 @@ pipeline {
parallel {
stage('Dialogue: Answer Extender using DialogueS2SGenerationModel') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
do_training=False \
model.dataset.data_dir=/home/TestData/nlp/ms-marco-qa \
Expand All @@ -1141,7 +1141,7 @@ pipeline {
}
stage('Dialogue: SGD Based Answer Extender using DialogueS2SGenerationModel') {
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
do_training=False \
model.dataset.data_dir=/home/TestData/nlp/sgd_small \
Expand Down Expand Up @@ -1182,7 +1182,7 @@ pipeline {
// parallel {
// stage('Dialogue: Answer Extender using DialogueGPTGenerationModel') {
// steps {
// sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/dialogue && \
// sh 'cd examples/nlp/dialogue && \
// python dialogue.py \
// do_training=False \
// model.dataset.data_dir=/home/TestData/nlp/ms-marco-qa \
Expand Down Expand Up @@ -1212,7 +1212,7 @@ pipeline {
parallel {
stage('Dialogue: Answer Extender using DialogueGPTGenerationModel') {
steps {
sh 'TRANSFORMERS_OFFLINE=0 && cd examples/nlp/dialogue && \
sh 'cd examples/nlp/dialogue && \
python dialogue.py \
do_training=False \
model.dataset.data_dir=/home/TestData/nlp/ms-marco-qa \
Expand Down Expand Up @@ -1336,7 +1336,7 @@ pipeline {
stage('BERT SQUAD 1.1') {
// Cannot do fast_dev_run because squad needs whole dev dataset
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/question_answering && \
sh 'cd examples/nlp/question_answering && \
python question_answering.py \
model.train_ds.file=/home/TestData/nlp/squad_mini/v1.1/train-v1.1.json \
model.dataset.use_cache=false \
Expand All @@ -1355,13 +1355,13 @@ pipeline {
trainer.precision=16 \
trainer.devices=[0] \
trainer.accelerator="gpu" \
exp_manager=null && TRANSFORMERS_OFFLINE=1'
exp_manager=null'
}
}
stage('BERT SQUAD 2.0') {
// Cannot do fast_dev_run because squad needs whole dev dataset
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/question_answering && \
sh 'cd examples/nlp/question_answering && \
python question_answering.py \
model.train_ds.file=/home/TestData/nlp/squad_mini/v2.0/train-v2.0.json \
model.dataset.use_cache=false \
Expand All @@ -1377,7 +1377,7 @@ pipeline {
trainer.precision=16 \
trainer.devices=[1] \
trainer.accelerator="gpu" \
exp_manager=null && TRANSFORMERS_OFFLINE=1'
exp_manager=null'
}
}
}
Expand All @@ -1395,7 +1395,7 @@ pipeline {
stage('BART SQUAD 1.1') {
// Cannot do fast_dev_run because squad needs whole dev dataset
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/question_answering && \
sh 'cd examples/nlp/question_answering && \
python question_answering.py \
model.train_ds.file=/home/TestData/nlp/squad_mini/v1.1/train-v1.1.json \
model.dataset.use_cache=false \
Expand All @@ -1415,13 +1415,13 @@ pipeline {
trainer.precision=16 \
trainer.devices=[0] \
trainer.accelerator="gpu" \
exp_manager=null && TRANSFORMERS_OFFLINE=1'
exp_manager=null'
}
}
stage('BART SQUAD 2.0') {
// Cannot do fast_dev_run because squad needs whole dev dataset
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/question_answering && \
sh 'cd examples/nlp/question_answering && \
python question_answering.py \
model.train_ds.file=/home/TestData/nlp/squad_mini/v2.0/train-v2.0.json \
model.dataset.use_cache=false \
Expand All @@ -1438,7 +1438,7 @@ pipeline {
trainer.precision=16 \
trainer.devices=[1] \
trainer.accelerator="gpu" \
exp_manager=null && TRANSFORMERS_OFFLINE=1'
exp_manager=null'
}
}
}
Expand All @@ -1456,7 +1456,7 @@ pipeline {
stage('GPT2 SQUAD 1.1') {
// Cannot do fast_dev_run because squad needs whole dev dataset
steps {
sh 'TRANSFORMERS_OFFLINE=0 && cd examples/nlp/question_answering && \
sh 'cd examples/nlp/question_answering && \
python question_answering.py \
model.train_ds.file=/home/TestData/nlp/squad_mini/v1.1/train-v1.1.json \
model.dataset.use_cache=false \
Expand All @@ -1476,13 +1476,13 @@ pipeline {
trainer.precision=16 \
trainer.devices=[0] \
trainer.accelerator="gpu" \
exp_manager=null && TRANSFORMERS_OFFLINE=1'
exp_manager=null'
}
}
stage('GPT2 SQUAD 2.0') {
// Cannot do fast_dev_run because squad needs whole dev dataset
steps {
sh 'TRANSFORMERS_OFFLINE=1 && cd examples/nlp/question_answering && \
sh 'cd examples/nlp/question_answering && \
python question_answering.py \
model.train_ds.file=/home/TestData/nlp/squad_mini/v2.0/train-v2.0.json \
model.dataset.use_cache=false \
Expand All @@ -1499,7 +1499,7 @@ pipeline {
trainer.precision=16 \
trainer.devices=[1] \
trainer.accelerator="gpu" \
exp_manager=null && TRANSFORMERS_OFFLINE=1'
exp_manager=null'
}
}
}
Expand Down

0 comments on commit 7d694d1

Please sign in to comment.