diff --git a/launcher_scripts/nemo_launcher/collections/dataprep_scripts/multimodal_dataprep/precache_encodings.py b/launcher_scripts/nemo_launcher/collections/dataprep_scripts/multimodal_dataprep/precache_encodings.py index ba058ee9b5..f083be2e3b 100644 --- a/launcher_scripts/nemo_launcher/collections/dataprep_scripts/multimodal_dataprep/precache_encodings.py +++ b/launcher_scripts/nemo_launcher/collections/dataprep_scripts/multimodal_dataprep/precache_encodings.py @@ -22,7 +22,7 @@ import hydra import numpy as np -import pytorch_lightning as pl +import lightning.pytorch as pl import torch import torch.utils.data as data import webdataset as wds diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_baichuan2.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_baichuan2.py index 1371989395..6206965871 100644 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_baichuan2.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_baichuan2.py @@ -38,7 +38,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_chatglm.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_chatglm.py index 9b3dcb3f14..8845d8e250 100644 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_chatglm.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_chatglm.py @@ -38,7 +38,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_falcon.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_falcon.py index 46a2ce521a..023e75094f 100755 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_falcon.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_falcon.py @@ -38,7 +38,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3.py index efe9f651e3..eb2d97c4b4 100755 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3.py @@ -41,7 +41,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3_prompt.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3_prompt.py index 5437ae946e..dbd2700e5f 100755 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3_prompt.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_gpt3_prompt.py @@ -30,7 +30,7 @@ from nemo.utils import logging from nemo.utils.app_state import AppState from nemo.utils.get_rank import is_global_rank_zero -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama.py index fd73cf5ee4..dfd0b86cd4 100755 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama.py @@ -38,7 +38,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama_prompt.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama_prompt.py index 3ecc81c884..831137fece 100755 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama_prompt.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_llama_prompt.py @@ -30,7 +30,7 @@ from nemo.utils import logging from nemo.utils.app_state import AppState from nemo.utils.get_rank import is_global_rank_zero -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mistral.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mistral.py index f14e9c3896..8bc9db7f72 100755 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mistral.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mistral.py @@ -38,7 +38,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mixtral.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mixtral.py index 751b67f90c..3912aff356 100755 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mixtral.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_mixtral.py @@ -38,7 +38,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate diff --git a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_qwen2.py b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_qwen2.py index 7eb7b7c67b..82a9a309ff 100644 --- a/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_qwen2.py +++ b/launcher_scripts/nemo_launcher/collections/eval_harness/lm_eval/models/nemo_qwen2.py @@ -38,7 +38,7 @@ from nemo.utils.get_rank import is_global_rank_zero from nemo.utils.model_utils import inject_model_parallel_rank from omegaconf import OmegaConf, open_dict -from pytorch_lightning.trainer.trainer import Trainer +from lightning.pytorch.trainer.trainer import Trainer from torch.nn.utils.rnn import pad_sequence from torch.utils.data import DataLoader, Dataset from torch.utils.data.dataloader import default_collate