You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Traceback (most recent call last):
File "E:\pyCode\Chinese-LLaMA-Alpaca\scripts\training\run_clm_sft_with_peft.py", line 468, in <module>
main()
File "E:\pyCode\Chinese-LLaMA-Alpaca\scripts\training\run_clm_sft_with_peft.py", line 205, in main
model_args, data_args, training_args = parser.parse_args_into_dataclasses()
File "D:\Python310\lib\site-packages\transformers\hf_argparser.py", line 346, in parse_args_into_dataclasses
obj = dtype(**inputs)
File "<string>", line 118, in __init__
File "D:\Python310\lib\site-packages\transformers\training_args.py", line 1333, in __post_init__
and (self.device.type != "cuda")
File "D:\Python310\lib\site-packages\transformers\training_args.py", line 1697, in device
return self._setup_devices
File "D:\Python310\lib\site-packages\transformers\utils\generic.py", line 54, in __get__
cached = self.fget(obj)
File "D:\Python310\lib\site-packages\transformers\training_args.py", line 1631, in _setup_devices
self.distributed_state = PartialState(backend=self.ddp_backend)
File "D:\Python310\lib\site-packages\accelerate\state.py", line 143, in __init__
torch.distributed.init_process_group(backend=self.backend, **kwargs)
File "D:\Python310\lib\site-packages\torch\distributed\distributed_c10d.py", line 907, in init_process_group
default_pg = _new_process_group_helper(
File "D:\Python310\lib\site-packages\torch\distributed\distributed_c10d.py", line 1013, in _new_process_group_helper
raise RuntimeError("Distributed package doesn't have NCCL " "built in")
RuntimeError: Distributed package doesn't have NCCL built in
详细描述问题
目前因为deepspeed和nccl两个库无法在win上使用,所以现在在win平台下进行训练是不可行的吗?
参考信息
依赖情况(代码类问题务必提供)
运行日志或截图
必查项目
The text was updated successfully, but these errors were encountered: