-
Notifications
You must be signed in to change notification settings - Fork 253
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Questions about multi-GPU training #2
Comments
Hello xhh232018, |
Hello TrinhQuoc, |
Hi xhh232018, |
@TrinhQuocNguyen how did u train yr own? |
@xhh232018 I am trying to use multi GPUs for training too. However, it seems to consume a lot of CPU. Did that occur in your training procedure? Also, I find it consumes much more space on the first GPU, which makes it hard to fully use all GPU. Could you give me any advice on that? Thank you a lot. |
I met the same problem. Have you guys found the way to deal with it? |
@xhh232018 I am trying to use multi GPUs for training,could you give me any advice on that? And when i trained on my datasets ,such ZeroDivisionError " Found 0 images belonging to 0 classes." appeared,how can i solve this problem?I need your help,thank you! |
@Mistariano Have you solved your problem?i have the same problem,thank you! |
@ZDD2009 I tried to build the models on my CPU first and then used My code likes this:
You can perform this trick on both pconv_model and vgg. It can exactly speed up the training. However, the first gpu still used much more mem than others after I did that. I have no idea how to deal with the problem. |
@Mistariano thank you very much! |
I've also been playing with multi-GPU implementation, but I've not been able to see any successful speedups. Seems like the VGG loss evaluations always happen on the first GPU, and so it doesn't scale well. If anyone figures out a solution for this, it'd be awesome. |
def init(self, img_rows=512, img_cols=512, vgg_weights="imagenet", inference_only=False, net_name='default', gpus=8, vgg_device=None): |
Here is solution. Best regards, |
Hi, due to the quite long training time, I want to know how can I use the keras.utils.multi_gpu_model?
The text was updated successfully, but these errors were encountered: