Model Parallelism

This article gives a good overview of different strategies for using multiple GPUs in neural network training. I currently train Argos Translate models on a single RTX 3090 so I haven’t done any of this but it will probably be necessary in the future.

OpenNMT supports using multiple GPUs and I think they use data parallelism. So it lets you train faster but not train larger models.

1 Like