Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Distributed Training #79

Open
Het-Shah opened this issue Aug 28, 2020 · 5 comments
Open

Distributed Training #79

Het-Shah opened this issue Aug 28, 2020 · 5 comments
Assignees
Labels
enhancement New feature or request

Comments

@Het-Shah
Copy link
Member

We need to add support for Distributed training, we can directly make use of Pytorch DDP if we want as of now. Let me know if anyone wants to take this up.

@Het-Shah Het-Shah added the enhancement New feature or request label Aug 28, 2020
@avishreekh
Copy link
Collaborator

I wouldn't mind taking this up. But I'd need a little time to do it. Let me know if that works

@Het-Shah
Copy link
Member Author

Het-Shah commented Sep 1, 2020

Yea take your time we don't really need to release this immediately anyway.

@srikar2097
Copy link

srikar2097 commented Apr 17, 2021

Hi @Het-Shah and @avishreekh, thanks for creating this wonderful library with support for multiple KD algorithms. The code and implementation is nicely done and structured.

Wanted to know if any update on distributed training is being done? Currently if I do python -m torch.distributed.launch --nproc_per_node=8 --master_port=1234 vanilla_kd.py the library does not run. Multi-GPU training is crucial for this library to be really useful as both model sizes and data sizes are increasing and we cannot get away from using multiple-GPU's for training.

thanks again!

@avishreekh
Copy link
Collaborator

Thank you @srikar2097. We are glad that this library could be useful to you.
We are working on the distributed training enhancement and hope to release it by Mid-may.

Thank you for your patience.

@avishreekh
Copy link
Collaborator

There are certain design choices that we are debating on currently. Will add this feature once it is decided how to efficiently accommodate it in the existing framework. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants