-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat(distributed): RPC-based distributed training support and add distributed MAML example #83
Conversation
Codecov ReportBase: 73.20% // Head: 66.59% // Decreases project coverage by
Additional details and impacted files@@ Coverage Diff @@
## main #83 +/- ##
==========================================
- Coverage 73.20% 66.59% -6.61%
==========================================
Files 33 37 +4
Lines 1515 1853 +338
==========================================
+ Hits 1109 1234 +125
- Misses 406 619 +213
Flags with carried forward coverage won't be shown. Click here to find out more.
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here. ☔ View full report at Codecov. |
6ba8eae
to
16add76
Compare
92cc6ea
to
d738101
Compare
d738101
to
1826df3
Compare
…tributed MAML example
1826df3
to
111ddb3
Compare
a4ef769
to
a29292f
Compare
Description
Describe your changes in detail.
Motivation and Context
Why is this change required? What problem does it solve?
If it fixes an open issue, please link to the issue here.
You can use the syntax
close #15213
if this solves the issue #15213Resolves #57
Types of changes
What types of changes does your code introduce? Put an
x
in all the boxes that apply:Implemented Tasks
New APIs
torchopt.distributed.is_available
torchopt.distributed.backward
World and Process Group:
torchopt.distributed.auto_init_rpc
torchopt.distributed.get_world_info
torchopt.distributed.get_world_rank
(torchopt.distributed.get_rank
)torchopt.distributed.get_world_size
torchopt.distributed.get_local_rank
torchopt.distributed.get_local_world_size
torchopt.distributed.barrier
Wrappers:
torchopt.distributed.auto_init_rpc
torchopt.distributed.on_rank
torchopt.distributed.not_on_rank
torchopt.distributed.rank_zero_only
torchopt.distributed.rank_non_zero_only
Remote call:
torchopt.distributed.remote_async_call
torchopt.distributed.remote_sync_call
torchopt.distributed.parallelize
(torchopt.distributed.parallelize_sync
)torchopt.distributed.parallelize_async
Misc:
torchopt.distributed.dim_partitioner
torchopt.distributed.batch_partitioner
torchopt.distributed.exclusive_batch_partitioner
torchopt.distributed.mean_reducer
torchopt.distributed.sum_reducer
Examples:
Checklist
Go over all the following points, and put an
x
in all the boxes that apply.If you are unsure about any of these, don't hesitate to ask. We are here to help!
make format
(required)make lint
(required)make test
pass. (required)