[Core][Distributed] make init_distributed_environment compatible with init_process_group #4014
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Currently we require
world_size
andrank
, but we can make them optional, and have a defaultdistributed_init_method
toenv://
.This way, the following code just works with
torchrun
, which makes test and development very convenient.This should not affect other part of the code, as it only provides some default value for the function.
I do think we can change the arg name
distributed_init_method
toinit_method
, which is more aligned with pytorch. But that will require more code change. Not sure if it is worth the change.cc @zhuohan123 for opinions.