Skip to content

Commit

Permalink
Merge branch 'fix219' into fix224
Browse files Browse the repository at this point in the history
  • Loading branch information
co63oc committed Jan 9, 2025
2 parents 076d918 + 6a0c53e commit 9e225a8
Show file tree
Hide file tree
Showing 2 changed files with 8 additions and 18 deletions.
13 changes: 4 additions & 9 deletions test/collective/collective_global_gather.py
Original file line number Diff line number Diff line change
Expand Up @@ -62,10 +62,8 @@ def run_trainer(self, args):
endpoints = args["endpoints"].split(",")
rank = args["trainerid"]
current_endpoint = args["currentendpoint"]
if args["dynamic_static_unified_comm"]:
paddle.distributed.collective._init_parallel_env(args["backend"])
else:
paddle.distributed.init_parallel_env()

paddle.distributed.collective._init_parallel_env(args["backend"])
nranks = 2
if args['backend'] == 'nccl':
device_id = int(os.getenv("FLAGS_selected_gpus", "0"))
Expand Down Expand Up @@ -112,11 +110,8 @@ def run_trainer(self, args):
)

if args['static_mode']:
result = (
self.get_model(train_prog, startup_prog, rank)
if args["dynamic_static_unified_comm"]
else self.get_model(train_prog, startup_prog, rank)
)
result = self.get_model(train_prog, startup_prog, rank)

fetch_list = []
for elem in result:
fetch_list.append(elem.name)
Expand Down
13 changes: 4 additions & 9 deletions test/collective/collective_global_scatter.py
Original file line number Diff line number Diff line change
Expand Up @@ -63,10 +63,8 @@ def run_trainer(self, args):
rank = args["trainerid"]
current_endpoint = args["currentendpoint"]
nranks = 2
if args["dynamic_static_unified_comm"]:
paddle.distributed.collective._init_parallel_env(args["backend"])
else:
paddle.distributed.init_parallel_env()

paddle.distributed.collective._init_parallel_env(args["backend"])
if args['backend'] == 'nccl':
device_id = int(os.getenv("FLAGS_selected_gpus", "0"))
place = base.CUDAPlace(
Expand All @@ -90,11 +88,8 @@ def run_trainer(self, args):
"float32"
)
if args['static_mode']:
result = (
self.get_model(train_prog, startup_prog, rank)
if args["dynamic_static_unified_comm"]
else self.get_model(train_prog, startup_prog, rank)
)
result = self.get_model(train_prog, startup_prog, rank)

exe = base.Executor(place)
exe.run(startup_prog)
fetch_list = []
Expand Down

0 comments on commit 9e225a8

Please sign in to comment.