Skip to content

Commit 153b6ed

Browse files
committed
Change argument name.
1 parent ec43edb commit 153b6ed

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

megatron/training/arguments.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1129,7 +1129,7 @@ def validate_args(args, defaults={}):
11291129
assert not args.distrib_optim_fully_reshardable_mem_efficient, \
11301130
'--distrib-optim-fully-reshardable-mem-efficient requires -enable-gloo-process-groups'
11311131

1132-
if args.fake_distributed_group:
1132+
if args.fake_process_group:
11331133
assert not args.enable_gloo_process_groups, "Fake distributed group requires disabling Gloo process groups."
11341134

11351135
# Checkpointing
@@ -2836,8 +2836,8 @@ def _add_distributed_args(parser):
28362836
"and must be consistent across all ranks.")
28372837
group.add_argument('--replication-factor', default=2, type=int,
28382838
help="Number of machines storing the replica of a given rank's data.")
2839-
group.add_argument('--fake-distributed-group', action='store_true', default=False,
2840-
help='If set, initialize with fake distributed group and all distributed communication operations will be skipped. \
2839+
group.add_argument('--fake-process-group', action='store_true', default=False,
2840+
help='If set, initialize with fake distributed process group and all distributed communication operations will be skipped. \
28412841
This is quite useful for profiling memory usage of distributed training with just one GPU. \
28422842
Setting WORLD_SIZE and RANK to the specific values for target distribtued scale.')
28432843
return parser

megatron/training/initialize.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -346,7 +346,7 @@ def _initialize_distributed(get_embedding_ranks, get_position_embedding_ranks, s
346346
'rank': args.rank,
347347
'timeout': timedelta(minutes=args.distributed_timeout_minutes),
348348
}
349-
if args.fake_distributed_group:
349+
if args.fake_process_group:
350350
from torch.testing._internal.distributed.fake_pg import FakeStore
351351
store = FakeStore()
352352
init_process_group_kwargs['backend'] = 'fake'

0 commit comments

Comments
 (0)