Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions pretrain_retro.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@

from functools import partial
import torch
from importlib import import_module

from megatron.training import get_args
from megatron.training import get_tokenizer
Expand All @@ -19,7 +20,6 @@
from megatron.core.enums import ModelType
from megatron.core.models.retro import get_retro_decoder_block_spec, RetroConfig, RetroModel
from megatron.core.models.retro.utils import get_all_true_mask
from megatron.core.tokenizers import MegatronTokenizer
from megatron.training import pretrain
from megatron.training.utils import get_ltor_masks_and_position_ids
from pretrain_gpt import (
Expand All @@ -46,7 +46,7 @@ def core_model_provider(pre_process=True, post_process=True):
else:
block_spec = get_retro_decoder_block_spec(config, use_transformer_engine=True)

print_rank_0('building GPT model ...')
print_rank_0('building Retro model ...')
model = RetroModel(
config=config,
transformer_layer_spec=block_spec,
Expand Down
Loading