Skip to content

Commit f60f9c8

Browse files
committed
Update doc
Signed-off-by: Asha Anoosheh <[email protected]>
1 parent c1ab42b commit f60f9c8

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

megatron/post_training/docs/distillation.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -75,7 +75,7 @@ Model Optimizer modifies the model using the loss criterion present in the disti
7575
defines a loss function between two module attribute names of the teacher and student model, respectively.
7676

7777
Default loss function used between logits is a KL-Divergence Loss and loss used among intermediate tensors is Cosine-Similarity,
78-
both defined in `megatron/inference/algos/distillation.py`.
78+
both defined in `modelopt.torch.distill.plugins.megatron`.
7979

8080
## Restrictions
8181

0 commit comments

Comments
 (0)