-
-
Notifications
You must be signed in to change notification settings - Fork 6
Description
HuggingFace is reporting an error at the start of the test step during an experiment run:
[ERROR|base.py:1149] 2025-02-28 12:05:52,437 >> The model 'OptimizedModule' is not supported for . Supported models are ['BartForConditionalGeneration', 'BigBirdPegasusForConditionalGeneration', 'BlenderbotForConditionalGeneration', 'BlenderbotSmallForConditionalGeneration', 'EncoderDecoderModel', 'FSMTForConditionalGeneration', 'GPTSanJapaneseForConditionalGeneration', 'LEDForConditionalGeneration', 'LongT5ForConditionalGeneration', 'M2M100ForConditionalGeneration', 'MarianMTModel', 'MBartForConditionalGeneration', 'MT5ForConditionalGeneration', 'MvpForConditionalGeneration', 'NllbMoeForConditionalGeneration', 'PegasusForConditionalGeneration', 'PegasusXForConditionalGeneration', 'PLBartForConditionalGeneration', 'ProphetNetForConditionalGeneration', 'Qwen2AudioForConditionalGeneration', 'SeamlessM4TForTextToText', 'SeamlessM4Tv2ForTextToText', 'SwitchTransformersForConditionalGeneration', 'T5ForConditionalGeneration', 'UMT5ForConditionalGeneration', 'XLMProphetNetForConditionalGeneration'].
However, the test step appears to work successfully for this experiment despite the error. The model is set to 'facebook/nllb-200-distilled-1.3B'.
Activity
mmartin9684-sil commentedon Mar 3, 2025
Another potential compatibility issue with the recent HuggingFace updates - this warning is reported at the start of training:
mmartin9684-sil commentedon Mar 3, 2025
An additional warning being reported by HF for recent experiments. This warning occurs at the end of preprocessing / start of training.
mmartin9684-sil commentedon Mar 3, 2025
ClearML warning at the start of the training step:
mmartin9684-sil commentedon Mar 3, 2025
Torch warning at the start of training:
mmartin9684-sil commentedon Mar 3, 2025
Warning at the end of training when the model is being saved:
bhartmoore commentedon Mar 3, 2025
I am seeing this warning during mid-training evals:
[WARNING|trainer.py:761] 2025-03-03 18:15:47,707 >> Trainer.tokenizer is now deprecated. You should use Trainer.processing_class instead.
bhartmoore commentedon Mar 3, 2025
Warning at the start of training, just before the
torch.cuda.amp.GradScaler(args...)
warning listed above:davidbaines commentedon Mar 21, 2025
This warning occurs after logging a checkpoint.
[INFO|integration_utils.py:1934] 2025-03-21 14:48:21,216 >> Logging checkpoint artifact
checkpoint-1000
. This may take some time.2025-03-21 14:49:10,742 - clearml.model - INFO - No output storage destination defined, registering local model /tmp/model_package.a7cazdje.zip
/usr/local/lib/python3.10/dist-packages/torch/utils/checkpoint.py:295: FutureWarning:
torch.cpu.amp.autocast(args...)
is deprecated. Please usetorch.amp.autocast('cpu', args...)
instead.It's nice to see how quickly checkpoints are saved.