Skip to content

Fail to load state dict for Megatron Model after PTQ #597

@yikaizhu-baseten

Description

@yikaizhu-baseten

Before submitting an issue, please make sure it hasn't been already addressed by searching through the existing and past issues.

Describe the bug

https://github.com/NVIDIA/TensorRT-Model-Optimizer/blob/main/modelopt/torch/opt/plugins/megatron.py#L80
When I try to use NeMo to load model for finetuning after PTQ, I got errors in this line.
I have no idea what this line, but I have tested in locally and it fails

import torch
import pickle
x = torch.randint(0, 10, (2,))
y = pickle.loads(x.numpy().tobytes())
  • ?

Steps/Code to reproduce bug

  • ?

Expected behavior

Who can help?

  • ?

System information

  • Container used (if applicable): ?
  • OS (e.g., Ubuntu 22.04, CentOS 7, Windows 10): ?
  • CPU architecture (x86_64, aarch64): ?
  • GPU name (e.g. H100, A100, L40S): ?
  • GPU memory size: ?
  • Number of GPUs: ?
  • Library versions (if applicable):
    • Python: ?
    • ModelOpt version or commit hash: ?
    • CUDA: ?
    • PyTorch: ?
    • Transformers: ?
    • TensorRT-LLM: ?
    • ONNXRuntime: ?
    • TensorRT: ?
  • Any other details that may help: ?

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions