You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
A clear and concise description of what the bug is.
The error is
Traceback (most recent call last):
File "/usr/lib/python3.12/multiprocessing/process.py", line 314, in _bootstrap
self.run()
File "/usr/lib/python3.12/multiprocessing/process.py", line 108, in run
self._target(*self._args, **self._kwargs)
File "/workdir/Megatron-LM/tools/checkpoint/saver_megatron.py", line 168, in save_checkpoint
validate_args(margs)
File "/workdir/Megatron-LM/megatron/training/arguments.py", line 405, in validate_args
args.main_grads_dtype = dtype_map[args.main_grads_dtype]
~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^
KeyError: torch.float32
To Reproduce
Steps to reproduce the behavior. The easier it is to reproduce the faster it will get maintainer attention.
Proposed fix
If you have a proposal for how to fix the issue state it here or link to a PR.
Additional context
Add any other context about the problem here.
When i try to install the transformers=4.31/4.32, there will be an error about compiling the tokenizer whl. So i choose to "pip install transformers" instead. I follow the guidence at https://github.com/NVIDIA/Megatron-LM/blob/main/docs/llama_mistral.md
The text was updated successfully, but these errors were encountered:
Sun2018421
changed the title
[BUG] When trying to convert llama2-7b/Llama3-8B model from HF format to megatron formate
[BUG] When trying to convert llama2-7b model from HF format to megatron format
Jan 6, 2025
Describe the bug
A clear and concise description of what the bug is.
The error is
To Reproduce
Steps to reproduce the behavior. The easier it is to reproduce the faster it will get maintainer attention.
Stack trace/logs
Environment (please complete the following information):
Proposed fix
If you have a proposal for how to fix the issue state it here or link to a PR.
Additional context
Add any other context about the problem here.
When i try to install the transformers=4.31/4.32, there will be an error about compiling the tokenizer whl. So i choose to "pip install transformers" instead. I follow the guidence at https://github.com/NVIDIA/Megatron-LM/blob/main/docs/llama_mistral.md
The text was updated successfully, but these errors were encountered: