You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[rank0]: File "/mnt/d/nanotron-pretrain/nanotron/src/nanotron/models/llama.py", line 529, in forward
[rank0]: (query_unpad, indices_q, cu_seqlens_q, max_seqlen_q) = bert_padding.unpad_input(
[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[rank0]: ValueError: too many values to unpack (expected 4)
Any help to resolve this issue will be greatly appreciated. Thanks.
The text was updated successfully, but these errors were encountered:
I was testing out the library for the model on a single GPU for training.
Used the following command to run the training,
Made some changes in the config_tiny_llama.yaml file which include,
The training ran smoothly and the checkpoints were generated, however when I try to run the model using,
I get the following error,
Any help to resolve this issue will be greatly appreciated. Thanks.
The text was updated successfully, but these errors were encountered: