You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I tried to run this with llama 8b model in 8*H100s
With batch size 8 it failed with OOM
I want batch_size smaller than n_gpus_per_node because of memory
Is there other way?
I tried to run this with llama 8b model in 8*H100s With batch size 8 it failed with OOM I want batch_size smaller than n_gpus_per_node because of memory Is there other way?
/verl/protocol.py", line 491, in chunk
请问是该错误是为何
The text was updated successfully, but these errors were encountered: