You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/_compile.py", line 32, in inner
return disable_fn(*args, **kwargs)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/_dynamo/eval_frame.py", line 632, in _fn
return fn(*args, **kwargs)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/utils/checkpoint.py", line 489, in checkpoint
return CheckpointFunction.apply(function, preserve, *args)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/autograd/function.py", line 575, in apply
return super().apply(*args, **kwargs) # type: ignore[misc]
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/utils/checkpoint.py", line 264, in forward
outputs = run_function(*args)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/transformers/models/mamba2/modeling_mamba2.py", line 649, in forward
hidden_states = self.mixer(
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/transformers/models/mamba2/modeling_mamba2.py", line 608, in forward
return self.torch_forward(hidden_states, cache_params, cache_position, attention_mask)
File "/home/tzl/.conda/envs/SEM/lib/python3.9/site-packages/transformers/models/mamba2/modeling_mamba2.py", line 535, in torch_forward
G_intermediate = C[:, :, :, None, :, :] * B[:, :, None, :, : ,:] # shape: (b, c, l, s, h, n)
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 16.00 GiB. GPU 0 has a total capacity of 23.64 GiB of which 9.08 GiB is free. Including non-PyTorch memory, this process has 14.56 GiB memory in use. Of the allocated memory 14.09 GiB is allocated by PyTorch, and 19.65 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
我确认GPU中无其它进程占用,但还是有CUDA out of memory的问题,因此想请教您一下,谢谢。
Expected behavior
No response
Others
No response
The text was updated successfully, but these errors were encountered:
Reminder
System Info
llamafactory
version: 0.9.1.dev0Reproduction
这是我遇到的问题:
我确认GPU中无其它进程占用,但还是有CUDA out of memory的问题,因此想请教您一下,谢谢。
Expected behavior
No response
Others
No response
The text was updated successfully, but these errors were encountered: