Replies: 3 comments
-
|
Similar issue to me. |
Beta Was this translation helpful? Give feedback.
-
|
Drop all caches with You might also want to give this a read and disable VPR carveout if it's enabled. The post is about a different board but the VPR carveout is still a thing on Orin. https://forums.developer.nvidia.com/t/jp-5-0-2-missing-1gb-volatile-memory/229214 |
Beta Was this translation helpful? Give feedback.
-
|
Looks like it's a bug that NVIDIA still hasn't fixed. Only reliable mitigation is to reflash your system with Jetpack 6.2.1 and not upgrade any packages. Always use the NVIDIA forum for Jetson support, very few users here have access to Jetson systems. https://forums.developer.nvidia.com/t/unable-to-allocate-cuda0-buffer-after-updating-ubuntu-packages/347862 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I am trying to launch a model with llama.cpp on a Jetson Orin Nano device, but I am getting an OOM error each time that I try to run with the full model.
I used the llama.cpp@03792ad, built with:
Then, I tested llama-cli with:
Using the whole model (31 layers) gave me this output:
As you read, the device has
device CUDA0 (Orin) (0000:00:00.0) - 6687 MiB freeand the model would only require3573.76 MiB.If someone had the same problem, could you guide me through? Also, I tried jetson-containers, but it seems outdated (got no gemma-3n model available error).
BTW, I also tried using
GGML_CUDA_ENABLE_UNIFIED_MEMORY=1 llama-clibut got the same error. (Not sure if the unified memory is working, I have a swap memory of 128Gb)Beta Was this translation helpful? Give feedback.
All reactions