Replies: 1 comment
-
|
check out the cpp version that just came out recently. It will run on the jetsons. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I've tried porting this code to NVIDIA Jetson machines without success so far 😬.
Jetson embedded systems have unified memory, and they're almost identical to NVIDIA DGX, which is based on Jetson design.
The Thor series 64-128GB RAM uses CUDA 13.0, and the Orin series 32-64GB RAM is still on CUDA 12.6.
On these systems you can't upgrade easily the CUDA version, it is because the CUDA release is deeply embedded with the hadware and OS. You can upgrade without errors only when a new embedded Jetpack is released.
Note that nvidia-smi is not very well supported on Jetson and in most systems the report is empty, even if inference of LLM models is working perfectly fine.
What changes are required to port the code?
Thanks in advance to everyone!
python scripts/check_gpu.py
But if I run: python profile_inference.py --mode tier-test
Beta Was this translation helpful? Give feedback.
All reactions