DGX Spark (GB10) llama.cpp build question #20405
Replies: 3 comments 4 replies
-
|
Does |
Beta Was this translation helpful? Give feedback.
-
|
|
Beta Was this translation helpful? Give feedback.
-
|
Finished building on the spark with Did a few llama-bench with KV cache set to Q8_0
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I dont see any specific mention of what flag to pass on during build of llama.cpp. Curious to ask what people have been doing with their llama.cpp builds? Specifically I wonder whether -DCMAKE_CUDA_ARCHITECTURES=121 flag is needed since GB10 is CUDA compute 12.1.
cmake -B build -DGGML_CUDA=ON -DGGML_CUDA_FA_ALL_QUANTS=ON -DCMAKE_CUDA_ARCHITECTURES=121 -DGGML_NATIVE=ON
Beta Was this translation helpful? Give feedback.
All reactions