ASUS Ascent GX10 - Having tons of issues
Posted by LivingHighAndWise@reddit | LocalLLaMA | View on Reddit | 16 comments
Hi all,
Looking for some advice with a GX10 I purchased about 4 months ago. I've been having all kind of issues trying to run local models on this device. It has constant crash/reboot events under heavy load. It doesn't matter if I run models using Ollama or Spark native. These crashes have corrupted the OS twice now, and both times I had to send it back to Asus to repair RMA. I've requested a new device twice now, but they keep "repairing" the current device and sending it back to me saying it passes their standard tests.
I just got it back yesterday, tried installing Nemotron 3 Super, using vLLM, and it crashs and rebooted the device again before it can finish a prompt.
The crash logs show power issues, but nothing definitive of the exact cause. At this point I want my money back, but Asus is not accepting returns on GX10s...
StardockEngineer@reddit
"Spark native" what does that even mean?
I had discovered on mine that having certain USB-c dongles to connect my wireless keyboard and mouse was causing power problems. Changed the dongle and haven't had any problems since.
LivingHighAndWise@reddit (OP)
TensorRT-LLM.
guai888@reddit
You might want to give this a try: GitHub - joeynyc/spark-doctor: Local diagnostic CLI for NVIDIA DGX Spark (GB10). Detects power caps, unified memory pressure, thermal risk, Docker/runtime issues, and validates vLLM/Ollama/llama.cpp/SGLang recipes. · GitHub
CalligrapherFar7833@reddit
You waited months instead of returning it immediately ? Thats stupid
kyrylogorbachov@reddit
I have two devices. Never experienced any of this, have been running them for weeks on full load, no issues. Maybe you want to check the warranty and get replacement.
FoundNil@reddit
I have one too. It took a while to get it working. The ecosystem is starting to catch up on cuda 13. You are almost certainly running out of memory trying to load models that are too big. Helps to have a memory monitor up like glances or nvitop to watch memory pressure.
Igot1forya@reddit
I just returned mine for 3 issues within the first 36h after unboxing.
I sent it back yesterday. I'm going to order a Founders Edition in its place to go next to the Founders Edition I already have.
LivingHighAndWise@reddit (OP)
I have simliar behavior. I'm fear I am beyond the return window..
ProfessionalSpend589@reddit
Does your device work at all with smaller models?
On a different platform I’ve seen OS become unresponsive when I’ve messed parameters for memory usage or when software and/or driver were unstable.
LivingHighAndWise@reddit (OP)
I have run very small models on it and I'm pretty sure they work (Hard to stess test the device with small models). Anything 20GB or larger seems to crash the device.
billy_booboo@reddit
I just submitted myself too claude for configuring this thing but now I have it running quite smooth. You gotta use patched images for everything. For example, here's what I'm using for vllm: https://github.com/AEON-7/Qwen3.6-NVFP4-DFlash/pkgs/container/vllm-spark-omni-q36
LivingHighAndWise@reddit (OP)
What context windows size you running with that patched image?
Powerful_Ad8150@reddit
If vLLM is a must use community build images. Spark run. For usual play use llama.cpp. As for me never got problems with gx10. Nemotron super - not a single crash. Maybe u have defective hardware?
deleted_by_reddit@reddit
[removed]
Excellent_Produce146@reddit
https://spark-arena.com/leaderboard - missed the mentioned Arena
vasimv@reddit
You probably want to check nvidia forum for more help: https://forums.developer.nvidia.com/c/accelerated-computing/dgx-spark-gb10/