🚀 NVIDIA DGX Spark vs. Alternatives: Escaping the RTX 3060 (6GB) for Medical LLM Research
Posted by Muted-Examination278@reddit | LocalLLaMA | View on Reddit | 8 comments
Hi r/LocalLLaMA 🚀 ,
I am currently struggling with my medical LLM research (language models only, no images/video) on my existing RTX 3060 6GB laptop GPU. As you can imagine, this is a major bottleneck—even simple LoRA experiments on small models are cumbersome due to the severe lack of VRAM. It's time to scale up.
Planned operations include: Intensive fine-tuning (LoRA/QLoRA), distillation, and pruning/quantization of large models (targeting 7B to 70B+) for clinical applications.
I am mainly considering two directions for a new setup:
- NVIDIA DGX Spark: Full power, maximum VRAM, and complete compatibility with the CUDA ecosystem. This is the ideal solution to ensure research freedom when loading and optimizing large LLMs.
- AMD-based Alternatives (e.g., future Strix Halo/similar): This option is theoretically cheaper, but I honestly dread the potential extra effort and debugging associated with ROCm and the general lack of ecosystem maturity compared to CUDA, especially for specialized LLM tasks (LoRA, QLoRA, distillation, etc.). I need to focus on research, not fighting drivers.
My questions to the community:
- For someone focused purely on research fine-tuning and optimization of LLMs (LoRA/Distillation), and who wants to avoid software friction—is the DGX Spark (or an equivalent H100 cluster) the only viable path?
- Are experiments like LoRA on 70B+ models even feasible when attempting to use non-NVIDIA/non-high-VRAM alternatives?
- Has anyone here successfully used AMD (Strix Halo or MI300 series) for advanced LLM research involving LoRA and distillation? How painful is it compared to CUDA?
Any perspective from an LLM researcher is greatly appreciated. Thank you!
Such_Advantage_6949@reddit
Cloud. Speed is money, the different between waiting days vs hours are huge. Spark should be good but be prepared that anything big, u probably end up using cloud
Prestigious_Fold_175@reddit
RTX 6000 pro
Muted-Examination278@reddit (OP)
Thank you for the suggestion! Unfortunately, solutions like the RTX 6000 significantly exceed my maximum budget (\~$4000 USD).
Prestigious_Fold_175@reddit
Mac Studio M5 Max ?
Ok_Appearance3584@reddit
For your use case, DGX Spark. But +70B is out of reach unless it's MoE like gpt-oss 120B.
Serprotease@reddit
The spark is good to tinker and prototype. You can fine tune 8b models fine with it.Â
But a 70b?? You probably can do it, but you definitely do not want to do it. That’s at least a couple of weeks of this thing running full tilt. Â
You want to rent gpus a100/h100 for this.Â
No-Refrigerator-1672@reddit
Making a LoRA for 70B+ dense model on DGX Spark will take months. The thing is painfully slow, it's only ever usable for MoE models at that size. From what you've described, you need a big, expensive, dedicated GPU rig. If you don't feel yourself confident enpugh to assemble such rig, then you can make DGX Spark work by restraining yourself to <30B quantized models, or ~100B quantized MoE. Using AMD for anything other than mainstream inference is also a no-no: most of the advanced stuff require CUDA those days, and for every cent you save by buying AMD you'll pay tenfold in your time spent making the software run.
AdDizzy8160@reddit
... making the software run and keep it running.