The 2020–2022 crypto boom made buying GPUs a nightmare, especially Nvidia’s RTX series. Back then, Team Red’s RX6000 lineup felt like a steal. I grabbed an AMD RX6750XT—same chip as the RX6700XT (gfx1031)—and skipped ray tracing for affordable rasterization. "Price over pixels," right?
Fast forward to 2024. Crypto’s fading, but the AI boom exploded. As a developer itching to experiment with LLMs and PyTorch, I realized my mistake: Nvidia’s CUDA dominance leaves AMD GPUs in the dust. ROCm—AMD’s answer to CUDA—officially supports RX6800XT (gfx1030) but snubs RX6700XT/6750XT (gfx1031). Cue the regret.
If I could time-travel, I’d slap my 2022 self and yell: “Buy an RTX 3070 instead!” But since I’m stuck with my RX6750XT, here’s how I forced it to play nice with machine learning workloads.
Tools Used:
amdgpu-install
Debian Package for RX6700XT from here.amdgpu-install --usecase=graphics,rocm
Add yourself to the video and render groups:
sudo usermod -a -G video $USER
sudo usermod -a -G render $USER
Reboot.
Post-reboot, run rocminfo
. You should be able to see your GPU here.
Here’s my working Ollama Docker command:
docker run -d --restart always \
--device /dev/kfd --device /dev/dri \
-v ollama:/root/.ollama -p 11434:11434 \
--name ollama \
-e HSA_OVERRIDE_GFX_VERSION=10.3.0 \
-e HCC_AMDGPU_TARGET=gfx1030 \
ollama/ollama:rocm
AMD’s improving ROCm, but driver issues and hardware restrictions linger. Hopefully the situation improves for future platforms.
References:
Notes: this blog post was edited by AI.