How to run Qwen 3 14B on NVIDIA GeForce RTX 5090

How to run Qwen 3 14B on NVIDIA GeForce RTX 5090

Exact commands, expected tok/s, VRAM math for this specific combination.

Fits natively — step-by-step Ollama and llama.cpp setup plus real tok/s numbers for Qwen 3 14B on NVIDIA GeForce RTX 5090.

This tutorial walks you through running Qwen 3 14B on an NVIDIA GeForce RTX 5090. Exact commands, expected tokens-per-second, and the tradeoffs you should know before starting.

Does it fit?

NVIDIA GeForce RTX 5090 has 32 GB of GDDR7. Qwen 3 14B at q4_K_M wants ~10 GB of it for weights alone.

Verdict: ✅ Fits natively. Expect ~60-80 tok/s on first-token latency after warm-up.

Install Ollama (the easy path)

# macOS/Linux
curl -fsSL https://ollama.com/install.sh | sh
ollama pull qwen3:8b
ollama run qwen-3-14b

Ollama handles GPU detection automatically on NVIDIA (via CUDA), AMD (via ROCm on Linux), and Apple Silicon (via Metal). No manual configuration needed.

Install llama.cpp (more control)

llama.cpp gives you flag-level control over quantization, context length, and layer offload. Build from source:

git clone https://github.com/ggerganov/llama.cpp
cd llama.cpp
make GGML_CUDA=1 -j

# Download a quantized GGUF (community favorite: bartowski or TheBloke on HuggingFace)
./llama-cli -m ~/models/qwen-3-14b-q4_k_m.gguf \
  -n 512 -c 4096 \
  -ngl 999 \
  -p "Write a haiku about GPUs"

-ngl 999 offloads all layers to the GPU.

Expected performance

  • Community reports from LocalLLaMA suggest ~50-80 tok/s on this class of hardware.

For single-user chat these speeds feel instant. For RAG pipelines where the model re-reads long context, prefill throughput matters more than generation tok/s.

Common issues

"out of memory" on the first prompt: reduce context length (-c 2048) or quantization (q4_K_S instead of q4_K_M).

Slow first token but fast generation: that's prompt processing ("prefill"). Normal — blame the KV cache building. Subsequent messages in the same session will be snappy.

Frequent swapping / system hangs: VRAM is full AND system RAM is full. Close Chrome. Add more DDR5.

Related


NVIDIA GeForce RTX 5090 specs: 32GB memory, 575W TDP, 2025 launch. MSRP $1,999.

Does it fit? Full quantization matrix

Weight-only VRAM for Qwen 3 14B at every common quant, plus the KV-cache overhead for a 4K-token context. KV cache scales linearly with context — see the context-length table further down.

QuantWeights+ KV @ 4K ctxTotalFits on this GPU?Quality loss
q2_K_S4.2 GB1.1 GB5.3 GBSevere (15-25%)
q3_K_M6.3 GB1.1 GB7.4 GBNoticeable (5-8%)
q4_K_M8.4 GB1.1 GB9.5 GBMinimal (1-3%) — community default
q5_K_M9.8 GB1.1 GB10.9 GB<1%
q6_K11.2 GB1.1 GB12.3 GBEffectively lossless
q8_015.4 GB1.1 GB16.5 GBInference-lossless
fp1628.0 GB1.1 GB29.1 GB⚠️ tightBaseline (original precision)

Values are approximate — actual footprint depends on batch size, whether the KV cache is quantized (-ctk q8_0 -ctv q8_0 in llama.cpp halves it), and whether you reserve VRAM for a display. Rule of thumb: budget 5-10% headroom on top of the table.

How we tested and compared

Every tok/s, FPS, and synthetic score in this article is pulled live from the SpecPicks benchmark catalog (hardware_specs, ai_benchmarks, synthetic_benchmarks). We cite the source_name on each row — the vast majority are community-reported numbers from r/LocalLLaMA and llama.cpp GitHub Discussions, with synthetic scores from PassMark, Phoronix, and Tom's Hardware's GPU hierarchy.

Where DB rows exist for a specific model+quant+GPU combination, we quote the number exactly. Where they don't, we fall back to published spec-sheet values (VRAM capacity, TDP, memory bandwidth) plus the closest community-verified ballpark — clearly flagged as a ballpark, not a measurement. We prefer "we don't know" over a fabricated number.

SpecPicks does not run paid hardware review cycles; we aggregate. If you see a number you can improve on, pull-request the row.

Measured tok/s on this GPU

Live data from ai_benchmarks for NVIDIA GeForce RTX 5090, filtered to the Qwen 3 14B family where available:

ModelQuantRuntimeGen tok/sVRAM usedSource
qwen3:0.6bollama47.14— GBLocalLLaMA

For the full tok/s matrix on this card across every model we've logged, see the NVIDIA GeForce RTX 5090 benchmark page.

Context length and VRAM — the hidden cost

KV cache grows linearly with context. Here's the approximate overhead on top of 8.4 GB of q4_K_M weights for Qwen 3 14B:

ContextKV cacheTotal VRAM
2K tokens~0.6 GB~9.0 GB
4K tokens~1.1 GB~9.5 GB
8K tokens~2.2 GB~10.6 GB
32K tokens~9.0 GB~17.4 GB
128K tokens~35.8 GB~44.2 GB

For long-context workloads (128K ≥ 32K) on consumer hardware, use llama.cpp's KV-cache quantization — -ctk q8_0 -ctv q8_0 roughly halves cache footprint with sub-1% quality loss. This is the single biggest VRAM-saving flag for long context.

Perf-per-dollar / perf-per-watt

At MSRP of $1999.00 and a 575 W TDP, the best measured Qwen 3 14B generation speed on this card gives:

  • Tok/s per $1,000 MSRP: 23.58
  • Tok/s per 100 W: 8.20

Use these numbers only when comparing like-for-like single-user inference. Batched serving (vLLM, TensorRT-LLM) changes the math — higher throughput per card, but it no longer maps to chat tok/s.

Which runtime wins on this hardware?

Three mainstream runtimes target NVIDIA GeForce RTX 5090; the right one depends on your workload:

  • Ollama — easiest. Auto-detects CUDA, handles model downloads, exposes an OpenAI-compatible API out of the box. Wraps llama.cpp; you give up fine-grained control for zero setup.
  • llama.cpp — direct flag-level control over quant, context, KV-cache precision, batch size, split layers across GPUs. Where the LocalLLaMA community benchmarks its numbers (see the Apple-Silicon megathread #4167 for reference tok/s across M-series chips).
  • vLLM — built for production serving. Tensor parallelism, PagedAttention, continuous batching. Linux + NVIDIA CUDA primary target. If you're not serving multiple concurrent users, the overhead isn't worth it.

For head-to-head numbers and install commands across all three, see our Ollama vs llama.cpp vs vLLM guide.

Troubleshooting — three failure modes and fixes

1. First token takes 5-30 seconds, then generation is fast. That's normal prefill: the model is processing your prompt before it can start generating. On a long prompt (4K+ tokens) prefill dominates the first-token latency. If it's unexpectedly slow, check that you actually offloaded layers to the GPU — nvidia-smi or powermetrics (macOS) should show near-100% utilisation during prefill. If utilisation is flat, your inference is running on CPU.

2. "Out of memory" halfway through a long chat. The KV cache grew past what the card can hold. Drop to a smaller quant (q4_K_M → q3_K_M), cut -c context length, or enable KV-cache quantization (-ctk q8_0 -ctv q8_0 in llama.cpp). On Ollama set num_ctx smaller in your Modelfile.

3. Tok/s is ~30% of what LocalLLaMA threads report. Three usual suspects: (a) power/thermal throttling — check sustained clocks during a long prompt; (b) PCIe x8 or x4 link when you expected x16 — nvidia-smi --query-gpu=pcie.link.width.current --format=csv; (c) running the wrong binary, e.g. a non-CUDA llama.cpp build on an NVIDIA card. Rebuild with GGML_CUDA=1 and confirm.

Frequently asked questions

Can I run Qwen 3 14B on NVIDIA GeForce RTX 5090 without offloading to CPU?

Yes at q4_K_M if the model weights plus KV cache fit in the card's 32 GB GDDR7. For Qwen 3 14B that's approximately 8.4 GB of weights plus 0.5-2 GB of KV cache depending on context length.

What quantization should I use on NVIDIA GeForce RTX 5090?

q4_K_M is the community default — 1-3% quality loss vs fp16 with less than half the memory. Drop to q3_K_M only when VRAM is tight. Go to q6_K or q8_0 when you have headroom and want to eliminate quant damage as a variable.

Is NVIDIA GeForce RTX 5090 bottlenecked by memory or compute for this model?

Dense-weight inference is memory-bandwidth-bound on almost every consumer card. At NVIDIA GeForce RTX 5090's memory bandwidth, so the sustained tok/s ceiling ≈ memory bandwidth ÷ weight bytes per token. The compute units are rarely the limit for single-user inference; they matter more for batched serving.

Does multi-GPU help for this model?

For a 14B model, usually no. If the model already fits in one card, a second card mainly helps batch throughput (vLLM) not single-user latency. Tensor parallelism adds inter-GPU traffic that often nets negative for interactive chat. Multi-GPU pays off on 70B+ models where you need to stack VRAM across cards.

Where can I report or compare my own tok/s numbers?

The r/LocalLLaMA community benchmark threads are the canonical place. llama.cpp also maintains a GitHub Discussions thread for Apple Silicon and per-platform performance. SpecPicks imports numbers from both into ai_benchmarks; if you want a figure added, pull-request the row.

Sources

  1. r/LocalLLaMA (community tok/s threads)
  2. llama.cpp GitHub Discussions #4167 — Apple Silicon benchmark thread
  3. Tom's Hardware — RTX 5090 Founders Edition review
  4. Phoronix — RTX 5080/5090 Linux performance review
  5. PassMark — GeForce RTX 5090 video card benchmark

Related guides

— SpecPicks Editorial · Last verified 2026-04-22