The NVIDIA GeForce RTX 5090 and AMD Radeon RX 7900 XTX often end up on the same shopping shortlist. This head-to-head pulls spec deltas, gaming FPS, AI inference tok/s, and synthetic scores from the live SpecPicks benchmark database, plus a decision matrix at the end.
Specs side by side
| NVIDIA GeForce RTX 5090 | AMD Radeon RX 7900 XTX | |
|---|---|---|
| Manufacturer | NVIDIA | AMD |
| Family | Blackwell | RDNA 3 |
| Release year | 2025 | 2022 |
| MSRP | $1,999 | $999 |
| VRAM | 32 GB GDDR7 | 24 GB GDDR6 |
| Memory bus | — | — |
| Shader units | 21760 | 6144 |
| TDP | 575 W | 355 W |
Price delta: the NVIDIA GeForce RTX 5090 is $1000 more expensive than the AMD Radeon RX 7900 XTX.
Synthetic benchmark deltas
Key synthetic scores pulled from the SpecPicks benchmark DB (PassMark, Cinebench, Geekbench, 3DMark):
| Benchmark | NVIDIA GeForce RTX 5090 | AMD Radeon RX 7900 XTX |
|---|---|---|
| Tom's Hardware GPU Hierarchy | 2 % | 959 fps |
| Phoronix: Linux Gaming | 1 reference | — |
| PassMark G3D Mark | 38,935 pts | 31,409 pts |
| PassMark G2D Mark | 1,412 pts | 1,285 pts |
AI inference (where it matters)
Real tok/s numbers for common LLMs at q4_K_M from the SpecPicks ai_benchmarks table:
| Model | NVIDIA GeForce RTX 5090 | AMD Radeon RX 7900 XTX |
|---|---|---|
| llama3.1:8b (q4_K_M) | — | — |
| qwen3:32b (q4_K_M) | — | — |
| llama3.1:70b (q4_K_M) | — | — |
For the full AI benchmark set for each card, see NVIDIA GeForce RTX 5090 benchmarks and AMD Radeon RX 7900 XTX benchmarks.
Power and thermals
TDP: NVIDIA GeForce RTX 5090 575W vs AMD Radeon RX 7900 XTX 355W. That's a 220W delta. Over 4 hours a day, 365 days — roughly 48 USD/yr in electricity at $0.15/kWh. Plan your PSU at least 1.5x peak draw.
Perf-per-dollar
If your primary workload is Llama 3.1 70B inference, the NVIDIA GeForce RTX 5090 at $1,999 yields X tok/s/$1k, and the AMD Radeon RX 7900 XTX at $999 yields Y. The cheaper card is often the better value per-dollar even when it's slower — run the math for YOUR workload, not someone else's.
Decision matrix
| Get the NVIDIA GeForce RTX 5090 if | Get the AMD Radeon RX 7900 XTX if |
|---|---|
| You need the most VRAM / cores in the comparison | Budget is tighter |
| Your workload scales with memory bandwidth | You want better perf-per-dollar |
| You're on a 2025-era platform anyway | You're keeping an older platform |
| You prioritize headroom for future larger models | You know exactly what you need today |
Don't bother with either if your real bottleneck is somewhere else — at the end of the day, both of these are competent parts. If you're gaming at 1080p, if your LLM workload is a single 7B model, if your renders fit in half this VRAM — get the cheaper part and save the delta.
Bottom line
For most buyers in 2026, the choice between the NVIDIA GeForce RTX 5090 and AMD Radeon RX 7900 XTX comes down to how much headroom you value. If you're certain your workload fits today's requirements, the cheaper card is the rational pick. If you're building a workstation you want to keep relevant for 2-3 years of increasingly hungry models, pay up for the VRAM.
Related
- NVIDIA GeForce RTX 5090 full benchmarks →
- AMD Radeon RX 7900 XTX full benchmarks →
- AI Rigs buyer's guide →
- Compare GPUs side-by-side →
How we tested and compared
Every tok/s, FPS, and synthetic score in this article is pulled live from the SpecPicks benchmark catalog (hardware_specs, ai_benchmarks, synthetic_benchmarks). We cite the source_name on each row — the vast majority are community-reported numbers from r/LocalLLaMA and llama.cpp GitHub Discussions, with synthetic scores from PassMark, Phoronix, and Tom's Hardware's GPU hierarchy.
Where DB rows exist for a specific model+quant+GPU combination, we quote the number exactly. Where they don't, we fall back to published spec-sheet values (VRAM capacity, TDP, memory bandwidth) plus the closest community-verified ballpark — clearly flagged as a ballpark, not a measurement. We prefer "we don't know" over a fabricated number.
SpecPicks does not run paid hardware review cycles; we aggregate. If you see a number you can improve on, pull-request the row.
AI inference: per-model tok/s from the SpecPicks catalog
Generation tok/s from ai_benchmarks. A dash means we don't have a matching DB row yet for that hardware + model + quant combination — contribute via pull request.
| Model | Quant | NVIDIA GeForce RTX 5090 (tok/s) | AMD Radeon RX 7900 XTX (tok/s) | Source |
|---|---|---|---|---|
| gemma:26b | q4_0 | 5.00 | — | LocalLLaMA |
| qwen3:0.6b | — | 47.14 | — | LocalLLaMA |
| qwen3:2b | int4 | — | 40.00 | LocalLLaMA |
Synthetic benchmark deltas
PassMark, Phoronix, and Tom's Hardware hierarchy scores, per the underlying source rows in synthetic_benchmarks.
| Benchmark | NVIDIA GeForce RTX 5090 | AMD Radeon RX 7900 XTX | Source |
|---|---|---|---|
| PassMark G2D Mark | 1412.00 pts | 1285.00 pts | PassMark |
| PassMark G3D Mark | 38935.00 pts | 31409.00 pts | PassMark |
| Phoronix: Linux Gaming | 1.00 reference | — | Phoronix |
| Tom's Hardware GPU Hierarchy | 2.00 % | 959.00 fps | Tom's Hardware |
Budget alternative
If both the NVIDIA GeForce RTX 5090 ($1999.00) and AMD Radeon RX 7900 XTX ($999.00) feel overkill, consider the tier below. For gaming at 1440p, an RTX 5070 at $549 or an RX 7900 GRE delivers 80-90% of the experience at less than half the cost — you give up headroom for 4K and some AI/ML work, but not much for modern AAA games.
For AI inference specifically, the cheapest card that holds a 14B q4 model natively in 2026 is the Arc B580 at $249. It's not fast, but it works — and the 12 GB VRAM buys you more headroom than an 8 GB GeForce at the same price.
Get neither if…
- Your actual bottleneck is CPU-limited single-threaded software (older games, emulators) — a cheaper GPU paired with a better CPU will outperform both of these in that workload.
- You only run 7-8B LLMs and don't plan to go larger — the NVIDIA GeForce RTX 5090 and AMD Radeon RX 7900 XTX are both massively over-provisioned for that use case. An RTX 4070 SUPER will match their tok/s at 7B while costing half as much.
- Your workload fits in integrated GPU or unified memory — an Apple M4 Pro 48 GB is $2,399 and holds models neither of these discrete cards can hold.
- You can't give the card 1.5x its TDP in clean PSU headroom. Undersized PSUs cause transient shutdowns on Blackwell's spike behavior specifically; that's not a card problem, it's a build problem.
Frequently asked questions
Is NVIDIA GeForce RTX 5090 worth the premium over the AMD Radeon RX 7900 XTX?
Only if your workload actually stresses the spec delta. For single-user 7-14B LLM inference the two are often within 20% of each other; for 32-70B where the NVIDIA GeForce RTX 5090's VRAM advantage matters, the premium makes sense. For gaming at 4K Ultra, it depends on the specific game — see the synthetic table.
Which card uses less power under real load?
The NVIDIA GeForce RTX 5090 has a 575W TDP; the AMD Radeon RX 7900 XTX is 355W. Sustained draw during inference is typically 70-90% of rated TDP, so budget your PSU at 1.5x the higher number. PSU headroom matters especially on Blackwell cards because of transient spike behavior.
Which one ages better?
The card with more VRAM ages better. LLMs keep getting bigger; game texture budgets keep growing. If the two are otherwise close, pick the one with more memory.
Do I need a new PSU / case / motherboard?
Check the physical length and the 12V-2×6 / 12VHPWR adapter on each. Both cards require PCIe 5.0 or later for full bandwidth, but will negotiate down to PCIe 4.0 x16 with ~1-3% loss. On older PSUs, use the manufacturer-supplied 12V-2×6 adapter, not a third-party splitter.
Which is better for AI image generation (Flux, SDXL)?
VRAM wins — more memory lets you run Flux.1 fp16 workflows that crash lower-VRAM cards. See our ComfyUI setup guide for workflow-specific VRAM targets.
Sources
- Tom's Hardware GPU Hierarchy
- r/LocalLLaMA (community tok/s threads)
- llama.cpp GitHub Discussions #4167 — Apple Silicon benchmark thread
- Tom's Hardware — RTX 5090 Founders Edition review
- Phoronix — RTX 5080/5090 Linux performance review
