RUNLOCALAIv38
→WILL IT RUNBEST GPUCOMPARETROUBLESHOOTSTARTPULSEMODELSHARDWARETOOLSBENCH
RUNLOCALAI

Operator-grade instrument for local-AI hardware intelligence. Hand-written verdicts. Real benchmarks. Reproducible commands.

OP·Fredoline Eruo
DIR
  • Models
  • Hardware
  • Tools
  • Benchmarks
  • Will it run?
GUIDES
  • Best GPU
  • Best laptop
  • Best Mac
  • Best used GPU
  • Best budget GPU
  • Best GPU for Ollama
  • Best GPU for SD
  • AI PC build $2K
  • CUDA vs ROCm
  • 16 vs 24 GB
  • Compare hardware
  • Custom compare
REF
  • Systems
  • Ecosystem maps
  • Pillar guides
  • Methodology
  • Glossary
  • Errors KB
  • Troubleshooting
  • Resources
  • Public API
EDITOR
  • About
  • About the author
  • Changelog
  • Latest
  • Updates
  • Submit benchmark
  • Send feedback
  • Trust
  • Editorial policy
  • How we make money
  • Contact
LEGAL
  • Privacy
  • Terms
  • Sitemap
MAIL · MONTHLY DIGEST
Get monthly local AI changes
Monthly recap. No spam.
DISCLOSURE

Some links on this site are affiliate links (Amazon Associates and other first-class retailers). When you buy through them, we earn a small commission at no extra cost to you. Affiliate links do not influence our verdicts — there are cards we rate highly that we don't have affiliate relationships with, and cards that sell well that we refuse to recommend. Read more →

SYS · ONLINEUPTIME · 100%2026 · operator-owned
RUNLOCALAI · v38
  1. >
  2. Home
  3. /Hardware
  4. /AMD Radeon 780M (Phoenix iGPU)
UNIT · AMD · GPU
entry·Reviewed May 2026

AMD Radeon 780M (Phoenix iGPU)

AMD's 780M iGPU (Ryzen 7040/8040 series Phoenix). Shares system RAM via unified memory architecture; 32 GB DDR5 system gives effective 16-20 GB usable for inference. Bandwidth (89 GB/s) is the bottleneck — ~6-12 tok/s on 7B Q4. The 'I have a thin laptop' audience can run AI but slowly.

Released 2023·89 GB/s memory bandwidth
RUNLOCALAI SCORE
See full leaderboard →
127/ 1000
DD-tier
Estimated
Throughput
26/ 500
VRAM-fit
0/ 200
Ecosystem
130/ 200
Efficiency
25/ 100

Extrapolated from 89 GB/s bandwidth — 8.9 tok/s estimated. No measured benchmarks yet.

WORKLOAD FIT
Try other hardware →

Plain-English: Doesn't fit modern chat models usefully — vision models won't fit.

7B chat✗
Doesn't fit
14B chat✗
Doesn't fit
32B chat✗
Doesn't fit
70B chat✗
Doesn't fit
Coding agent✗
Doesn't fit
Vision (≤8B VLM)✗
Doesn't fit
Long context (32K)✗
Doesn't fit
✓Comfortable — fits with headroom
~Tight — works, no slack
△Marginal — needs aggressive quant
✗Doesn't fit usefully

Verdicts extrapolated from catalog VRAM + bandwidth + ecosystem flags. Hover any chip for the rationale. Want measured numbers? Submit your own run with runlocalai-bench --submit.

BLK · VERDICT

Our verdict

OP · Fredoline Eruo|VERIFIED MAY 10, 2026
2.1/10

This card is for the operator who already owns a Ryzen 7040/8040 laptop and wants to experiment with local AI without buying a discrete GPU. It runs 7B Q4 models at ~6-12 tok/s — usable for chat but too slow for real-time interaction. 13B models dip to ~3-5 tok/s, and anything larger is impractical. The shared memory architecture means the system RAM (ideally 32 GB) is split between GPU and OS, leaving ~16-20 GB for models. What breaks: any model above 13B, any workload requiring sustained throughput, and ROCm support on iGPUs is still rough — expect manual setup and limited compatibility. Pass if you need interactive speeds or plan to run models larger than 7B; a used RTX 3060 12GB will cost ~$200 and deliver 5-10x the performance. Price/value note: the 780M is free with the CPU, so its value is purely incremental — if you already have the laptop, it's a bonus, not a reason to buy.

›Why this rating

The 780M is a capable iGPU for casual local AI experimentation, but its shared memory and low bandwidth severely limit model size and speed. It earns a 3.0 for being a free add-on that can run small models, but it's not a primary inference card.

BLK · OVERVIEW

Overview

AMD's 780M iGPU (Ryzen 7040/8040 series Phoenix). Shares system RAM via unified memory architecture; 32 GB DDR5 system gives effective 16-20 GB usable for inference. Bandwidth (89 GB/s) is the bottleneck — ~6-12 tok/s on 7B Q4. The 'I have a thin laptop' audience can run AI but slowly.

BLK · SPECS

Specs

VRAM0 GB
Power draw28 W
Released2023
Backends
ROCm
Vulkan

Frequently asked

Does AMD Radeon 780M (Phoenix iGPU) support CUDA?

No — AMD Radeon 780M (Phoenix iGPU) is an AMD card. Use ROCm (Linux) or the Vulkan backend in llama.cpp instead. CUDA-only tools won't work.

Where next?

Buyer guides
  • Best GPU for local AI →
  • Best laptop for local AI →
  • Best Mac for local AI →
  • Best used GPU for local AI →
Troubleshooting
  • CUDA out of memory →
  • Ollama running slowly →
  • ROCm not detected →
  • Model keeps crashing →

Reviewed by RunLocalAI Editorial. See our editorial policy for how we research and verify hardware specifications.

Compare alternatives

Hardware worth comparing

Same VRAM tier and the one step above and below — so you can frame the buying decision against real options.

Same VRAM tier
Cards in the same memory band
  • NVIDIA GeForce GTX 1050 Ti
    nvidia · 4 GB VRAM
    1.3/10
  • NVIDIA GeForce GTX 1650
    nvidia · 4 GB VRAM
    1.8/10
  • AMD Radeon 880M (Strix Point iGPU)
    amd · 0 GB VRAM
    2.4/10
  • NVIDIA GeForce RTX 3050
    nvidia · 8 GB VRAM
    5.3/10
  • NVIDIA GeForce GTX 1060 6GB
    nvidia · 6 GB VRAM
    2.6/10
  • NVIDIA GeForce GTX 1060 3GB
    nvidia · 3 GB VRAM
    1.1/10
Step up
More VRAM — bigger models, more context
  • NVIDIA GeForce GTX 1660
    nvidia · 6 GB VRAM
    2.8/10
  • NVIDIA GeForce RTX 3060 12GB
    nvidia · 12 GB VRAM
    7.0/10
  • NVIDIA GeForce GTX 1070 Ti
    nvidia · 8 GB VRAM
    5.1/10
Step down
Less VRAM — cheaper, more constrained
No verdicted hardware in the next tier down yet.