IPEX-LLM
Intel's PyTorch extension for low-bit LLM inference on Intel GPUs / CPUs / NPUs. Strongest community-supported path for running LLMs on Intel Arc A770 / B580 and on Lunar Lake NPUs. Compatible with Hugging Face Transformers + LangChain + Ollama-on-Intel.
Overview
Intel's PyTorch extension for low-bit LLM inference on Intel GPUs / CPUs / NPUs. Strongest community-supported path for running LLMs on Intel Arc A770 / B580 and on Lunar Lake NPUs. Compatible with Hugging Face Transformers + LangChain + Ollama-on-Intel.
Pros
- First-class Intel Arc GPU support — fills the gap left by vLLM / llama.cpp
- PyTorch-native — Hugging Face checkpoints work directly
- Active Intel maintenance — kernel optimizations land regularly
Cons
- Intel-only — doesn't help on NVIDIA / Apple / AMD
- Documentation density behind the mainline runtimes
- Community size smaller than the NVIDIA-centric runtimes
Compatibility
| Operating systems | Linux Windows |
| GPU backends | Intel Arc GPU Intel CPU Intel NPU |
| License | Open source · free + open-source |
Runtime health
Operator-grade signals on how actively IPEX-LLM is being maintained, how fresh its measurements are, and what failure classes operators have flagged. Every label below is anchored to a real date or count — we never infer maintainer activity we can't show.
Release cadence
Derived from the most recent editorial signal on this row.
6 days since last refresh · source: lastUpdated
Benchmark freshness
How recent the editorial measurements on this runtime are.
No editorial benchmarks for this runtime yet.
Community reproduction
Submissions that match an editorial measurement on similar hardware.
No community reproductions on file yet.
Get IPEX-LLM
Frequently asked
Is IPEX-LLM free?
What operating systems does IPEX-LLM support?
Which GPUs work with IPEX-LLM?
Reviewed by RunLocalAI Editorial. See our editorial policy for how we evaluate tools.
Related — keep moving
Verify IPEX-LLM runs on your specific hardware before committing money.