Which old AMD GPU should I get for Local LLMs on an eGPU?
1. Since the compute should all be happening on the card, an eGPU should be fine. The slow connection should only be a factor if you try running a model that exceeds the VRAM capacity of the card, but that's a bad idea in general - being an eGPU will just make it even worse.
2. PCIe 3.0 should be fine - again, as long as you don't run out of VRAM, this shouldn't be a big factor in compute performance.
3. The reason people aren't doing these comparisons with old AMD cards is that old AMD cards had terrible AI performance. Until this generation, AMD's AI compute was abysmal, and they still haven't caught up to Nvidia, not to mention that the majority of AI applications still expect CUDA. You can certainly find comparisons with the most recent AMD cards - LTT's reviews of the 9060XT and 9070 series had AI benchmarks, for instance. And there are LLMs that can work on AMD. However, you're not going to get good performance on them unless you're buying something from this generation.
Asking to run AI on old AMD is like asking to run AI on old Nvidia - as in, 10 series and older. Even the 1080 Ti, the best consumer Nvidia graphics card without Tensor cores, gets curb-stomped by anything from the 20 series and newer in AI performance. Old AMD does likewise, except in some ways even worse, because they lack CUDA, so some applications just refuse to work. Here's an article from Puget Systems that shows a 1080 Ti against other cards for AI - it ain't pretty.
Create an account or sign in to comment
You need to be a member in order to leave a comment
Create an account
Sign up for a new account in our community. It's easy!
Register a new accountSign in
Already have an account? Sign in here.
Sign In Now