Jump to content

Which old AMD GPU should I get for Local LLMs on an eGPU?

Go to solution Solved by YoungBlade,

1. Since the compute should all be happening on the card, an eGPU should be fine. The slow connection should only be a factor if you try running a model that exceeds the VRAM capacity of the card, but that's a bad idea in general - being an eGPU will just make it even worse.

 

2. PCIe 3.0 should be fine - again, as long as you don't run out of VRAM, this shouldn't be a big factor in compute performance.

 

3. The reason people aren't doing these comparisons with old AMD cards is that old AMD cards had terrible AI performance. Until this generation, AMD's AI compute was abysmal, and they still haven't caught up to Nvidia, not to mention that the majority of AI applications still expect CUDA. You can certainly find comparisons with the most recent AMD cards - LTT's reviews of the 9060XT and 9070 series had AI benchmarks, for instance. And there are LLMs that can work on AMD. However, you're not going to get good performance on them unless you're buying something from this generation.

 

Asking to run AI on old AMD is like asking to run AI on old Nvidia - as in, 10 series and older. Even the 1080 Ti, the best consumer Nvidia graphics card without Tensor cores, gets curb-stomped by anything from the 20 series and newer in AI performance. Old AMD does likewise, except in some ways even worse, because they lack CUDA, so some applications just refuse to work. Here's an article from Puget Systems that shows a 1080 Ti against other cards for AI - it ain't pretty.

TLDR: I want to buy a used AMD GPU for AI. I don't know which is best to buy or how to compare cards. I am also limited to PCIe 3.0 in an external chassis

I'm currently building a new pc and I have an external GPU chassis that I want to put a card in for dedicated AI loads. I am using AMD GPUs in my pc so I would like an AMD gpu in the eGPU as well. Most of my local AI I am currently using is used for DND NPCs and keeping an output speed like an IRL conversation is important to me. I realize I can just change the AI parameters, but I want to maximize value with the card I pick. I'm unsure which card to get for a few reasons:
1. Since this is an external GPU Chassis, does that mean the speed will be too slow for simulating conversations? I used to use the eGPU for video games on a laptop and didn't have issues with the information transferred over a cable so I'm assuming it won't be an issue, but I would love to know more. 

 

2. The GPU Chassis is PCIe 3.0, I saw some reddit posts saying that cards with PCIe 4.0 or higher will work in PCIe 3.0. But since I care mostly about the compute speed, does this mean a PCIe 4.0 card will just run the same as a PCIe 3.0 card for local AIs? Is getting a PCIe 4.0 or higher card just a waste of money for me?

 

3. I have no idea how to compare AMD cards for AI. I see a lot of AMD Instinct cards on ebay for about $50-$100 (I have no problem with used cards). I don't know how these would compare to getting an old GPU for a similar price. If this was an Nvidia card, I could find graphs with Tokens/second comparisons such as 2080s and K40s. Seems a lot of Youtubers and articles just don't do the same for AMD, probably because AMD has been unusable for AI for a while. If someone can provide some context on which metric to use for best comparing AI GPUs to regular GPUs that'd be great. I've mostly been looking at this wiki for comparisons, but I don't really know which metric listed is the best for comparing price to performance. 

Asking for a lot of stuff I don't know, I appreciate your response in advance 🙂

 

Link to post
Share on other sites

AMD is still basically unusable for AI at this time. packages either don't support AMD, or run way worse. Nvidia's CUDA cores are the way to go.

I might be experienced, but I'm human and I do make mistakes. Trust but Verify! I edit my messages after sending them alot, please refresh before posting your reply. Please try to be clear and specific, you'll get a better answer. Please remember to mark solutions once you have the information you need. Expand this signature for common PC building advice, a short bio and a list of my components.

 

Common build advice:

1) Buy the cheapest (well reviewed) motherboard that has the features you need. Paying more typically only gets you features you won’t use. 2) only get as much RAM as you need, getting more won’t (typically) make your PC faster. 3) While I recommend getting an NVMe drive, you don’t need to splurge for an expensive drive with DRam cache, DRamless drives are fine for gamers. 4) paying for looks is fine, just don’t break the bank. 5) Tower coolers are usually good enough, unless you go top tier Intel or plan on OCing. 6) OCing is a dead meme, you probably shouldn’t bother. 7) "Bottlenecks" rarely matter and "Future-proofing" is a myth. 8) AIOs don't noticably improve performance past 240mm.

 

Useful Websites:

https://www.productchart.com - helps compare monitors, https://uk.pcpartpicker.com - makes designing a PC easier.

 

Bio:

He/Him - I'm a PhD student working in the fields of reinforcement learning and traffic control. PCs are one of my hobbies and I've built many PCs and performed upgrades on a few laptops (for myself, friends and family). My personal computers include 4 windows (10/11) machines and a TrueNAS server (and I'm looking to move to dual booting Linux Mint on my main machine in future). Aside from computers, I also dabble in modding/homebrew retro consoles, support Southampton FC, and enjoy Scuba Diving and Skiing.

Fun Facts

1) When I was 3 years old my favourite toy was a scientific calculator. 2) My father is a British Champion ploughman in the Vintage Hydraulic Class. 3) On Speedrun.com, I'm the world record holder for the Dream Bobsleigh event on Mario & Sonic at the Olympic Winter Games 2010.

 

My Favourite Games: World of Tanks, Runescape, Subnautica, Metroid (Fusion and Dread), Spyro: Year of the Dragon (Original and Reignited Trilogy), Crash Bash, Mario Kart Wii, Balatro

 

My Computers: Primary: My main gaming rig - https://uk.pcpartpicker.com/user/will0hlep/saved/NByp3C Second: Hosts Discord bots as well as a Minecraft and Ark server, and also serves as a reinforcement learning sand box - https://uk.pcpartpicker.com/user/will0hlep/saved/cc9K7P NAS: TrueNAS Scale NAS hosting SMB shares, DDNS updater, pi-hole, and a Jellyfin server - https://uk.pcpartpicker.com/user/will0hlep/saved/m37w3C Foldatron: My folding@home and BOINC rig (partially donated to me by Folding Team Leader GOTSpectrum) - Mobile: Mini-ITX gaming rig for when I'm away from home -

Link to post
Share on other sites

Could you be a bit more precise about budget and intended use ? Which LLM, what kind of power you need ?

And budget ?

But I agree with @will0hlep an old AMD GPU will be crap for AI

And why an eGPU ? Using a chassis makes sense for laptops, but not if you plan to build a desktop...

AMD R9  7950X3D CPU/ Asus ROG STRIX X670E-E board/ 2x32GB G-Skill Trident Z Neo 6000CL30 RAM ASUS TUF Gaming AMD Radeon RX 7900 XTX OC Edition GPU/ Phanteks P600S case /  Arctic Liquid Freezer III 360 ARGB cooler/  2TB WD SN850 NVme + 2TB Crucial T500  NVme  + 4TB Toshiba X300 HDD / Corsair RM850x PSU/ Alienware AW3420DW 34" 120Hz 3440x1440p monitor / ASUS ROG AZOTH keyboard/ Logitech G PRO X Superlight mouse / Audeze Maxwell headphones

Link to post
Share on other sites

1. Since the compute should all be happening on the card, an eGPU should be fine. The slow connection should only be a factor if you try running a model that exceeds the VRAM capacity of the card, but that's a bad idea in general - being an eGPU will just make it even worse.

 

2. PCIe 3.0 should be fine - again, as long as you don't run out of VRAM, this shouldn't be a big factor in compute performance.

 

3. The reason people aren't doing these comparisons with old AMD cards is that old AMD cards had terrible AI performance. Until this generation, AMD's AI compute was abysmal, and they still haven't caught up to Nvidia, not to mention that the majority of AI applications still expect CUDA. You can certainly find comparisons with the most recent AMD cards - LTT's reviews of the 9060XT and 9070 series had AI benchmarks, for instance. And there are LLMs that can work on AMD. However, you're not going to get good performance on them unless you're buying something from this generation.

 

Asking to run AI on old AMD is like asking to run AI on old Nvidia - as in, 10 series and older. Even the 1080 Ti, the best consumer Nvidia graphics card without Tensor cores, gets curb-stomped by anything from the 20 series and newer in AI performance. Old AMD does likewise, except in some ways even worse, because they lack CUDA, so some applications just refuse to work. Here's an article from Puget Systems that shows a 1080 Ti against other cards for AI - it ain't pretty.

Link to post
Share on other sites

1 hour ago, ImLuciano said:

TLDR: I want to buy a used AMD GPU for AI. I don't know which is best to buy or how to compare cards. I am also limited to PCIe 3.0 in an external chassis

I'm currently building a new pc and I have an external GPU chassis that I want to put a card in for dedicated AI loads. I am using AMD GPUs in my pc so I would like an AMD gpu in the eGPU as well. Most of my local AI I am currently using is used for DND NPCs and keeping an output speed like an IRL conversation is important to me. I realize I can just change the AI parameters, but I want to maximize value with the card I pick. I'm unsure which card to get for a few reasons:
1. Since this is an external GPU Chassis, does that mean the speed will be too slow for simulating conversations? I used to use the eGPU for video games on a laptop and didn't have issues with the information transferred over a cable so I'm assuming it won't be an issue, but I would love to know more. 

 

2. The GPU Chassis is PCIe 3.0, I saw some reddit posts saying that cards with PCIe 4.0 or higher will work in PCIe 3.0. But since I care mostly about the compute speed, does this mean a PCIe 4.0 card will just run the same as a PCIe 3.0 card for local AIs? Is getting a PCIe 4.0 or higher card just a waste of money for me?

 

3. I have no idea how to compare AMD cards for AI. I see a lot of AMD Instinct cards on ebay for about $50-$100 (I have no problem with used cards). I don't know how these would compare to getting an old GPU for a similar price. If this was an Nvidia card, I could find graphs with Tokens/second comparisons such as 2080s and K40s. Seems a lot of Youtubers and articles just don't do the same for AMD, probably because AMD has been unusable for AI for a while. If someone can provide some context on which metric to use for best comparing AI GPUs to regular GPUs that'd be great. I've mostly been looking at this wiki for comparisons, but I don't really know which metric listed is the best for comparing price to performance. 

Asking for a lot of stuff I don't know, I appreciate your response in advance 🙂

 

AMD isn't good for AI unless you get a newer one (RDNA 4) but you said you want an older one.

AMD Ryzen™ 5 5600g w/ Radeon Graphics | 16GB DDR4-3200 RAM | 256GB NVME SSD + 2TB HDD | Amazon Basics 2.0 Speakers

                                                                                            I'M JUST A REAL-LIFE TOM SAWYER

Link to post
Share on other sites

44 minutes ago, PDifolco said:

Could you be a bit more precise about budget and intended use ? Which LLM, what kind of power you need ?

And budget ?

But I agree with @will0hlep an old AMD GPU will be crap for AI

And why an eGPU ? Using a chassis makes sense for laptops, but not if you plan to build a desktop...

LLM is flexible, most of what I've done so far is all entry level and can be moved to different models. I plan to find which model works best once I've finished the setup. Budget is less than $200, I was planning to get an older AI GPU off ebay, but it seems like old generations aren't worth it from the replies I've seen here.
I am using an eGPU because it is what I already have. I don't have extra space in my new Case for a card specific to AI.

Link to post
Share on other sites

34 minutes ago, YoungBlade said:

1. Since the compute should all be happening on the card, an eGPU should be fine. The slow connection should only be a factor if you try running a model that exceeds the VRAM capacity of the card, but that's a bad idea in general - being an eGPU will just make it even worse.

 

2. PCIe 3.0 should be fine - again, as long as you don't run out of VRAM, this shouldn't be a big factor in compute performance.

 

3. The reason people aren't doing these comparisons with old AMD cards is that old AMD cards had terrible AI performance. Until this generation, AMD's AI compute was abysmal, and they still haven't caught up to Nvidia, not to mention that the majority of AI applications still expect CUDA. You can certainly find comparisons with the most recent AMD cards - LTT's reviews of the 9060XT and 9070 series had AI benchmarks, for instance. And there are LLMs that can work on AMD. However, you're not going to get good performance on them unless you're buying something from this generation.

 

Asking to run AI on old AMD is like asking to run AI on old Nvidia - as in, 10 series and older. Even the 1080 Ti, the best consumer Nvidia graphics card without Tensor cores, gets curb-stomped by anything from the 20 series and newer in AI performance. Old AMD does likewise, except in some ways even worse, because they lack CUDA, so some applications just refuse to work. Here's an article from Puget Systems that shows a 1080 Ti against other cards for AI - it ain't pretty.

I was under the impression most of what AMD was lacking was on the driver side and therefore older cards would get to a usable state when newer cards were optimized. Clearly not the case. 

Link to post
Share on other sites

3 minutes ago, ImLuciano said:

I was under the impression most of what AMD was lacking was on the driver side and therefore older cards would get to a usable state when newer cards were optimized. Clearly not the case. 

Yeah, No. AMD's older cards are lacking the hardware to be useful for AI work.

I might be experienced, but I'm human and I do make mistakes. Trust but Verify! I edit my messages after sending them alot, please refresh before posting your reply. Please try to be clear and specific, you'll get a better answer. Please remember to mark solutions once you have the information you need. Expand this signature for common PC building advice, a short bio and a list of my components.

 

Common build advice:

1) Buy the cheapest (well reviewed) motherboard that has the features you need. Paying more typically only gets you features you won’t use. 2) only get as much RAM as you need, getting more won’t (typically) make your PC faster. 3) While I recommend getting an NVMe drive, you don’t need to splurge for an expensive drive with DRam cache, DRamless drives are fine for gamers. 4) paying for looks is fine, just don’t break the bank. 5) Tower coolers are usually good enough, unless you go top tier Intel or plan on OCing. 6) OCing is a dead meme, you probably shouldn’t bother. 7) "Bottlenecks" rarely matter and "Future-proofing" is a myth. 8) AIOs don't noticably improve performance past 240mm.

 

Useful Websites:

https://www.productchart.com - helps compare monitors, https://uk.pcpartpicker.com - makes designing a PC easier.

 

Bio:

He/Him - I'm a PhD student working in the fields of reinforcement learning and traffic control. PCs are one of my hobbies and I've built many PCs and performed upgrades on a few laptops (for myself, friends and family). My personal computers include 4 windows (10/11) machines and a TrueNAS server (and I'm looking to move to dual booting Linux Mint on my main machine in future). Aside from computers, I also dabble in modding/homebrew retro consoles, support Southampton FC, and enjoy Scuba Diving and Skiing.

Fun Facts

1) When I was 3 years old my favourite toy was a scientific calculator. 2) My father is a British Champion ploughman in the Vintage Hydraulic Class. 3) On Speedrun.com, I'm the world record holder for the Dream Bobsleigh event on Mario & Sonic at the Olympic Winter Games 2010.

 

My Favourite Games: World of Tanks, Runescape, Subnautica, Metroid (Fusion and Dread), Spyro: Year of the Dragon (Original and Reignited Trilogy), Crash Bash, Mario Kart Wii, Balatro

 

My Computers: Primary: My main gaming rig - https://uk.pcpartpicker.com/user/will0hlep/saved/NByp3C Second: Hosts Discord bots as well as a Minecraft and Ark server, and also serves as a reinforcement learning sand box - https://uk.pcpartpicker.com/user/will0hlep/saved/cc9K7P NAS: TrueNAS Scale NAS hosting SMB shares, DDNS updater, pi-hole, and a Jellyfin server - https://uk.pcpartpicker.com/user/will0hlep/saved/m37w3C Foldatron: My folding@home and BOINC rig (partially donated to me by Folding Team Leader GOTSpectrum) - Mobile: Mini-ITX gaming rig for when I'm away from home -

Link to post
Share on other sites

Get a 2060 Super if you want the cheapest GPU with 8GB and tensor cores, or go for a 3060 if you want 12GB of vram.

 

If you are ok with slower performance but want to run larger models, get a P40 or even a P100.

FX6300 @ 4.2GHz | Gigabyte GA-78LMT-USB3 R2 | Hyper 212x | 3x 8GB + 1x 4GB @ 1600MHz | Gigabyte 2060 Super | Corsair CX650M | LG 43UK6520PSA
ASUS X550LN | i5 4210u | 12GB
Lenovo N23 Yoga

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×