Jump to content

Budget Build for deploy Large-Language Model locally?

Alkaid

Budget (including currency): ideally $500, and the maximum is $1500.

Country: USA

Games, programs or workloads that it will be used for: LLM

 

I want to deploy a local chatbot, which requires a large-ram (my target is 24GB) Nvidia GPU. My main desktop is an ITX build with Arc A770 GPU, which is not an ideal GPU for this task and has no space for a second GPU.

 

So I either

 

(1) swap the current GPU with an RTX3090/4090/4060Ti

Because I need to run many daily tasks on my main desktop, I require some gaming performance and stability, which implies a relatively new GeForce GPU. But new large-ram GeForce GPUs are pretty expensive. Only 3090 (Ti)and 4090 have 24GB memory. 4090 is just expensive. A second-hand 3090 is about $700 now, but I do worry about its reliability considering it might have been used for mining. I've had a bad experience with second-hand card before, so I am hesitant to plug a second-hand card into my main machine.

 

4060ti 16GB is also on the table if the 24GB plan is not very achievable.

 

OR

(2) build a dedicated PC with Nvidia Tesla P40

The benefit of this is that I can use Nvidia Tesla P40, the cheapest 24GB Nvidia card. ( about $200) This card does not have video output and is very old (unreliable), but totally fine for AI usage. The issue is I need to buy a full set of CPU, MB, RAM, PSU, etc. I used 10 minutes and was able to find i7-9700+32GB ram+mb+psu+case for about $400. Is there a lot of room for price optimization? And I am also worried about the reliability of Tesla P40, a card released in 2016, nearly a decade ago. I am not a big fan of the second-hand market, and my first DIY build was Intel 10th Gen, so I have very limited experience with old hardware.

 

Which is better? Or is there any other plan I should look for?

 

 

 

 

Main Desktop:

Case:Jonsbo A4

CPU: Intel Core i9-12900K

GPU: Intel Arc A770 Limited Edition

CPU Cooling: Asus ROG Strix LC II 240mm AIO, with 4x Noctua A12x25 Fan.

Motherboard: Gigabyte Z690i Aorus Ultra DDR4

RAM: G.Skill DDR4 3600CL18 32GBx2
System Drive: Samsung 970Pro 1TB.

File Drive: Samsung 980Pro 2TB.

Media Drive: Intel S4510 3.84TBx2 (RAID 0)

PSU: Silverstone SX650G 650W SFX

 

Mini Home Server:

Case: Akasa Plato X8

CPU: Intel i5-8259U

Cooling: Passive Cooling Case

Motherboard: NUC8 i5BEH

RAM: 2400Mhz 16GBx2

Drive: Samsung 970Pro 512GB.

 

NAS :

Case: Jonsbo N1

CPU: Intel Xeon D-1521

Motherboard: X10SDV-4C-TLN2F

RAM: Micron 2400MHz ECC RDIMM 32GBx4

Pool 1: Data: WestDigit Red Plus 8TBx5 (RAID-Z1) Metadata: Intel P1600X 118GBx2 (Mirror)

Pool 2: Intel 670P 2TBx3 (RAID-Z1)

PSU: Silverstone SX300-B SFX

 

Mobile Devices: Microsoft Surface Laptop 4 (13.5-inch edition) | Microsoft Surface Go 2 | Samsung Galaxy Z Flip 4 | Samsung Galaxy Watch 5 | Microsoft Surface Headphones | Sony Linkbuds S | Sony Linkbuds

Link to comment
Share on other sites

Link to post
Share on other sites

Used 3090 + psu upgrade. The performance of that tesla is crap and the vram amount wont make up for that. You also need to make a custom fan for it because ti DOES NOT have any way to cool itself. It is designed to be in a server airflow config.

Link to comment
Share on other sites

Link to post
Share on other sites

Yeah I'd agree with the guy above me, go with the second hand 3090 (TI) you'll save a lot of money and the Tesla is really only an option for a server and the performance is probably worse anyway.

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×