Jump to content

OpenAI ChatGPT Server

Budget (including currency): $2500

Country: USA

Games, programs or workloads that it will be used for: OpenAI

Other details (existing parts lists, whether any peripherals are needed, what you're upgrading from, when you're going to buy, what resolution and refresh rate you want to play at, etc): 

I asked the chat server what it would take to build a openai chat-gpt, and below is the response:  

 

"To create a ChatGPT model within Ubuntu 22.04, you will need to have a machine with a CUDA-enabled GPU, as well as the following software installed:

Python 3.7 or later
CUDA 11.0 or later
cuDNN 8.0 or later
The Hugging Face transformers library
The torch library
Once you have all the necessary software installed, you can use the Hugging Face transformers library to fine-tune a pre-trained ChatGPT model on a dataset of your choice.
You will also need to download the pre-trained weights for GPT-2 or GPT-3 models from Hugging Face model hub or OpenAI API.
You can use the provided code examples and documentation from the transformers library to train and fine-tune the model to your specific use case.
It's worth noting that GPT-3 models are quite large and resource-intensive, so it might take a powerful machine with a large amount of memory and storage to train them.

Please note that creating a GPT model requires a good knowledge of deep learning and computer science, if you are new to this field I recommend to start with a simple project or to follow a tutorial that can guide you through the process."

 

From that response, I started to look at GPU's with CUDA-enabled, however quick to think I have no idea what I am looking at much less what to look for on price and specs.  At the end of this month is the last semester of college, studying python, how cool would it be to have figure this out by then.  

 

JD

Link to comment
Share on other sites

Link to post
Share on other sites

doesn't GPT-3 need like 1 terabyte of vram?

If your question is answered, mark it so.  | It's probably just coil whine, and it is probably just fine |   LTT Movie Club!

Read the docs. If they don't exist, write them. | Professional Thread Derailer

Desktop: i7-8700K, RTX 2080, 16G 3200Mhz, EndeavourOS(host), win10 (VFIO), Fedora(VFIO)

Server: ryzen 9 5900x, GTX 970, 64G 3200Mhz, Unraid.

 

Link to comment
Share on other sites

Link to post
Share on other sites

4 minutes ago, Takumidesh said:

doesn't GPT-3 need like 1 terabyte of vram?

From some googling, looks like ~300-350GB VRAM for the smallest version of it, so yeah closer to 1TB for the largest sounds about right. I don't think you can get a single 48GB VRAM card for $2500, let alone the multiple OP would need. 

Intel HEDT and Server platform enthusiasts: Intel HEDT Xeon/i7 Megathread 

 

Main PC 

CPU: i9 7980XE @4.5GHz/1.22v/-2 AVX offset 

Cooler: EKWB Supremacy Block - custom loop w/360mm +280mm rads 

Motherboard: EVGA X299 Dark 

RAM:4x8GB HyperX Predator DDR4 @3200Mhz CL16 

GPU: Nvidia FE 2060 Super/Corsair HydroX 2070 FE block 

Storage:  1TB MP34 + 1TB 970 Evo + 500GB Atom30 + 250GB 960 Evo 

Optical Drives: LG WH14NS40 

PSU: EVGA 1600W T2 

Case & Fans: Corsair 750D Airflow - 3x Noctua iPPC NF-F12 + 4x Noctua iPPC NF-A14 PWM 

OS: Windows 11

 

Display: LG 27UK650-W (4K 60Hz IPS panel)

Mouse: EVGA X17

Keyboard: Corsair K55 RGB

 

Mobile/Work Devices: 2020 M1 MacBook Air (work computer) - iPhone 13 Pro Max - Apple Watch S3

 

Other Misc Devices: iPod Video (Gen 5.5E, 128GB SD card swap, running Rockbox), Nintendo Switch

Link to comment
Share on other sites

Link to post
Share on other sites

This is my reasoning behind building my own system, when to the website, and typed in the question and received the response  "An error occurred. If this issue persists please contact us through our help center at help.openai.com."  , high volume of people are trying to access the system, thus making it unavailable to the average user.  

Link to comment
Share on other sites

Link to post
Share on other sites

HP L40549-001 8GB Nvidia Quadro RTX 4000 Turing Graphics Card - GDDR6 - 256-Bit - 2304 Cuda Cores - PCIe 3.0 X16 - 1005 MHz - Multi Monitor (Renewed) with tax was $730 on amazon.  

Link to comment
Share on other sites

Link to post
Share on other sites

5 minutes ago, AllGoodNamesRGone said:

HP L40549-001 8GB Nvidia Quadro RTX 4000 Turing Graphics Card - GDDR6 - 256-Bit - 2304 Cuda Cores - PCIe 3.0 X16 - 1005 MHz - Multi Monitor (Renewed) with tax was $730 on amazon.  

so at that price it is looking like $27,000 for the smallest version of gpt-3

 

edit: just for the GPUs

 

additionally, I don't think openAI will give you the model for personal use, and you won't be able to train a model like it without the HUGE dataset thatt they had. keep in mind you need the hundreds of gigs of vram just to run the model not train it.

 

 

If you want to learn about AI and neural networks you don't need anything other than the computer you have now. you can start training basic networks with a cpu.

 

If your question is answered, mark it so.  | It's probably just coil whine, and it is probably just fine |   LTT Movie Club!

Read the docs. If they don't exist, write them. | Professional Thread Derailer

Desktop: i7-8700K, RTX 2080, 16G 3200Mhz, EndeavourOS(host), win10 (VFIO), Fedora(VFIO)

Server: ryzen 9 5900x, GTX 970, 64G 3200Mhz, Unraid.

 

Link to comment
Share on other sites

Link to post
Share on other sites

1 hour ago, AllGoodNamesRGone said:

I started to look at GPU's with CUDA-enabled

That's any Nvidia GPU.

1 hour ago, AllGoodNamesRGone said:

At the end of this month is the last semester of college, studying python, how cool would it be to have figure this out by then.  

By college you mean high school or university? If university, then it worries me a bit lol

 

Otherwise, if you're in high school, start learning about machine learning and make use of google colab to learn, no need to build a new system if you have no idea about what you're doing, it'll be just a waste of money and time.

 

FWIW, GPT-3 is a 800GB model, you can't train/fine-tune it in a regular computer, but will rather need a cluster of those with many V100s/A100s/H100s. Things like GPT-j do exist and you can fine tune those for your need, but without grasping the basics of neural networks you won't go very far.

 

Start off by toying with something like MNIST and go from there.

 

 

For the curious, here are some time/cost estimations for GPT-3:

 

Over 1000 A100s running full tilt for over a month, around $5M in cost.

 

FX6300 @ 4.2GHz | Gigabyte GA-78LMT-USB3 R2 | Hyper 212x | 3x 8GB + 1x 4GB @ 1600MHz | Gigabyte 2060 Super | Corsair CX650M | LG 43UK6520PSA
ASUS X550LN | i5 4210u | 12GB
Lenovo N23 Yoga

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×