Jump to content

Used Tesla cards, are they any good?

Aleksey
Go to solution Solved by igormp,
1 hour ago, Aleksey said:

Yes, this seems to be a better option. I ditched P40 idea. It has no value. I decided to go with 3090. It has 24GB of RAM. Two 3060 may be a bit cheaper but it is slower than one 3090 and it may require some software tweaks.

As a 3090 owner who pretty much only uses it for ML, that was a nice pick haha

I'll likely pick a second 3090 next month or so 馃檪

There are some used Nvidia Tesla cards on the market at relatively low price, like $200-$300, some of them seem to have come from China. Are they OK cards? Will they work for machine learning tasks?

Link to comment
Share on other sites

Link to post
Share on other sites

Yes, but the use cases for them require the support for them specifically. They also require a forced air cooling solution to run in a desktop. I had a handful of Tesla T4 cards but sold them as they werent as useful as I had hoped for in widows, I think linux has better support.

For the money a used quadro P series, or Quadro RTX works well. A p5000 is great with its 16gb of VRAM.

Link to comment
Share on other sites

Link to post
Share on other sites

Thanks. I use Linux so that's OK. Specifically the Tesla P40 cards with 24Gb of RAM seems to be very interesting.

Link to comment
Share on other sites

Link to post
Share on other sites

14 minutes ago, Aleksey said:

Thanks. I use Linux so that's OK. Specifically the Tesla P40 cards with 24Gb of RAM seems to be very interesting.

As said above they are quite hard to get running properly in a mainstream system,聽 You have to bear I mind that they usually have blowymatrons cooling them which is loud.

CPU : Ryzen 7 7800X3D @ -18mv all core except -13mv on Core 5 because its a pig.

CPU Cooler : Deepcool AK620 Zero Dark

Mobo : MSI B650M-A Wifi MATX

Ram : 32GB (2X16GB) Corsair Vengeance DDR5 6000MHZ CL34

GPU : Reference Design RX7900XT sold by Saphire running at 1050MV undervolt and +15% PL (355w)

Storage : 1TB WD SN770 + 2TB Samsung 970 Evo

PSU : Corsair HX750w Platinum

Case : Asus Prime AP201 All Mesh MATX

Case Fans : Arctic p12's everywhere i can fit them in , 7 In total.

Monitor : LG 27GP850-B.BEK 1440p Nano IPS 180Hz

Keyboard :聽HyperX Alloy Core RGB

Mouse : Corsair M65 Elite RGB

Headset : Corsair HS35 Gaming Headset

Link to comment
Share on other sites

Link to post
Share on other sites

10 hours ago, Bagzie said:

As said above they are quite hard to get running properly in a mainstream system,聽 You have to bear I mind that they usually have blowymatrons cooling them which is loud.

There are many adapter models that can be 3D printed. E.g. -聽https://www.yeggi.com/q/nvidia+tesla+p100+p40+k80+by/聽An alternative would be to pay Amazon over $1,000/month. I can build an entire computer from used parts for these money.

Link to comment
Share on other sites

Link to post
Share on other sites

19 hours ago, Aleksey said:

Thanks. I use Linux so that's OK. Specifically the Tesla P40 cards with 24Gb of RAM seems to be very interesting.

Yeah, those should be plug and play on linux (after you install the driver ofc), but since you mentioned you'll be doing ML, be aware that a 3060 will be faster than it and be able to work on the same amount of data since it has proper half precision support (which the p40 lacks).

Turing/volta were a big game changer for ML with their tensor cores.

FX6300 @ 4.2GHz | Gigabyte GA-78LMT-USB3 R2 | Hyper 212x | 3x 8GB + 1x 4GB @ 1600MHz | Gigabyte 2060 Super | Corsair CX650M | LG 43UK6520PSA
ASUS X550LN | i5 4210u | 12GB
Lenovo N23 Yoga

Link to comment
Share on other sites

Link to post
Share on other sites

19 hours ago, Bagzie said:

blowymatrons

I'm using this term someday. I don't know how or why, but I will randomly use the word "blowymatron" at some point in my life.

Aerocool DS are the best fans you've never tried.

Link to comment
Share on other sites

Link to post
Share on other sites

4 hours ago, igormp said:

Yeah, those should be plug and play on linux (after you install the driver ofc), but since you mentioned you'll be doing ML, be aware that a 3060 will be faster than it and be able to work on the same amount of data since it has proper half precision support (which the p40 lacks).

Why? According to this data聽it supports half precision. According to this comparison chart, it is indeed slower, but there is a catch - 3060 only has 12GB of RAM, while P40 has 24GB, which is important for running modern AI models. Most LLM do not fit into 12GB.

Link to comment
Share on other sites

Link to post
Share on other sites

Actually if we compare P40 specification聽and 3060 specification聽the former is almost 15 times faster when using half precision - 183.7 GFLOPS聽vs 12.74 GFLOPS. Furthermore - 3060 does not give any performance advantage for half precision. This look too good to be true. I am not even sure if the data is correct.

Link to comment
Share on other sites

Link to post
Share on other sites

1 hour ago, Aleksey said:

Why? According to this data聽it supports half precision

Notice how it's 64 times slower than regular fp32, while on volta/turing and newer fp16 is much faster than fp32.

1 hour ago, Aleksey said:

According to this comparison chart, it is indeed slower,

That's not a credible source for comparisons.

One practical comparison is for the stable diffusion benchmarks, notice how a 3060 is 2~3x faster than a P40:

https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html

1 hour ago, Aleksey said:

but there is a catch - 3060 only has 12GB of RAM, while P40 has 24GB, which is important for running modern AI models. Most LLM do not fit into 12GB.

Given how important quantization is, you won't benefit from it, so you will have to load fp32 models using 24gb, while you could load the exact same model with fp16 using only 12gb, or even lower precision such as int8 and even int4, using 6/3gb ov vram (or making use of that to use bigger models).

At that price point you could also buy 2x used 3060s for a total of 24gb of vram and much faster compute.

1 hour ago, Aleksey said:

Actually if we compare P40 specification聽and 3060 specification聽the former is almost 15 times faster when using half precision - 183.7 GFLOPS聽vs 12.74 GFLOPS. Furthermore - 3060 does not give any performance advantage for half precision. This look too good to be true. I am not even sure if the data is correct.

It is true, and you can take a look at Ampere's whitepaper, those tensor cores are magic 馃檪

1 hour ago, Aleksey said:

I think you mixed it up with M40. M40 does not support half precision, P40 - does.

See explanation above. Anything older than volta is pretty much useless for modern ML unless you're really tight on budget or don't have any other options.

FX6300 @ 4.2GHz | Gigabyte GA-78LMT-USB3 R2 | Hyper 212x | 3x 8GB + 1x 4GB @ 1600MHz | Gigabyte 2060 Super | Corsair CX650M | LG 43UK6520PSA
ASUS X550LN | i5 4210u | 12GB
Lenovo N23 Yoga

Link to comment
Share on other sites

Link to post
Share on other sites

19 hours ago, igormp said:

Notice how it's 64 times slower than regular fp32, while on volta/turing and newer fp16 is much faster than fp32.

21 hours ago, Aleksey said:

OMG, that Gigaflops compared to Teraflops. That's bad...

Link to comment
Share on other sites

Link to post
Share on other sites

On 10/7/2023 at 2:03 AM, igormp said:

Given how important quantization is, you won't benefit from it, so you will have to load fp32 models using 24gb, while you could load the exact same model with fp16 using only 12gb, or even lower precision such as int8 and even int4, using 6/3gb ov vram (or making use of that to use bigger models).

At that price point you could also buy 2x used 3060s for a total of 24gb of vram and much faster compute.

Yes, this seems to be a better option. I ditched P40 idea. It has no value. I decided to go with 3090. It has 24GB of RAM. Two 3060 may be a bit cheaper but it is slower than one 3090 and it may require some software tweaks.

Link to comment
Share on other sites

Link to post
Share on other sites

1 hour ago, Aleksey said:

Yes, this seems to be a better option. I ditched P40 idea. It has no value. I decided to go with 3090. It has 24GB of RAM. Two 3060 may be a bit cheaper but it is slower than one 3090 and it may require some software tweaks.

As a 3090 owner who pretty much only uses it for ML, that was a nice pick haha

I'll likely pick a second 3090 next month or so 馃檪

FX6300 @ 4.2GHz | Gigabyte GA-78LMT-USB3 R2 | Hyper 212x | 3x 8GB + 1x 4GB @ 1600MHz | Gigabyte 2060 Super | Corsair CX650M | LG 43UK6520PSA
ASUS X550LN | i5 4210u | 12GB
Lenovo N23 Yoga

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now