Jump to content

Folding @ Home on AWS GPU instances

Has anybody actually tried running an AWS GPU instance to see how effective it is versus just buying a nice GPU yourself and paying for the electricity?  

 

I'm looking at one of their GPU instances -- p2.xlarge -- and their spot pricing is pretty consistently $.27 per hour to run a GPU that seems like it could run about 50% of the throughput of my 2080 Ti (granted, I'm not fully aware of what F@H primarily uses from the card -- I'm assuming CUDA cores).  Given the fairly consistent spot pricing, I could theoretically run the instance for about a week for the cost of the electricity it takes to run my home PC.  

 

I'm interested in trying it, but the big question is whether their GPUs can actually perform F@H work efficiently.  Has anybody done this?  What were your results?  

#Muricaparrotgang

 

Folding@Home Stats | Current PC Loadout:

Small                        Bigger				Biggerer				Biggest
Fractal Design Focus G       NZXT H1				Lian LI O11 Dynamic XL			Fractal Design Meshify C
FX-8320                      Ryzen 3 3200G			Ryzen 5 3600				Ryzen 7 3700X
120mm AIO                    120mm AIO				Custom 280mm loop			Noctua NH-D15
A motherboard                ASRock B450 mobo			MSI x570 mobo				MSI x570 mobo
16gb DDR3                    16gb DDR4 @ 3200			16gb DDR4 @ 3200			16gb DDR4 @ 3600
a melange of HDDs/SSDs       WD 1tb m.2				WD 500gb m.2				WD 1tb m.2/2tb HDD
PNY GTX 1070 x2              GTX 1070				GTX 1070 FE				MSI RTX 2080 TI
some 650w PSU                650W SFX-L 80+ Gold		MSI RTX 2080 Super			EVGA SuperNova 750w 80+ GOLD 
								Corsair RM850x 80+ GOLD

 

Link to comment
Share on other sites

Link to post
Share on other sites

They should do the work fine, its just a system with a gpu.

 

Why not try it?

 

If you want to help, your probably better off donating money than paying for the aws time. Then let them decide if thats the best use of their. Also there is a good amount of markup, so thats why there getting 1/2 of the performance for the price.

 

Also what are your power prices? You would have to have very high power prices for a 2080ti to coast $.27 a hour in power to run

Link to comment
Share on other sites

Link to post
Share on other sites

Just now, Electronics Wizardy said:

They should do the work fine, its just a system with a gpu.

 

Why not try it?

 

If you want to help, your probably better off donating money than paying for the aws time. Then let them decide if thats the best use of their. Also there is a good amount of markup, so thats why there getting 1/2 of the performance for the price.

 

Also what are your power prices? You would have to have very high power prices for a 2080ti to coast $.27 a hour in power to run

I don't have a voltmeter so I'm estimating the cost of running my own GPU -- I'm guessing the cost of running my system full bore to be about 500 watts on the high end.  That's about 360 kwh a month to keep it going 24/7.  At my local pricing ($.12  per kwh), that's about 40 bucks a month.  At that pricing, I could run an AWS instance at $.27 for roughly 7 days.  

 

That's a good point about just donating the money, though.  I just happened to look at AWS for some reason their free tier offer got me to spin up an instance, then I got curious and looked into their GPU instance pricing.  It's just low enough to make me think it's worth doing every now and again when I want more PRECIOUS POINTS, but plenty high enough for me to figure I should just build another system: over the course of a year or so, the new system should be way cheaper.  

#Muricaparrotgang

 

Folding@Home Stats | Current PC Loadout:

Small                        Bigger				Biggerer				Biggest
Fractal Design Focus G       NZXT H1				Lian LI O11 Dynamic XL			Fractal Design Meshify C
FX-8320                      Ryzen 3 3200G			Ryzen 5 3600				Ryzen 7 3700X
120mm AIO                    120mm AIO				Custom 280mm loop			Noctua NH-D15
A motherboard                ASRock B450 mobo			MSI x570 mobo				MSI x570 mobo
16gb DDR3                    16gb DDR4 @ 3200			16gb DDR4 @ 3200			16gb DDR4 @ 3600
a melange of HDDs/SSDs       WD 1tb m.2				WD 500gb m.2				WD 1tb m.2/2tb HDD
PNY GTX 1070 x2              GTX 1070				GTX 1070 FE				MSI RTX 2080 TI
some 650w PSU                650W SFX-L 80+ Gold		MSI RTX 2080 Super			EVGA SuperNova 750w 80+ GOLD 
								Corsair RM850x 80+ GOLD

 

Link to comment
Share on other sites

Link to post
Share on other sites

On 5/20/2020 at 8:44 AM, Rybo said:

I'm interested in trying it, but the big question is whether their GPUs can actually perform F@H work efficiently.  Has anybody done this?  What were your results?  

Oh and the results can be quite good. Here's the contest we had last month and I ran entirely on AWS

 

 

Link to comment
Share on other sites

Link to post
Share on other sites

20 hours ago, Macaw2000 said:

Works quite well. My cluster runs on g4dn.xlarge instances when the spot pricing is under $0.15. 

 

Darn.  

I did end up trying it on p2.xlarge and while it was working, it was taking like 15 minutes per fold (i.e., not good).  It did start messing up and dumping work units so I gave up before running up a bill (I installed it on Amazon's custom command line Linux AMI, so I probably just needed to jump further down the "configure it properly" rabbit hole), but I wonder if the tensor GPUs in the g4dn instances are better aligned to F@H workloads or if my instance was simply not running properly. 

 

I'm looking forward to a long 4 day weekend, so I might take another stab at this.  I'd love to get it working if only to sate my curiosity (and to claw my way to 100m points a bit quicker...and to be able to ramp up production very quickly when the next F@H event hits)

#Muricaparrotgang

 

Folding@Home Stats | Current PC Loadout:

Small                        Bigger				Biggerer				Biggest
Fractal Design Focus G       NZXT H1				Lian LI O11 Dynamic XL			Fractal Design Meshify C
FX-8320                      Ryzen 3 3200G			Ryzen 5 3600				Ryzen 7 3700X
120mm AIO                    120mm AIO				Custom 280mm loop			Noctua NH-D15
A motherboard                ASRock B450 mobo			MSI x570 mobo				MSI x570 mobo
16gb DDR3                    16gb DDR4 @ 3200			16gb DDR4 @ 3200			16gb DDR4 @ 3600
a melange of HDDs/SSDs       WD 1tb m.2				WD 500gb m.2				WD 1tb m.2/2tb HDD
PNY GTX 1070 x2              GTX 1070				GTX 1070 FE				MSI RTX 2080 TI
some 650w PSU                650W SFX-L 80+ Gold		MSI RTX 2080 Super			EVGA SuperNova 750w 80+ GOLD 
								Corsair RM850x 80+ GOLD

 

Link to comment
Share on other sites

Link to post
Share on other sites

I've used AWS for folding and boinc before... the only thing you need to consider is your bill goes up fast. I wouldn't recommend it honestly. You can also look into doing P100 instances for 0.25/hr on Azure. It pulls about 1.4m to 1.6m. If you can, I would just get a gpu and run it from home. It's a better investment in the long run.

 

Regarding the AMI, I would use the Nvidia Deep Learning AMI. Easier to setup because you don't have to deal with drivers.

https://aws.amazon.com/marketplace/pp/NVIDIA-NVIDIA-Deep-Learning-AMI/B076K31M1S

 

Also consult this guide on setting it up:

 

Link to comment
Share on other sites

Link to post
Share on other sites

9 hours ago, Macaw2000 said:

If Docker is interesting... it's a far cleaner solution than an AMI.  Nice example of the Docker config here: https://github.com/raykrueger/FoldingAtHomeContainer

 

Edit: Docker cleaner, AMI much easier barrier to entry and works well!

 

About that...Docker seems to only work in the AWS cloud project (not plain old manually created EC2 instances), which seems to only be trivially accessible to the outside world via a load balancer.  I'm sure I could eventually learn how to set up access tokens to get my local pc into the VPC and hook directly up to each instance in the cluster (I would like to have all of them linked up to my FAH control on my PC for monitoring), but bleh.  Learning curve. 

What I'll probably end up doing is harvesting the initialization commands and just executing them manually in manually created instances.  It looks like you're using Amazon's AMI anyways, so presumably that closes the one gap I had in my first attempt -- configuring the instance with the proper drivers.   

 

I did spend like 3 hours yesterday adding configuration to the FoldingOnEcs project so I could customize the spot pricing and cluster size via config file before I realized how profoundly...enterprisey the AWS cloud solution is.  Not sure I'm up for the learning curve just for the sake of this project ;)  Buuuut, I finally got to use Typescript without React, so I consider it time well spent anyways.  

#Muricaparrotgang

 

Folding@Home Stats | Current PC Loadout:

Small                        Bigger				Biggerer				Biggest
Fractal Design Focus G       NZXT H1				Lian LI O11 Dynamic XL			Fractal Design Meshify C
FX-8320                      Ryzen 3 3200G			Ryzen 5 3600				Ryzen 7 3700X
120mm AIO                    120mm AIO				Custom 280mm loop			Noctua NH-D15
A motherboard                ASRock B450 mobo			MSI x570 mobo				MSI x570 mobo
16gb DDR3                    16gb DDR4 @ 3200			16gb DDR4 @ 3200			16gb DDR4 @ 3600
a melange of HDDs/SSDs       WD 1tb m.2				WD 500gb m.2				WD 1tb m.2/2tb HDD
PNY GTX 1070 x2              GTX 1070				GTX 1070 FE				MSI RTX 2080 TI
some 650w PSU                650W SFX-L 80+ Gold		MSI RTX 2080 Super			EVGA SuperNova 750w 80+ GOLD 
								Corsair RM850x 80+ GOLD

 

Link to comment
Share on other sites

Link to post
Share on other sites

19 hours ago, ConceptOW said:

Regarding the AMI, I would use the Nvidia Deep Learning AMI. Easier to setup because you don't have to deal with drivers.

Good call.  Gave it another stab on the g4dn.xlarge instance with the Amazon's base deep learning linux AMI.  After a few minutes of ramping up, it seems to be matching my 2080 Ti in TPF: about 1 minutes 10 seconds.  No unpleasant folding errors yet at 7% progress.  

 

I'll let it run for a few hours and see if it will actually work this time :)   

#Muricaparrotgang

 

Folding@Home Stats | Current PC Loadout:

Small                        Bigger				Biggerer				Biggest
Fractal Design Focus G       NZXT H1				Lian LI O11 Dynamic XL			Fractal Design Meshify C
FX-8320                      Ryzen 3 3200G			Ryzen 5 3600				Ryzen 7 3700X
120mm AIO                    120mm AIO				Custom 280mm loop			Noctua NH-D15
A motherboard                ASRock B450 mobo			MSI x570 mobo				MSI x570 mobo
16gb DDR3                    16gb DDR4 @ 3200			16gb DDR4 @ 3200			16gb DDR4 @ 3600
a melange of HDDs/SSDs       WD 1tb m.2				WD 500gb m.2				WD 1tb m.2/2tb HDD
PNY GTX 1070 x2              GTX 1070				GTX 1070 FE				MSI RTX 2080 TI
some 650w PSU                650W SFX-L 80+ Gold		MSI RTX 2080 Super			EVGA SuperNova 750w 80+ GOLD 
								Corsair RM850x 80+ GOLD

 

Link to comment
Share on other sites

Link to post
Share on other sites

When I had my network at full power on Folding @home. I was really used lot power. Although I went from just starting to about 550 place on ltt group in about month half. Then it got warm out. Were I would had to to run the AC to stay comfortable in place. So I have turned most computers off. Expect the one machine and sometimes run the other media center computer with it. I would guess that it would be cheaper just to run your own system. I also running your own computers maxed out give me little sense pride. Since most of  the system that I ran were over clocked in one way or another.  I did let some of my ESXI server run with extra cpu cycles. Other then enable XMP on them they at stock settings.

 

Link to comment
Share on other sites

Link to post
Share on other sites

17 hours ago, Rybo said:

About that...Docker seems to only work in the AWS cloud project (not plain old manually created EC2 instances), which seems to only be trivially accessible to the outside world via a load balancer.  I'm sure I could eventually learn how to set up access tokens to get my local pc into the VPC and hook directly up to each instance in the cluster (I would like to have all of them linked up to my FAH control on my PC for monitoring), but bleh.  Learning curve. 

What I'll probably end up doing is harvesting the initialization commands and just executing them manually in manually created instances.  It looks like you're using Amazon's AMI anyways, so presumably that closes the one gap I had in my first attempt -- configuring the instance with the proper drivers.   

 

I did spend like 3 hours yesterday adding configuration to the FoldingOnEcs project so I could customize the spot pricing and cluster size via config file before I realized how profoundly...enterprisey the AWS cloud solution is.  Not sure I'm up for the learning curve just for the sake of this project ;)  Buuuut, I finally got to use Typescript without React, so I consider it time well spent anyways.  

Easier way to do it is create a snapshot of your instance and then create an ami. Then create a template for your instances with that ami. It will boot up with everything configured already.

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now

×