Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (2024)

samspenc 8 months ago | parent | context | favorite | on: Show HN: I made a GPU VRAM calculator for transfor...


Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those cost $1000-2000 each. You can get higher VRAM but need enterprise GPUs which are in the five figures, easily starting at $30K a pop.

Per this calculator, for training, only gpt2-large and gpt2-medium would work with those two top-of-the-line GPUs.

For inference it's certainly a bit better, only the Llama-2-70b-hf and Llama-2-13b-hf don't fit in that much VRAM, all the other models do.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (1)

alexhutcheson 8 months ago | next [–]


Nvidia’s workstation cards are available with more RAM than the consumer cards, at a lower price than the datacenter cards. RTX 6000 Ada has 48 GB VRAM and retails for $6800, and RTX 5000 Ada has 32 GB VRAM and retails for $4000[1].

Very large models have to be distributed across multiple GPUs though, even if you’re using datacenter chips like H100s.

[1] https://store.nvidia.com/en-us/nvidia-rtx/store/

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (2)

slabity 8 months ago | parent | next [–]


Other than power consumption, is there any reason to prefer a single workstation card over multiple consumer cards then?

A single $6800 RTX 6000 Ada with 48GB of VRAM vs 6x 7900XTX with a combined total of 144GB of VRAM honestly makes this seem like a no brainer to me.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (3)

alexhutcheson 8 months ago | root | parent | next [–]


You can only fit 1-2 graphics cards in a “normal” ATX case (each card takes 2-3 “slots”). If you want 4 cards on one machine, you need a bigger/more expensive motherboard, case, PSU, etc. I haven’t personally seen anyone put 6 cards in a workstation.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (4)

oceanplexian 8 months ago | root | parent | next [–]


In a water cooled config the cards only take 1 slot. I’ve got 2 3090s and am buying another two shortly. Preemtively upgraded the power to 220v, found a 2kw PSU, and installed a dedicated mini split. I’m also undervolting the cards to keep power and heat down, because even 2000w is not enough to run 4 and a server grade CPU without tripping. When you start accumulating GPUs you also run into all kinds of thermal and power problems for the room, too.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (5)

kkielhofner 8 months ago | root | parent | next [–]


This is impressive.

I was fortunate enough to scoop up a bunch of Gigabyte RTX 3090 Turbos. Cheap used eight slot SuperMicro (or whatever), a cabling kit, four 3090s, boot.

Those were the days!

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (6)

alexhutcheson 8 months ago | root | parent | prev | next [–]


Sincere question: Is installing and running a mini split actually cheaper than racking them in a colo, or paying for time on one of the GPU cloud providers?

Regardless, I can understand the hobby value of running that kind of rig at home.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (7)

oceanplexian 8 months ago | root | parent | next [–]


I personally haven’t done the calculation. I have rented colo space before and they are usually quite stingy on power. The other issue is, there’s a certain element to having GPUs around 24/7/365 to play with that I feel is fundamentally different than running it on a Cloud Provider. You’re not stressing out about every hour it’s running. I think in the long run (2yr+) it will be cheaper, and then you can swap in the latest and greatest GPU without any additional infrastructure cost.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (8)

ttt3ts 8 months ago | root | parent | prev | next [–]


You have to pass the context between GPUs for large models that don't fit in VRAM. Often ends up slower. Also, tooling around AMD GPUs is still poor in comparison.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (9)

mciancia 8 months ago | prev | next [–]


Used 3090 are going for ~600usd these days (at least in Europe) thanks to crypto mining crash - building a workstation with 2 of these is fairly easy for 48GB of vram, with 4 a bit more tricky but still doable and affordable IMO

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (10)

ngoro7bd 8 months ago | parent | next [–]


Recently bought 24GB 3090 for $700 in US. Used but never tweaked, runs stable for 6 months despite heavy workloads.

nVidias play seems obvious. Game graphics don’t move that fast these days. Used market flush with 3090s and down is fine to them while they focus on extracting top dollar from fast moving AI researchers/VCs

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (11)

namibj 8 months ago | prev | next [–]


You can easily use pipeline-parallelism though.Especially if you have 8-16 lanes of PCIe4 with direct P2P access between the cards.

IIRC you want micro-batching though, to overlap pipeline phases.

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (12)

nox100 8 months ago | prev [–]


I haven't a clue how they compare but a Studio Mac with an M2 Ultra can get 192GB of unified ram for $5700 (PS: not a mac fan, a curious)

Consumer grade GPUs like NVidia's 3090 and 4090 max out at 24 GB VRAM, and those... (2024)
Top Articles
Explained: PoolTogether, the crypto lottery where you never lose - CNBC TV18
CFA and FRM: Best Way to Get Both Quickly | Zell Education
Costco Gas Barstow
Gomovies Spiderman
Eternal Sunshine Of The Spotless Mind Parents Guide
Florence Alabama Police
Craigslist Portland Motorhomes
Onlinewagestatements Lifepoint
Pennys Department Store Near Me
Pitco Foods San Leandro
International Medical Insurance for Employers from Aetna - Global Health Plans for Expatriates & Business Travel Employees | Aetna
Terramia Brick Oven Pizza & Trattoria Menu
Unit 7 Lesson 2 Joshua's Law
7Starhd Movies
Can Kool Aid Powder Go Bad
Herbalism Guide Tbc
Black Adam Showtimes Near Amc Brazos Mall 14
Khn Remote Access
War Thunder M60
Rescare Training Online
Uw Madison Kb
Lake Ridge Ixl
Catholic Church Near Seatac Airport
Lanipopvip
Student Choice Odysseyware
Loopnet Properties For Sale
Home Depot Roto Rooter Rental
Metro 72 Hour Extension 2022
Enola Bedard Net Worth 2022
Missing 2023 Showtimes Near Cinemark Chesapeake Square
Last Cloudia Radiance Of The World
Pewdiepieisprettydarncool
Violent Night Showtimes Near Amc Fashion Valley 18
Craigslist Edmond
Lbl A-Z
Teacup Yorkie For Sale Up To $400 In South Carolina
Restored Republic December 1 2022
A Closer Look at Woman With a Parasol by Claude Monet
Brokensilenze Rupaul
Tetroid Addicting Games
Doomz.io Unblocked Games 76
Trap Candy Strain Leafly
Google Flights Calgary
Straightup Internet Hotspot Pass
OPINION: My grandfather was Delaware's last Republican governor. I'm supporting Harris
Gran Turismo Showtimes Near Century Rio 24 And Xd
Amazon Gru Costume
Watch Wrestling.up
Summer in South Korea | University of Nevada, Reno
Quiktrip Gas Price Today
Whyp It Downloader
Nyu Salary Bands
Latest Posts
Article information

Author: Allyn Kozey

Last Updated:

Views: 5663

Rating: 4.2 / 5 (63 voted)

Reviews: 86% of readers found this page helpful

Author information

Name: Allyn Kozey

Birthday: 1993-12-21

Address: Suite 454 40343 Larson Union, Port Melia, TX 16164

Phone: +2456904400762

Job: Investor Administrator

Hobby: Sketching, Puzzles, Pet, Mountaineering, Skydiving, Dowsing, Sports

Introduction: My name is Allyn Kozey, I am a outstanding, colorful, adventurous, encouraging, zealous, tender, helpful person who loves writing and wants to share my knowledge and understanding with you.