Recommendations on new 2 x RTX 3090 setup

Uh, to be honest, I’d avoid both. I urge you to adopt ECC memory for serious applications.
Go for a Xeon. If you like to stay with socket 2066, you can both adopt Xeon W-22XX (counterparts to 109XX), or W-21XX (Skylake). The latter can be found very cheapo in China, they ship worldwide.

Particularly cheap are the W-21x0B, originally made for iMac Pros. The W-2150B has 10 cores, the W-2170B does have 14 cores. Both do have 120w tdp, and produce less heat than your average 2066 cpu. The latter can be bought for ~400$.

Advice for a mainboard (even if you decide to go for 109XX): https://www.gigabyte.com/Server-Motherboard/MW51-HP0-rev-10#ov

Let us know what you buy :wink:

1 Like

Thanks again!
Will look into my options and share some RGB pics soon! :smiley:

Unfortunately the great options suggested by you would take >1 month to reach my country (and will include some insane customs, Ex: on the Titan, I paid ~1000$ in customs!!) so I’d be forced to go with the 10-Gen processors :frowning:

Understood. Go for the 14 cores then, if you can afford it. The 10 cores is also solid for configurations with <= 2 GPUs. I do have its Xeon counterpart on my windows daily driver, and I’m perfectly satisfied with it!

1 Like

Many thanks again Andrea (and to everyone) for all the help in this discussion! :slight_smile:

Time to spend my life savings on hardware now :grin:

1 Like

Thank you for the interesting conversation!

1 Like

Why ECC is so important?
What si th epoint of ECC on the CPU and not on the GPU?

1 Like

Can some of you expert recommends all the parts here https://pcpartpicker.com/list/ or similar website…? It will be nice for noobs like me who never build to have all the components so we can just order them and build ? =)

2 Likes

@DrHB could you please share your budget/use-cases?

Are you also aiming for a 2x3090 build?

1 Like

Yes 2 x 3090 =) My budget will be 5-6k and its mostly for deep learning / Kaggle … But maybe it will be good to have several brackets ? I am pretty sure there are a lot of people who will benefit from this =)

2 Likes

Oh, I just recalled you’re in the US! :sweat_smile:

You could also consider:

  • Cheap 2nd hand 2080Tis: get 4 of them and liquid cool
  • Get a cheap CPU via Ebay, please see the messages by Andrea above

These aren’t possible in my country

2 Likes

Had anyone taken a look at double 3090 but undervolted to Max-Q? Wondering what sort of cooling that ends up taking.

I just don’t find the extra performance for power/heat worth it. Memory for increased batch size is all I really care about.

1 Like

ECC Ram is not important for deep learning activities per se. It is important if you want to preserve an efficient and stable system in the long run.
Every time that a bit flip occurs in memory (and that’s more common than one thinks, see the relevant literature), and a file is open in memory, when the OS writes it back, it can write a corrupted file upon the filesystem. If that’s a system file, you will end up with problems.
I was a system administrator once, and I managed systems both with and without ECC. The latter always drifted towards instability, and required system reinstalls. The former, never.
Finally, ECC on the GPUs is only required if you do high precision numerical analysis with them (FP64 calculations). Indeed, the cards aimed at that (tesla, Gx100…) do have ECC.

1 Like

That’s a very good idea. People have run extensive benchmarking about it, and it’s not even required to do awkward tricks like manual undervolting… Using nvidia-smi will suffice. It turns out that if you cap the maximum allowed power draw at 60%, you still enjoy some 90% of the card’s performance.
I do that regularly with my four GPUs (no space between them).

Often, we don’t appreciate the fact that manufacturers run their GPUs at their very limit in terms of power draw/heat produced/efficiency ratio, and that’s because even one single fps more in a game whatsoever is what really matters to them. For any serious use case that is inefficient and unadvisable.

1 Like

Let’s take out 3k for the cards. This leaves us with some 2.5k for the rest.
If you don’t require sleep/hibernation, I’d go for a ROMED8-2T motherboard and the best Epyc you can buy with the rest of the money. Otherwise, go for a socket 3647 solution.

2 Likes

Thank you for the recommendations in this thread so far.

@balnazzar, in addition to your recommended ROMED8-2T motherboard and an Epyc,
is there anything to pay special attention to (with regards to 2x rtx 3090)?

  1. Will a InWin R400N 4U Rackmount Server have sufficient space for the rtxs? any other recommendations?
  2. Any additional airflow needed?
  3. Will a Corsair MP600, Force Series fit under the rtxs on the board?

Also while looking at the Epyc 7282 I can only see that it as 64 lanes, not 128. I guess there a 7302 would be required?

  1. Hard to say. It seems ok, but mind that the 3090 has unusual height, and the lock-bar of the case could interfere. I guess you should send an e-mail to in-win.

  2. In my experience, space between cards counts more than case airflow, especially for non-blower card designs like the 3090 FE.

  3. It will fit, but it could develop overheating issues, see post #9 in this thread. It would certainly overheat with designs like the previous generation Founder’s. The new Ampere’s heatsink avoids blowing how air orthogonally on the mainboard, but still, it will be better to install the ssd on a pcie add-on cards, since the ROMED8 has plenty of slots…

  4. It has 128 lanes: EPYC 7282 - AMD - WikiChip. But for two 3090 it doesn’t really matter. They would work fine and without bottlenecks even with eight gen4 lanes each. Read it as: if you don’t care about ECC and registered ram (but you should), you can use a 3950X or any other processor with sixteen gen4 lanes.

2 Likes

Prior to go straight for two 3090, I would advise to wait a bit for knowing the street price of the new RTX A6000. Less headaches with parallelization, a lot less weight and torque exerted, standard card dimensions, no special requirements for the case, reasonable power draw, same amount of vram w.r.t. 2x3090.
Of course it would be interesting only if priced not much more than 3000$…

1 Like

To add to the discussion, I’ve used a lambda labs PC at work, which has a 12 core i9-9920X processor and 4x 2080ti. For all my deep learning workloads, the 12 core processor is surprisingly sufficient. The avg load is about 6.0 on htop during training and I don’t think I have noticable I/O issues. An easy way to overcome I/O problems is to load your datasets into ramfs, so having 128GB RAM really comes in handy.

3 Likes

It’s a 44 lanes processor, gen3. It would be interesting to know which motherboard this machine uses: is there a PLX/PEX switch? If not, it confirms that even for a 4 gpu setup, 8 gen3 lanes are sufficient.

I noticed the same with an e5-2660v3 (10 cores, 40 lanes) and four gpus. Still, a beefier processor would come in handy for a plethora of tasks not directly related to gpu training (augmentation, preprocessing of big tabular data, etc…).

Agreed. The more, the better. And here is where rdimms win: much higher density per module, better price per gb.

1 Like

Well its all plugged in and working! I am still trying to get the latest nvidia drivers (455) installed in ubuntu but everything on the windows side is working great. Maybe I can answer some questions if someone can lead me in the right direction to getting 455 installed in ubuntu. Right now cuda is not seeing the 3090s.

My bigger plan is to do a 6x 3090 FE setup using a dual case and water cooling. Basically I want to max out a consumer machine and not go server grade with epyc processors and a6000s or a100s

edit: Here is the current hardware:

CPU : AMD 3990x 64 core running at stock speed.
Motherboard : ASUS ROG Zenith II Extreme Alpha TRX40
RAM : G.SKILL Trident Z Royal Series 256GB (8 x 32GB) DDR4 3200
GPUs : Dual 3090 FEs with +100 on core clock and +200 memory clock. 114% on power
Boot drive : Samsung 2TB 970 EVO Plus NVMe
PSU : CORSAIR AXi Series, AX1600i, 1600 Watt
Case : Fractal Design Define 7 XL

Eventually I want to get a much bigger case, 2x 2000w power supplies, and 6x 3090 FEs. I want to mount the cards up and down vertically with the display ports at the bottom in a custom 3d printed enclosure. I would use a double case with the water cooling and motherboard on one side and the video cards on the other… Ekwb rads and blocks. I am thinking 4x 420 rads with two pumps and a large res. Each psu would be on a dedicated circuit with 4x 3090s on one psu and the rest of the system on the other psu.

4 Likes