THIS is NVIDIA's New AI GPU for up to 10,000 GPU Installations ft. Supermicro

THIS is NVIDIA's New AI GPU for up to 10,000 GPU Installations ft. Supermicro

ServeTheHome

6 месяцев назад

123,348 Просмотров

Ссылки и html тэги не поддерживаются


Комментарии:

@godsmackssa
@godsmackssa - 28.01.2024 23:28

You become a billionaire, to some govern agent make you "fly" from some building randomly :)

Ответить
@godsmackssa
@godsmackssa - 28.01.2024 23:26

Nowadays it's dangerous be a billionaire. Just survive and be an average "asset".

Ответить
@josueveguilla9069
@josueveguilla9069 - 20.01.2024 01:59

AI = Artificial Idiot

Ответить
@MARKXHWANG
@MARKXHWANG - 14.01.2024 07:13

Very good in depth study

Ответить
@for2utube
@for2utube - 13.01.2024 05:39

I think I need the Torque Test Channel for these.

Ответить
@geekinasuit8333
@geekinasuit8333 - 11.01.2024 23:48

Yeah the marketing was more than over the top in this one, but this is Nvidia for you. The L40S is a low-end device made to get around regulations limiting capabilities sold to China. Never-the-less, $300K for a few low-end GPU's is unreal, I can't see the price gouging continue for much longer now that competition is rolling in.

Ответить
@bogartICQa
@bogartICQa - 09.01.2024 00:23

Mhm. Irrelevant for me for a number of reasons, but have fun.

Ответить
@blender_wiki
@blender_wiki - 08.01.2024 08:33

If you are considering upgrading your models from an L40 to an L40S, it's important to observe that the tensor cores exhibit similar speeds on both cards at INT4 and you don't have any concrete adavantage. Additionally, if you currently use or own a few RTX6000 GPUs, upgrading to the L40S may not yield significant performance improvements; the differences in performance are marginal, except for certain specific aspects. However, if you are making a new investment, opting for the L40S could be a more favorable choice. It is advisable to carefully plan the amortization of this investment over an 18-month period.

Ответить
@zyxwvutsrqponmlkh
@zyxwvutsrqponmlkh - 08.01.2024 07:20

Need more ram.
Also just use 4090's 10x cheaper for the same performance. But less ram :'c(

Ответить
@theworddoner
@theworddoner - 08.01.2024 02:40

The prohibitive prices we are seeing here is why I gave up on vram products for inferencing.

APU’s with much cheaper but slower ddr5 memory is the way to go. The Apple m1 ultra showed that this is possible.

We just need Intel, AMD, Qualcomm to make this product with a more open and repairable ecosystem.

I won’t say no to affordable vram products but at the end of the day, they will only supplement the APU.

Ответить
@koijoijoe
@koijoijoe - 07.01.2024 22:16

Soooooo is there a plan to turn them on, and compute something with them for a different video? Or did you just get them for footage haha. Cause sorry for the criticism, this was a good video topic but cmon -__-

Ответить
@chromerims
@chromerims - 07.01.2024 21:20

Unique coverage, fabulous 👍
Thank you STH

Ответить
@kwazar6725
@kwazar6725 - 07.01.2024 20:22

Looks like a nice workstation hedt to play games with.

Ответить
@JeffreyPeckham_abic
@JeffreyPeckham_abic - 07.01.2024 20:04

Love the videos, love the content. Following data center tech has always been fascinating for me, even though it's been 10 years since I've had to step foot inside one.

I never thought I'd have to say this, considering I'm known for being an incredibly fast talker, but this was the first video I've ever had to slow down to 75% in order to not feel burned out.

Ответить
@chubbysumo2230
@chubbysumo2230 - 07.01.2024 18:49

wow, so they found a way to ship more consumer GPUs into the professional market, and have fully abandoned the consumer market. expect the 4090 prices to go wild, and probably completely go away. how long before they start doing it with the AD103 die too? fuck Nvidia for this.

Ответить
@aldvelothi755
@aldvelothi755 - 07.01.2024 17:36

One big difference between the L40 and L40s that you conveniently "forgot" to mention, is that the L40s is 1 gram heavier than the L40. How much did big GPU pay you to leave this detail out of the video? This is exactly why I always do my own research.

Ответить
@orthodoxNPC
@orthodoxNPC - 07.01.2024 14:21

What's up with Supermicro's dual socket H13 boards lacking PCIe slots? Are they relying on "ribbons" to optimize? Asking for a broadwell friend

Ответить
@benjaminsmekens2344
@benjaminsmekens2344 - 07.01.2024 13:59

Just procured a bunch of L40S GPU's for our inference tasks, and they rip. We also have bunch of H100's for the initial training of our LLM's. You touched on it during your talk but power and cooling requirements are the biggest hurdle when it comes to the H/A100's. The L40S GPU's are a lot more reasonable, and the price and lead times are nice bonusses.

Ответить
@ProjectPhysX
@ProjectPhysX - 07.01.2024 11:50

It's a bit too much toxic marketing for me. Can't build multi-GPU servers with cheap gaming cards anymore because Nvidia killed 2-slot 3090s/4090s to prevent exactly this. Their coolers are so heavy that they rip out the PCIE slot under their own weight, and so hilariously large they don't fit in even normal PC cases.
L40S is identical to the RTX 4090, except for the extra 24GB memory chips which cost $100 in retail, yet they increase the price by an extra $9000.
And NVLink on the A100/H100 is kept locked to their proprietary CUDA language, so you can't use it with portable OpenCL codes.

Ответить
@bogganalseryd2324
@bogganalseryd2324 - 07.01.2024 09:45

But can it play crysis? 😂

Ответить
@shadowbrony33
@shadowbrony33 - 07.01.2024 09:26

As someone who works on these things, the only billionaire those GPUs will be making is Jensen Huang

Ответить
@zeekjones1
@zeekjones1 - 07.01.2024 05:52

"Look at the baby future president(s)!"
Chonky models will better manage countries than officials.

Ответить
@0xKruzr
@0xKruzr - 07.01.2024 05:33

about to build a 40-node cluster with one of these in each machine. they're really remarkably flexible cards.

Ответить
@OTechnology
@OTechnology - 07.01.2024 05:31

As someone who got disappointed by switching to a dual 3090 setup from a dual Titan RTX setup, I just want to add that an important metric that Nvidia buries in the spec sheet that's important for DL performance is the FP16 with FP32 Accumulate. Where Nvidia half rates it on the Geforce cards to stop them from beating the pro cards.

Ответить
@Veptis
@Veptis - 07.01.2024 05:28

For me, memory matters. H100 is 80GB, and RTX600 Ada, RTX5880 Ada, L40S and L40 are all 48GB. RTX4090 is just 24GB, but so is the RTX 3090ti, which is much cheaper.
There is water cooled RTX6000 Ada, but is there water cooled L40S?

Sadly, it's still not a good solution. Intel has nothing for workstation, AMD is missing software. At this point... I am hoping for Qualcomm to sell the Cloud AI100 Ultra with 128GB LPDDR5X for workstation

Ответить
@ajhieb
@ajhieb - 07.01.2024 04:20

What about Plex transcoding? /s

Ответить
@FlaxTheSeedOne
@FlaxTheSeedOne - 07.01.2024 03:42

By buying these GPUs I know someone who definetly is gonna be a Bilionare. But the more you buy the more you save.

Ответить
@AI-xi4jk
@AI-xi4jk - 07.01.2024 03:00

Hi Patrick, thanks for the specs table. Makes it much easier to visualise. Nvidia publishes specs in a different format for different GPUs so it’s tedious to compare especially when naming is different. How does A800 stack up to these?

Ответить
@geofrancis2001
@geofrancis2001 - 07.01.2024 02:57

This is where your 4090 shortage stems from....

Ответить
@HeNryous
@HeNryous - 07.01.2024 02:46

Supermicro estore link doesn’t work…

Ответить
@shephusted2714
@shephusted2714 - 07.01.2024 02:44

this is proof we are still in the stone age as far as ai goes - real ai doesn't hit sme/smb mkt for 5 years

Ответить
@Buy_YT_Views.426
@Buy_YT_Views.426 - 07.01.2024 02:03

I love how relatable you are in your vlog 👍

Ответить
@blackryan5291
@blackryan5291 - 07.01.2024 02:03

I know its off topic but yo P...thats a nice watch

Ответить
@-_-_-FIF-_-_-
@-_-_-FIF-_-_- - 07.01.2024 01:53

nice! thanks Patrick

Ответить
@DeclanBirmingham
@DeclanBirmingham - 07.01.2024 01:28

But can it run Crysis?

Ответить
@hotrodhunk7389
@hotrodhunk7389 - 07.01.2024 01:20

I can't wait to see videos in the years coming of people using these once they're like $500 on the used market. 😂😂😂 LTT will be all over that.

Ответить
@pcislocked
@pcislocked - 07.01.2024 01:16

Is this shot at the new studio or old? Can't quite tell idk

Ответить
@notreal5311
@notreal5311 - 07.01.2024 01:12

The sharesale Supermicro store link is broken FYI

Ответить
@TripleMachine
@TripleMachine - 07.01.2024 01:05

Can't wait to buy a H100 for $100 in 8-10 years

Ответить
@gl.72637
@gl.72637 - 07.01.2024 01:04

What is the closest GPU(s) from AMD against these H100 and L40S?

Ответить
@LerrySanders
@LerrySanders - 07.01.2024 00:57

Thanks for the info Patrick. Great learning about things Ill be able to afford in 20 years when it hits ebay.

Ответить
@Mpdarkguy
@Mpdarkguy - 07.01.2024 00:53

Isn't vGPU/SR-IOV all software sauce? I really dislike nvidia price gating features that should be everywhere

Ответить
@Darkk6969
@Darkk6969 - 07.01.2024 00:53

Crazy amount of GPU computing power in one box.

Ответить
@trapexit
@trapexit - 07.01.2024 00:51

Since I work in research the L40S just isn't a good choice. We need the VRAM. But for "smaller" or quantized LLM inference they can be fine.

Unfortunately, my data centers are not setup for high density servers so currently we are using 4x setups.

Ответить
@lurick
@lurick - 07.01.2024 00:37

The more you buy the more you save!!

Ответить
@SatsJava
@SatsJava - 07.01.2024 00:35

Wow

Ответить