Settings

Theme

Jensen Huang – Nvidia GTC 2025 Keynote

nvidia.com

75 points by moondistance 10 months ago · 74 comments

Reader

nh23423fefe 10 months ago

> "geforce is sold out all over the world"

not a flex

  • accrual 10 months ago

    Agree. It's a terrible time to want a new Nvidia card for gaming or AI projects. I had a 4070 Ti Super in my cart for a while at a cool $700 on Amazon - now it's out of stock and secondhand I can find it for double that now.

    If one doesn't require the latest DLSS or need first-class AI support, get an AMD card.

    • SV_BubbleTime 10 months ago

      AMD’s is coming up in support for things like torch and xformers. But it isn’t there yet.

      If you want for AI, your best bet is a used 3090 for $700-800. The ram is more important and that card is still faster than a 4070.

      • codedokode 10 months ago

        AMD has documentation for its assembly language as I remember. Or one needs more documentation to run AI?

        • SV_BubbleTime 10 months ago

          Isn’t a documentation issue.

          The entire ecosystem is built around CUDA currently. There was someone implementing Cuda on AMD hardware but as I understand it, AMD shut that down.

          There are two alternatives to CUDA for AMD right now, but I forget their names.

          AMD has STRIX coming, which is a single board embedded ram machine, and is being targeted at AI loads with lots of low latency ram. So we’ll see what happens there.

    • colonCapitalDee 10 months ago

      Thankfully the ONNX runtime supports AMD's ROCm. The performance is nowhere close to Nvidia's TensorRT, but doing inference on an AMD GPU is doable. I haven't ran the numbers, but I wouldn't be surprised if you could get more throughput per dollar on an AMD GPU for at least some scenarios. AFAIK there aren't any great options for training though, so if you want to do anything besides running ONNX models your options are limited.

  • timtom123 10 months ago

    easy to do when you don't ship any cards :P

  • spullara 10 months ago

    oddly just bought one (5090) right after you posted this so not quite true

    • KeplerBoy 10 months ago

      For how much? In Europe they are in stock. For 3500€. For the MSRP, I'd buy it in a heartbeat.

      • spullara 10 months ago

        It was a full computer, $5200 for an i9-14900KF, 64 GB, 4TB NVMe machine with a 5090.

    • AuryGlenz 10 months ago

      I’ve been trying since release with stock notifications and have barely had the chance to even try to add one to a cart.

  • colonelspace 10 months ago

    > "we've left money on the table because we can't manage our supply chain. Oh, and we've pissed off an entire market segment."

    • dralley 10 months ago

      It's probably false to say they're leaving money on the table. They'd be leaving a lot more money on the table if they allocated chips towards consumer gaming GPUs instead of maxing out the server AI/GPU compute segment. The entire gaming market constitutes like 15%-or-less of their revenue nowadays.

      And Nvidia has enough mindshare that they could piss on consumers for the next 3 release cycles and still have more than half the market. I don't like it but it's reality.

      • whywhywhywhy 10 months ago

        The top end should at least be stocked because a lot of us are using cards locally for AI that eventually runs in the cloud.

        So yeah they did leave $2600 from me on the table that is now becoming more likely to be spent on a bootleg 48GB 4090 than a 5090 and if I get that they won’t see money from me for many years till they beat 48GB in consumer form factor.

      • m4rtink 10 months ago

        "You PC people got us where we are now - so screw you, we don't need you anymore because the AI bubble will go forever!"

  • PeterStuer 10 months ago

    Yep, all 10 they shipped sold instantly at 2x MSRP. Prices on 3090's (not a typo) are going up ffs, because there is nothing out there on the NVIDIA front.

    • incognito124 10 months ago

      My 3090 Ti appreciated in price for about 20% in 2 years (1300->1600)

      nVidia GPUs have basically been an inflation hedge for the past few years lol

      • davely 10 months ago

        This made me curious, so I just took a look on eBay.

        My 4070 Ti that I bought (new, oops) this past December has appreciated 20 - 25%! At least according to what people have them listed for... no idea if anyone is actually buying them at those prices.

        Of course even if I managed to sell it, everything else has gone up in that time, so it's not like I'd get to make money on the deal. Pretty wild, nonetheless!

    • renewiltord 10 months ago

      This is sick. Lucky I kept mine https://news.ycombinator.com/item?id=42618996

  • fransje26 10 months ago

    Or quite a flex. If you want to create artificial scarcity to keep your prices high.

    Reminds me of the current egg "crisis".

  • ecshafer 10 months ago

    NVidia doesn't make the chips, they just design them, so it is a flex. People are literally buying their cards faster than they can be produced. TSMC is building new fabs to make more nvidia chips faster.

    • atq2119 10 months ago

      NVidia decide how they allocate their wafer starts at TSMC. Consumer chips have lower margin than datacenter parts, so they almost certainly allocated comparatively little volume to consumer chips.

      On the one hand, this is a great situation to be in for Nvidia in terms of overall revenue.

      On the other hand, this has allowed AMD to grab market share with the RX 9000 series launch, at least in the short term. So the narrow point that Geforce is sold out is decidedly not a flex.

      • YetAnotherNick 10 months ago

        But the datacenter GPUs are sold out too even at ~90% profit margin. I heard somewhere lead time is in years.

    • whywhywhywhy 10 months ago

      Literally the worst thing about Nvidia is the fact they don’t take the core product end to end serious enough.

      The idea it’s hard to buy a standard Nvidia GPU in the consumer lineup is absurd.

cdfuller 10 months ago

GTC stands for GPU Technology Conference, for anybody else who wasn't familiar with the acronym. Surprisingly they don't have it spelled out anywhere on the homepage.

  • queuebert 10 months ago

    While we're at it, what does Nvidia mean?

    • bcatanzaro 10 months ago

      It means “envy” in Latin and Spanish etc. Which is why the logo is an acid green evil eye. And why NVIDIA’s headquarters are two buildings named “ENdeavor” and “Voyager”.

wewewedxfgdf 10 months ago

Nvidia has lost the hearts and minds - I'm not just not interested in or excited by anything Nvidia - there's a baseline expectation of their products being overpriced and aimed at stiffing the customer.

An outcome of greed.

  • numba888 10 months ago

    what's the better alternative, DIY AMD? No, thanks. I'm interested in the result, not the process. In terms of robotic software there is just nobody close. Any attempt to port it to different hardware will be just a big pain and limited result at best. So NVidia looks to be the way to go till China catches up.

  • comeonbro 10 months ago

    "Nobody goes there nowadays, it's too crowded"

  • artemonster 10 months ago

    honestly, seeing a company that craps so hard on a core customer group still succeed THAT much just because of some random trends is just sad. and yet people are still willing and continue to vote with their wallets (as the comment below was proving). what a mess

    • andrepd 10 months ago

      They're really the luckiest bastards lol. First crypto, then the "AI" craze...

      • bobsondugnut 10 months ago

        success is the combination of luck and preparation. NVIDIA has been well prepared + anticipating deep learning to take off for a long while.

baobabKoodaa 10 months ago

The video glitches out at 2:14:30 where Jensen would be introducing DGX Spark and Station. Does anyone have a link to a working video for that segment?

  • adrian_b 10 months ago

    The specifications are at:

    https://www.nvidia.com/en-us/products/workstations/dgx-spark...

    DGX Spark has the same memory as AMD Strix Halo, a weaker CPU, but perhaps a stronger GPU, except that for now there is no data about the GPU, besides that it might be stronger for AI inference (only FP4 speed is given). For now it is not known whether for graphics it will be better than Strix Halo.

    While DGX Spark might be weaker than AMD Strix Halo for anything else except AI Inference, it will still be stronger than any mini-PC made with Intel Arrow Lake H or with AMD Strix Point.

    • canucker2016 10 months ago

      CPU specs from the nVidia page above:

      20 core Arm, 10 Cortex-X925 + 10 Cortex-A725 Arm

      Anandtech's 2024 article about the X925 & A725 Arm cpu cores:

      https://www.anandtech.com/show/21399/arm-unveils-2024-cpu-co...

      • adrian_b 10 months ago

        For operations with floating-point numbers, big integer numbers or arrays, a Cortex-X925 matches 3/4 Zen 5 full cores at the same clock frequency, while Cortex-A725 matches 1/4 Zen 5 full cores.

        So DGX Spark is equivalent with at most 10 * 3/4 + 10 * 1/4 = 10 Zen 5 cores versus 16 Zen 5 cores of Strix Halo.

        In reality DGX Spark will be even slower, because it will have a lower clock frequency (especially the Cortex-A725 cores) and a worse cache memory.

        For irregular code that does only operations with pointers and integers, the advantage of AMD Strix Halo will be significantly less, but even in that case the 10+10 cores of DGX Spark are unlikely to match more than 15 Zen 5 cores at the same clock frequency and less than that at the real clock frequencies.

        On the other hand, like I have said, DGX Spark should be faster than Intel Arrow Lake H, the best that Intel can offer in a mini-PC of this size.

    • KeplerBoy 10 months ago

      I'm somewhat surprised they are not explicitly mentioning the networking capabilities. There should be two QSFP(?) ports but they don't mention if their speeds and supported protocols (possibly not just Ethernet but also infiniband, but probably not).

      Could be 2x200 Gbit/s, could also be much less.

    • KeplerBoy 10 months ago

      It will be better than strix halo. Even the Orin AGX GPU is better than strix halo and that is Ampere.

      • adrian_b 10 months ago

        Perhaps you are confusing Strix Halo with Strix Point.

        The GPU of Strix Halo is many times faster than the GPU of the biggest Orin AGX.

        The Strix Halo has 25% more "CUDA cores" (2560 vs. 2048), which work at a clock frequency that is more than double (2.9 GHz vs. 1.3 GHz) and which can have for some operations a double throughput even at the same clock frequency. The memory throughput is also higher by 25%.

        The GPU of DGX Spark will have to be 4 or 5 times faster than Orin to match Strix Halo as a GPU. This is not at all certain because NVIDIA has stressed only AI/ML applications without saying anything about graphics.

  • baobabKoodaa 10 months ago

    NVIDIA uploaded the missing part in a separate upload now: https://www.youtube.com/watch?v=6p4U1kSiegg

lousken 10 months ago

I feel like I am seeing same stuff every year for like half a decade Self driving cars? Simulations? Robots? What's changed?

  • comeonbro 10 months ago

    You seem to have somehow missed one of the most significant advancements in human history suddenly exploding out of the pages of science fiction and fundamentally altering our existence forever. Minor oversight.

    • mepian 10 months ago

      Ah yes, a chat bot that is consistently failing to do such advanced tasks as counting letters in words is definitely life-changing, right next to the metaverse and NFTs.

      • davely 10 months ago

        Pointing out this one error is reductive and greatly simplifies the profound effect that LLMs (and research around them) are having across a wide range of industries, unlike the metaverse and NFTs.

      • john_minsk 10 months ago

        Metaverse is changing the world. Just as you could see in this presentation - Data centers are first being built as digital twins in Omniverse and later in real world.

        It is a profound change, but it will take time for significant number of digital twins being build. However digital twin is a CAPEX in a way. You build 3D model of your warehouse only once and then use it for years to manage operations, robots and everything else.

  • tim333 10 months ago

    It's all getting better though, especially the robots seem to be coming along.

    Compare the Disney robots at the end of 2024 and 2025

    2040: https://youtu.be/Y2F8yisiS6E?t=6968

    2025: https://youtu.be/_waPvOwL9Z8?t=9042

    • lousken 10 months ago

      I mean it's a cool demo, but it doesn't really show anything meaningful

      • john_minsk 10 months ago

        It means that Disney, company which is a CGI/movie/cartoon producing company(Pixar created OpenUSD format) can use the same tools and make physical robot walk on stage.

        If the venue, where the presentation took place also has readily available 3D model made with OpenUSD in SimReady level of details, Pixar could just download this model to their cartoon tools and start training virtual model of the robot in the virtual model of the venue before the event and troubleshoot any issues with the program.

        Isn't it mindblowing?

      • tim333 10 months ago

        Ah I had a typo. The vids are from 2024 and 2025. It seems to me the 2024 robots are quite clunky moving as you kind of expect from robots, whereas the 2025 one moves quite well with comparable dexterity to an animal. Which according to the talk is down to nvidia's new tech. I mean you see robots move like that in movies but I think it's the first time I've seen it in real life. Apparently they are going to put them in the Disney theme parks.

        It's significant in the scheme of things in that LLMs have got quite good at text chat but for AI do to real world things like build you a house or fix your car it's got to get good at physical robot stuff too.

nmstoker 10 months ago

Some interesting points but he seems a bit less polished and confident than usual.

  • canucker2016 10 months ago

    At 32:30, Jensen Huang mentions that he's not following a script and he's not using a teleprompter.

numba888 10 months ago

2 hours in, don't see anything about Digits, their 'personal supercomputer'. Did I miss something?

  • Rooster61 10 months ago

    The stream cut out for the main debut, but it showed up later once it came back. It's called DGX Spark

    Very unfortunate time for it to cut out

    • numba888 10 months ago

      It looks to be the whole line with different performance. But no dates and times

nialv7 10 months ago

DGX Spark only has 273 GB/s memory bandwidth?! Massive disappointment, this is just on par with Framework's desktop PC.

  • Gracana 10 months ago

    That's what everybody was predicting based on the memory technology and the pictures. I'm just glad to hear about DGX Station, though I doubt I'll be able to afford one.

    • numba888 10 months ago

      > glad to hear about DGX Station

      how much is it expected, my guess barely fits in 5 digits. Would be nice to have something in between Spark and Station. I.e. some desktop withing $20K.

      • numba888 10 months ago

        well, "RTX Pro Blackwell" desktop and workstation version looks to be the answer for midrange. Some should be available starting in April.

    • canucker2016 10 months ago

      theVerge article says the DGX Spark will cost US$3K, about the same ballpark range as the Framework desktop with the maxed out 128GB RAM option.

      https://www.theverge.com/news/631957/nvidia-dgx-spark-statio...

      10Gb vs 5Gb Ethernet

      1 HDMI 2.1, 4 x USB TypeC vs 1 HDMI 2.1, 2 Displayport, 2 USB TypeC

      • InTheArena 10 months ago

        The unified memory model w/ CUDA support should be the huge differentiator for this hardware.

        I think a better comparison might be the Mac Studio ultra.

    • nialv7 10 months ago

      I think most people (including me) were expecting 384-bit and ~400G/s. Can't believe they went with 256.

      • Gracana 10 months ago

        Based on the chip count in the promo picture it looked like it was going to be a 256 bit interface, but it'd be unfair to say that it was known at that point. I certainly wish it was better.

john_minsk 10 months ago

Amazing video of Omniverse. So many announcements!

shmerl 10 months ago

I'll leave this here for lulz - "Nvidia engineer explains RTX 5000 series": https://www.youtube.com/watch?v=pCyNp8aPcsY

numba888 10 months ago

cool robot at the end ;)

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection