Settings

Theme

Jensen Huang – Nvidia GTC 2025 Keynote

nvidia.com

75 points by moondistance a year ago · 74 comments

Reader

nh23423fefe a year ago

> "geforce is sold out all over the world"

not a flex

  • accrual a year ago

    Agree. It's a terrible time to want a new Nvidia card for gaming or AI projects. I had a 4070 Ti Super in my cart for a while at a cool $700 on Amazon - now it's out of stock and secondhand I can find it for double that now.

    If one doesn't require the latest DLSS or need first-class AI support, get an AMD card.

    • SV_BubbleTime a year ago

      AMD’s is coming up in support for things like torch and xformers. But it isn’t there yet.

      If you want for AI, your best bet is a used 3090 for $700-800. The ram is more important and that card is still faster than a 4070.

      • codedokode a year ago

        AMD has documentation for its assembly language as I remember. Or one needs more documentation to run AI?

        • SV_BubbleTime a year ago

          Isn’t a documentation issue.

          The entire ecosystem is built around CUDA currently. There was someone implementing Cuda on AMD hardware but as I understand it, AMD shut that down.

          There are two alternatives to CUDA for AMD right now, but I forget their names.

          AMD has STRIX coming, which is a single board embedded ram machine, and is being targeted at AI loads with lots of low latency ram. So we’ll see what happens there.

    • colonCapitalDee a year ago

      Thankfully the ONNX runtime supports AMD's ROCm. The performance is nowhere close to Nvidia's TensorRT, but doing inference on an AMD GPU is doable. I haven't ran the numbers, but I wouldn't be surprised if you could get more throughput per dollar on an AMD GPU for at least some scenarios. AFAIK there aren't any great options for training though, so if you want to do anything besides running ONNX models your options are limited.

  • timtom123 a year ago

    easy to do when you don't ship any cards :P

  • spullara a year ago

    oddly just bought one (5090) right after you posted this so not quite true

    • KeplerBoy a year ago

      For how much? In Europe they are in stock. For 3500€. For the MSRP, I'd buy it in a heartbeat.

      • spullara a year ago

        It was a full computer, $5200 for an i9-14900KF, 64 GB, 4TB NVMe machine with a 5090.

    • AuryGlenz a year ago

      I’ve been trying since release with stock notifications and have barely had the chance to even try to add one to a cart.

  • colonelspace a year ago

    > "we've left money on the table because we can't manage our supply chain. Oh, and we've pissed off an entire market segment."

    • dralley a year ago

      It's probably false to say they're leaving money on the table. They'd be leaving a lot more money on the table if they allocated chips towards consumer gaming GPUs instead of maxing out the server AI/GPU compute segment. The entire gaming market constitutes like 15%-or-less of their revenue nowadays.

      And Nvidia has enough mindshare that they could piss on consumers for the next 3 release cycles and still have more than half the market. I don't like it but it's reality.

      • whywhywhywhy a year ago

        The top end should at least be stocked because a lot of us are using cards locally for AI that eventually runs in the cloud.

        So yeah they did leave $2600 from me on the table that is now becoming more likely to be spent on a bootleg 48GB 4090 than a 5090 and if I get that they won’t see money from me for many years till they beat 48GB in consumer form factor.

      • m4rtink a year ago

        "You PC people got us where we are now - so screw you, we don't need you anymore because the AI bubble will go forever!"

  • PeterStuer a year ago

    Yep, all 10 they shipped sold instantly at 2x MSRP. Prices on 3090's (not a typo) are going up ffs, because there is nothing out there on the NVIDIA front.

    • incognito124 a year ago

      My 3090 Ti appreciated in price for about 20% in 2 years (1300->1600)

      nVidia GPUs have basically been an inflation hedge for the past few years lol

      • davely a year ago

        This made me curious, so I just took a look on eBay.

        My 4070 Ti that I bought (new, oops) this past December has appreciated 20 - 25%! At least according to what people have them listed for... no idea if anyone is actually buying them at those prices.

        Of course even if I managed to sell it, everything else has gone up in that time, so it's not like I'd get to make money on the deal. Pretty wild, nonetheless!

    • renewiltord a year ago

      This is sick. Lucky I kept mine https://news.ycombinator.com/item?id=42618996

  • fransje26 a year ago

    Or quite a flex. If you want to create artificial scarcity to keep your prices high.

    Reminds me of the current egg "crisis".

  • ecshafer a year ago

    NVidia doesn't make the chips, they just design them, so it is a flex. People are literally buying their cards faster than they can be produced. TSMC is building new fabs to make more nvidia chips faster.

    • atq2119 a year ago

      NVidia decide how they allocate their wafer starts at TSMC. Consumer chips have lower margin than datacenter parts, so they almost certainly allocated comparatively little volume to consumer chips.

      On the one hand, this is a great situation to be in for Nvidia in terms of overall revenue.

      On the other hand, this has allowed AMD to grab market share with the RX 9000 series launch, at least in the short term. So the narrow point that Geforce is sold out is decidedly not a flex.

      • YetAnotherNick a year ago

        But the datacenter GPUs are sold out too even at ~90% profit margin. I heard somewhere lead time is in years.

    • whywhywhywhy a year ago

      Literally the worst thing about Nvidia is the fact they don’t take the core product end to end serious enough.

      The idea it’s hard to buy a standard Nvidia GPU in the consumer lineup is absurd.

cdfuller a year ago

GTC stands for GPU Technology Conference, for anybody else who wasn't familiar with the acronym. Surprisingly they don't have it spelled out anywhere on the homepage.

  • queuebert a year ago

    While we're at it, what does Nvidia mean?

    • bcatanzaro a year ago

      It means “envy” in Latin and Spanish etc. Which is why the logo is an acid green evil eye. And why NVIDIA’s headquarters are two buildings named “ENdeavor” and “Voyager”.

wewewedxfgdf a year ago

Nvidia has lost the hearts and minds - I'm not just not interested in or excited by anything Nvidia - there's a baseline expectation of their products being overpriced and aimed at stiffing the customer.

An outcome of greed.

  • numba888 a year ago

    what's the better alternative, DIY AMD? No, thanks. I'm interested in the result, not the process. In terms of robotic software there is just nobody close. Any attempt to port it to different hardware will be just a big pain and limited result at best. So NVidia looks to be the way to go till China catches up.

  • comeonbro a year ago

    "Nobody goes there nowadays, it's too crowded"

  • artemonster a year ago

    honestly, seeing a company that craps so hard on a core customer group still succeed THAT much just because of some random trends is just sad. and yet people are still willing and continue to vote with their wallets (as the comment below was proving). what a mess

    • andrepd a year ago

      They're really the luckiest bastards lol. First crypto, then the "AI" craze...

      • bobsondugnut a year ago

        success is the combination of luck and preparation. NVIDIA has been well prepared + anticipating deep learning to take off for a long while.

baobabKoodaa a year ago

The video glitches out at 2:14:30 where Jensen would be introducing DGX Spark and Station. Does anyone have a link to a working video for that segment?

  • adrian_b a year ago

    The specifications are at:

    https://www.nvidia.com/en-us/products/workstations/dgx-spark...

    DGX Spark has the same memory as AMD Strix Halo, a weaker CPU, but perhaps a stronger GPU, except that for now there is no data about the GPU, besides that it might be stronger for AI inference (only FP4 speed is given). For now it is not known whether for graphics it will be better than Strix Halo.

    While DGX Spark might be weaker than AMD Strix Halo for anything else except AI Inference, it will still be stronger than any mini-PC made with Intel Arrow Lake H or with AMD Strix Point.

    • canucker2016 a year ago

      CPU specs from the nVidia page above:

      20 core Arm, 10 Cortex-X925 + 10 Cortex-A725 Arm

      Anandtech's 2024 article about the X925 & A725 Arm cpu cores:

      https://www.anandtech.com/show/21399/arm-unveils-2024-cpu-co...

      • adrian_b a year ago

        For operations with floating-point numbers, big integer numbers or arrays, a Cortex-X925 matches 3/4 Zen 5 full cores at the same clock frequency, while Cortex-A725 matches 1/4 Zen 5 full cores.

        So DGX Spark is equivalent with at most 10 * 3/4 + 10 * 1/4 = 10 Zen 5 cores versus 16 Zen 5 cores of Strix Halo.

        In reality DGX Spark will be even slower, because it will have a lower clock frequency (especially the Cortex-A725 cores) and a worse cache memory.

        For irregular code that does only operations with pointers and integers, the advantage of AMD Strix Halo will be significantly less, but even in that case the 10+10 cores of DGX Spark are unlikely to match more than 15 Zen 5 cores at the same clock frequency and less than that at the real clock frequencies.

        On the other hand, like I have said, DGX Spark should be faster than Intel Arrow Lake H, the best that Intel can offer in a mini-PC of this size.

    • KeplerBoy a year ago

      I'm somewhat surprised they are not explicitly mentioning the networking capabilities. There should be two QSFP(?) ports but they don't mention if their speeds and supported protocols (possibly not just Ethernet but also infiniband, but probably not).

      Could be 2x200 Gbit/s, could also be much less.

    • KeplerBoy a year ago

      It will be better than strix halo. Even the Orin AGX GPU is better than strix halo and that is Ampere.

      • adrian_b a year ago

        Perhaps you are confusing Strix Halo with Strix Point.

        The GPU of Strix Halo is many times faster than the GPU of the biggest Orin AGX.

        The Strix Halo has 25% more "CUDA cores" (2560 vs. 2048), which work at a clock frequency that is more than double (2.9 GHz vs. 1.3 GHz) and which can have for some operations a double throughput even at the same clock frequency. The memory throughput is also higher by 25%.

        The GPU of DGX Spark will have to be 4 or 5 times faster than Orin to match Strix Halo as a GPU. This is not at all certain because NVIDIA has stressed only AI/ML applications without saying anything about graphics.

  • baobabKoodaa a year ago

    NVIDIA uploaded the missing part in a separate upload now: https://www.youtube.com/watch?v=6p4U1kSiegg

lousken a year ago

I feel like I am seeing same stuff every year for like half a decade Self driving cars? Simulations? Robots? What's changed?

  • comeonbro a year ago

    You seem to have somehow missed one of the most significant advancements in human history suddenly exploding out of the pages of science fiction and fundamentally altering our existence forever. Minor oversight.

    • mepian a year ago

      Ah yes, a chat bot that is consistently failing to do such advanced tasks as counting letters in words is definitely life-changing, right next to the metaverse and NFTs.

      • davely a year ago

        Pointing out this one error is reductive and greatly simplifies the profound effect that LLMs (and research around them) are having across a wide range of industries, unlike the metaverse and NFTs.

      • john_minsk a year ago

        Metaverse is changing the world. Just as you could see in this presentation - Data centers are first being built as digital twins in Omniverse and later in real world.

        It is a profound change, but it will take time for significant number of digital twins being build. However digital twin is a CAPEX in a way. You build 3D model of your warehouse only once and then use it for years to manage operations, robots and everything else.

  • tim333 a year ago

    It's all getting better though, especially the robots seem to be coming along.

    Compare the Disney robots at the end of 2024 and 2025

    2040: https://youtu.be/Y2F8yisiS6E?t=6968

    2025: https://youtu.be/_waPvOwL9Z8?t=9042

    • lousken a year ago

      I mean it's a cool demo, but it doesn't really show anything meaningful

      • john_minsk a year ago

        It means that Disney, company which is a CGI/movie/cartoon producing company(Pixar created OpenUSD format) can use the same tools and make physical robot walk on stage.

        If the venue, where the presentation took place also has readily available 3D model made with OpenUSD in SimReady level of details, Pixar could just download this model to their cartoon tools and start training virtual model of the robot in the virtual model of the venue before the event and troubleshoot any issues with the program.

        Isn't it mindblowing?

      • tim333 a year ago

        Ah I had a typo. The vids are from 2024 and 2025. It seems to me the 2024 robots are quite clunky moving as you kind of expect from robots, whereas the 2025 one moves quite well with comparable dexterity to an animal. Which according to the talk is down to nvidia's new tech. I mean you see robots move like that in movies but I think it's the first time I've seen it in real life. Apparently they are going to put them in the Disney theme parks.

        It's significant in the scheme of things in that LLMs have got quite good at text chat but for AI do to real world things like build you a house or fix your car it's got to get good at physical robot stuff too.

nmstoker a year ago

Some interesting points but he seems a bit less polished and confident than usual.

  • canucker2016 a year ago

    At 32:30, Jensen Huang mentions that he's not following a script and he's not using a teleprompter.

numba888 a year ago

2 hours in, don't see anything about Digits, their 'personal supercomputer'. Did I miss something?

  • Rooster61 a year ago

    The stream cut out for the main debut, but it showed up later once it came back. It's called DGX Spark

    Very unfortunate time for it to cut out

    • numba888 a year ago

      It looks to be the whole line with different performance. But no dates and times

nialv7 a year ago

DGX Spark only has 273 GB/s memory bandwidth?! Massive disappointment, this is just on par with Framework's desktop PC.

  • Gracana a year ago

    That's what everybody was predicting based on the memory technology and the pictures. I'm just glad to hear about DGX Station, though I doubt I'll be able to afford one.

    • numba888 a year ago

      > glad to hear about DGX Station

      how much is it expected, my guess barely fits in 5 digits. Would be nice to have something in between Spark and Station. I.e. some desktop withing $20K.

      • numba888 a year ago

        well, "RTX Pro Blackwell" desktop and workstation version looks to be the answer for midrange. Some should be available starting in April.

    • canucker2016 a year ago

      theVerge article says the DGX Spark will cost US$3K, about the same ballpark range as the Framework desktop with the maxed out 128GB RAM option.

      https://www.theverge.com/news/631957/nvidia-dgx-spark-statio...

      10Gb vs 5Gb Ethernet

      1 HDMI 2.1, 4 x USB TypeC vs 1 HDMI 2.1, 2 Displayport, 2 USB TypeC

      • InTheArena a year ago

        The unified memory model w/ CUDA support should be the huge differentiator for this hardware.

        I think a better comparison might be the Mac Studio ultra.

    • nialv7 a year ago

      I think most people (including me) were expecting 384-bit and ~400G/s. Can't believe they went with 256.

      • Gracana a year ago

        Based on the chip count in the promo picture it looked like it was going to be a 256 bit interface, but it'd be unfair to say that it was known at that point. I certainly wish it was better.

john_minsk a year ago

Amazing video of Omniverse. So many announcements!

shmerl a year ago

I'll leave this here for lulz - "Nvidia engineer explains RTX 5000 series": https://www.youtube.com/watch?v=pCyNp8aPcsY

numba888 a year ago

cool robot at the end ;)

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection