this post was submitted on 27 Feb 2024
427 points (98.2% liked)

Technology

59597 readers
3377 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

cross-posted from: https://sopuli.xyz/post/9700996

Nvidia's AI customers are scared to be seen courting other AI chipmakers for fear of retaliatory shipment delays, says rival firm

you are viewing a single comment's thread
view the rest of the comments
[–] utopiah 1 points 8 months ago

If the major cloud providers end up using their own custom silicon, that’s a major blow for Nvidia.

They already but AFAIK not at scale. What I believe, but that's my intuition I don't have numbers to back that up, is that Alphabet for GCP, Microsoft for Azure, Amazon for AWS and others do design their own chips, their own racks, etc but mostly do it as promotional R&D. They do want to show investors that they are acutely aware of their dependency on NVIDIA and thus try to be more resilient by having alternatives. What is still happening though is that in terms of compute per watt and thus per dollar, NVIDIA through its entire stack, both hardware (H100, A100, 40xx, etc) and software (mostly CUDA here) but also trust from CTOs, is the de facto standard. Consequently my bet is that GCP, Azure, AWS do have their custom silicon running today but it let's than 1% of their compute and they probably even provide it at a discount price to customers. It's a bit like China and their billions poured into making their own chips. Sure they are showing that they can (minus the dependency on ASML...) but at what cost? Making some chipset at equivalent performance with the state of the art is a research feat not to be downplayed but doing it at scale in a commercially competitive way is quite different.

Anyway that's just my hunch so if anybody has data to contradict that please do share.