this post was submitted on 18 Dec 2024
11 points (92.3% liked)

Hardware

758 readers
182 users here now

All things related to technology hardware, with a focus on computing hardware.


Rules (Click to Expand):

  1. Follow the Lemmy.world Rules - https://mastodon.world/about

  2. Be kind. No bullying, harassment, racism, sexism etc. against other users.

  3. No Spam, illegal content, or NSFW content.

  4. Please stay on topic, adjacent topics (e.g. software) are fine if they are strongly relevant to technology hardware. Another example would be business news for hardware-focused companies.

  5. Please try and post original sources when possible (as opposed to summaries).

  6. If posting an archived version of the article, please include a URL link to the original article in the body of the post.


Some other hardware communities across Lemmy:

Icon by "icon lauk" under CC BY 3.0

founded 2 years ago
MODERATORS
top 12 comments
sorted by: hot top controversial new old
[–] vzq 4 points 5 hours ago

The CUDA moat is pretty deep, but the primitive are starting to solidify and almost no one uses CUDA directly. Increasingly popular libraries are going multi-backend (thanks Apple silicon).

My guess is that as soon as cheap accelerators with LARGE memory banks hit the market, the libraries will support whatever API those need and CUDA dominance will be essentially shattered forever.

But we are not there yet because making good numerical hardware is fucking hard.

[–] anamethatisnt 5 points 6 hours ago* (last edited 6 hours ago) (2 children)

Had a discussion with @[email protected] touching on this over at [email protected] yesterday. (https://lemmy.world/post/23245782 )
Well, inexperienced me asked bruce questions to be exact.
The most interesting part for me would be how the rumored clamshell ARC gpus could upset the balance if the price is right.
If a 24gb b580 or 32gb b770 for a much lower price then nvidia/amd offerings is available how would that affect market share and software development in the field?

[–] brucethemoose 3 points 5 hours ago* (last edited 5 hours ago)

An important thing to note is Intel does not have an enterprise class GPU anymore, and appear to have abandoned most plans for such a thing.

So while consumer GPU inference is great, and would seed support for their future laptop/desktop IGPs, Intel is not in the same boat as AMD anymore, who's consumer efforts would seed support for the enterprise MI300X.

I see basically zero chance of Intel supplanting CUDA for this reason, especially if they don't foster cooperation with anyone else.

[–] Alphane_Moon 2 points 6 hours ago (1 children)

I can't speak for the nitty-gritty details and enterprise-scale technology, but from a consumer perspective (for local ML upscale and LLM using both proprietary and free tools), Nvidia clearly has the upper hand in terms of software support.

[–] anamethatisnt 2 points 6 hours ago (2 children)

How cheap must rivalling high vram offerings be to upset the balance and move devs towards Intel/AMD?
Do you think their current platform offerings are mature enough to grab market share with "more for less" hardware or is the software support advantage just too large?

[–] vzq 3 points 5 hours ago (1 children)

They need to be substantially cheaper and (more importantly) they need loads more memory.

The problem is that everyone (chiefly nvidia, but not only) is afraid to hurt their professional offerings by introducing consumer grade ML cards. They are not afraid of Joe having to use a smaller model to do AI on his security cameras, they are afraid of large companies ditching all their A100 cards for consumer equipment.

So they try and segment the market any way they can think of and Joe gets screwed.

It’s classic market failure really.

[–] brucethemoose 4 points 5 hours ago

The bizarre thing about this is that AMD's workstation card volume is comically small, and Intel's is probably nonexistant.

On the high end... Intel literally discontinued their HPC GPUs. The AMD MI300X is doing OK, but clearly suffering from a lack of grassroots software support.

WTF are they afraid of losing?

[–] Alphane_Moon 2 points 6 hours ago (2 children)

From my limited consumer-level perspective, Intel/AMD platforms aren't mature enough. Try looking into any open-source/commercial ML software aimed at consumers, Nvidia is guaranteed and first class.

The situation is arguably different in gaming.

[–] brucethemoose 4 points 5 hours ago* (last edited 5 hours ago)

Intel is not as bad in LLM land as you'd think. Llama.cpp support gets better every day.

Nvidia may be first class, but in this case, it doesn't matter if the model you want doesn't fit in VRAM. I'd trade my 3090 for a 48GB Arc card without even blinking, even if the setup is an absolute pain.

[–] anamethatisnt 2 points 6 hours ago (1 children)

Thanks for the insight. Kinda sad how selfhosted LLM or ML means Nvidia is a must have for the best experience.

[–] brucethemoose 2 points 5 hours ago

Only because AMD/Intel aren't pricing competitively. I define "best experience" as the largest LLM/context I can fit on my GPU, and right now that's essentially dictated by VRAM.

That being said, I get how most wouldn't want to go through the fuss of setting up Intel/AMD inference.

[–] Alphane_Moon 3 points 6 hours ago

Some interesting points at the attempts to overcome Nvidia's CUDA dominance (and the significant challenges to doing so), but the real answer to this question will be market share (if the AI hype doesn't die down by the point AMD and Intel make signiciant improvements to their software stack).