this post was submitted on 08 Nov 2023
219 points (96.6% liked)

Linux

48624 readers
1693 users here now

From Wikipedia, the free encyclopedia

Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).

Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.

Rules

Related Communities

Community icon by Alpár-Etele Méder, licensed under CC BY 3.0

founded 5 years ago
MODERATORS
 

Qualcomm brought a company named Nuvia, which are ex-Apple engineers that help designed the M series Apple silicon chips to produce Oryon which exceeds Apple’s M2 Max in single threaded benchmarks.

The impression I get is than these are for PCs and laptops

I’ve been following the development of Asahi Linux (Linux on the M series MacBooks) with this new development there’s some exciting times to come.

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 82 points 1 year ago (3 children)

I'm just eager to know how much laptops will cost with the new Qualcomm chip. I don't want to pop champagne too early only to realize that new ARM laptops cost $2000.

[–] fuckwit_mcbumcrumble 31 points 1 year ago (3 children)

I'd expect them to start around 1k. Not many people are going to be buying these devices so there's no economies of scale.

Also I love how qualcomm announced this CPU and a day later Apple releases the M3 which is finally a real upgrade from the M1.

[–] [email protected] 21 points 1 year ago (1 children)

1k like the Macbook air or 1k with actually good memory and storage specs?

[–] [email protected] 11 points 1 year ago

or gasp something mildly modular you can upgrade if you need to.

[–] [email protected] 9 points 1 year ago (2 children)

Lots of tech companies might be interested. For example, at my work we are now stuck half way between x64 and arm, both on the server side and on the developers side (Linux users are on x64 and Mac users are on arm). While multiarch OCI/docker containers minimize the pains caused by this, it would still be easier to go back to a single architecture.

load more comments (2 replies)
[–] [email protected] 7 points 1 year ago

I’m sure Qualcomm knew what they were doing

[–] [email protected] 11 points 1 year ago* (last edited 1 year ago) (1 children)

New tech always comes at a cost, hopefully with the many manufacturers partnering with Qualcomm in this project we’ll have competitive pricing better than the current offering that Apple silicon provides.

[–] anon_8675309 9 points 1 year ago* (last edited 1 year ago) (1 children)

Used to be, each year-ish computers got faster AND cheaper. So, it doesn’t “always” have to be that way.

[–] [email protected] 5 points 1 year ago

That’s not happening anymore due to real world constraints, though. Dennard scaling combined with Moore’s Law allowed us to get more performance per watt until around 2006-2010, when Dennard scaling stopped applying - transistors had gotten small enough that thermal issues and other current leakage related challenges meant that chip manufacturers were no longer able to increase clock frequencies each generation.

Even before 2006 there was still a cost to new development, though, us consumers just got more of an improvement per dollar a year later than we do now.

[–] [email protected] 4 points 1 year ago

Youre right, just like the first risc-v laptop which was more than 1k with awful performances. This will probably follow the M series trend at about 1,5k , but arm has a lot of competitors...

[–] [email protected] 26 points 1 year ago (1 children)

I hope for Microsoft to just give up and build a new "windows“ which is just an other Linux distro xD

Ducking windows can’t even clone the Linux kernel right now

[–] [email protected] 4 points 1 year ago (2 children)

That'd be based, but I don't think there's anything in that for them.

load more comments (2 replies)
[–] [email protected] 26 points 1 year ago (2 children)

Qualcomm you say?

I'll believe it when it ships

[–] [email protected] 16 points 1 year ago

Qualcomm is my main fear also. They will ship it with lots of closed source firmware digitally signed with their private keys which users can't replace so expect a shitty bootloader and don't forget about always running hypervisior, trust zone and world most kept secret modem

[–] [email protected] 9 points 1 year ago* (last edited 1 year ago)

[This comment has been deleted by an automated system]

[–] xohshoo 23 points 1 year ago (2 children)

don't care about absolute performance, more interested in performance/watt

[–] AnUnusualRelic 8 points 1 year ago

I'm more interested in something that has an actual hardware and software ecosystem. I'm no longer interested in soldering my computer and it's peripherals together.

load more comments (1 replies)
[–] [email protected] 19 points 1 year ago (1 children)

Would definitely upgrade to that instead of my current Lenovo. I want x86 to die already.

[–] semperverus 26 points 1 year ago (2 children)

If you want to kill x86, you need to do what Valve and the Wine foundation did with Proton/WINE (mostly proton at this point though), but for x86 to ARM and maybe other architectures like RISCV (especially because the milkV pioneer is a thing).

There is too much legacy software that will never be converted that people still use to this day. Once you make it easy to transition, it will slowly but steadily start to happen.

Box86/Box64 are promising, but need help from contributors like you. If you want it to happen, go make it happen, or continue to live in the world you have now.

[–] [email protected] 4 points 1 year ago

Well legacy software is fine, that stuff mostly runs on old machines/servers/etc. ARM will be more easily to move towards by focusing the consumer market, where legacy issue is less of an issue because their programs are frequently updated. Some old server using outdated software that people are afraid to touch, we don't need to worry about converting that lol.

[–] [email protected] 12 points 1 year ago

As long as memory and ssd are upgradable and not soldered on the board, I would buy this laptop

[–] OscarRobin 12 points 1 year ago

The limited benchmarks I've seen put the new X Elite at slightly less efficient than the M2 Pro (let alone M3 Pro). It only gets marginally higher scores when operating at 3x the wattage.

Also, let's not imagine even for a second that notoriously terrible ARM are going to make it easy to support this chip, especially not in the long term.

[–] Parastie 10 points 1 year ago (10 children)

The benchmarks for the M3 have the single core and multicore performances way past similar Intel and AMD chips. Qualcomm's mobile chips are still no where near Apple's mobile chips. I do not believe for a second that Qualcomm will catch up to the M2 on their first release.

[–] [email protected] 17 points 1 year ago* (last edited 1 year ago)

That's absolutely not true. The M3 Max just about brings Apple performance up to similar levels as Intel and AMD. The Ryzen 9 7945HX3D for example is a laptop processor which trades blows with the M3 on benchmarks - single core the M3's slightly faster and multi core the Ryzen's slightly faster - and in performance per watt the Ryzen's marginally better. So really it's just catching up with older laptop processors from other manufacturers.

And if you venure outside the laptop space to compare ultimate speed it's nowhere near the fastest, particularly in multi-threaded. Its multi-threaded performance is around 13% of the AMD EPYC 9754 Bergamo for example.

[–] fuckwit_mcbumcrumble 9 points 1 year ago

Keep in mind this is with up to an 80 watt TDP vs an effectively 3 year old architecture in a select few tests. The M2 was basically just an overclocked M1, with the Pro/Max models getting 2 extra cores. This is qualcomms best case scenario.

[–] [email protected] 4 points 1 year ago

Apple cooks only with water too.

[–] [email protected] 3 points 1 year ago

The Qualcomm x elite benchmarks as faster than the M3 for multicore. Not too surprising as I think it's 12 cores vs 10. For single core its something like 2700 vs 3200.

Laptops running x elite are supposed to be available mid 2024.

load more comments (6 replies)
[–] [email protected] 9 points 1 year ago (1 children)

I don't wanna repeat myself, but: 7840u for the next few years, then I hope RISC V will be mature enough to kick some ass (and that framework releases a board for it).

That's all I dream of.

[–] semperverus 5 points 1 year ago

Check out the milkV Pioneer.

[–] anon_8675309 7 points 1 year ago (2 children)

I see a bunch of lawsuits in the future. Because that’s what big companies do.

[–] [email protected] 2 points 1 year ago

They don't compete, they sue or buy out.

load more comments (1 replies)
[–] kittenzrulz123 5 points 1 year ago

If I can get decent performance on a tablet style laptop for cheap I would buy it new

[–] drkhrse96 5 points 1 year ago* (last edited 1 year ago) (1 children)

It’s interesting as a comparison to M3 now and at different power limits. I’m hoping it may hopefully benefit the asahi project also. As a windows product I don’t think it’ll be good at all unless Microsoft has a Rosetta like emulation layer that is nearly as good as Apple. Without that this product will not do well.

[–] fuckwit_mcbumcrumble 5 points 1 year ago

Microsoft has a pretty good translation layer, it's the hardware x86 acceleration that most windows ARM chips lack, that Apple's CPUs have.

[–] [email protected] 2 points 1 year ago* (last edited 1 year ago) (1 children)

I can’t wait for the hardware Android continuity… that’s the only thing I’m waiting on now to switch to Android besides the raw performance being equal.

load more comments (1 replies)
load more comments
view more: next ›