I think it looks badass, I've been watching streams on twitch and it's right up my alley so I'm buying it. If for any reason to piss off the jackasses that are hating just for hates sake.
Defaced
At least Xbox made some sense, it was originally going to be called the DirectXbox, thankfully they shortened the name to something catchier.
It looked good, but the gyro is apparently awful and the trigger travel is basically non-existent.
Well I didn't say they were perfect.
While the 9000 series looks decent, I honestly think Intel has a really interesting platform to build off of with the core ultra chips. It feels like Intel course correcting with poor decisions made for the 13th and 14th gen chips. Wendel from Level1 techs made a really good video about the good things Intel put into the chips while also highlighting some of the bad things, things like a built-in NPU and how they're going to use that to pull in profiles for applications and games with ML, or the fact that performance variance occurs between chipset makers more often with the core ultra. It's basically a step forwards in tech but a step backwards in price/performance.
Not sure you know what swap is...I looked at my m1 after a night of gaming on GeForce now and filling out forms on Google Chrome. My swap was at 0, my used ram was at 4GB used out of the 8GB and didn't show down at all. I'm sorry you have had a terrible experience with your Mac, I love my Mac mini and will enjoy it as a really cool piece of tech.
This is some real doom and gloom shit right here. You really think it's a single-vendor market? I guess if you want to buy a 4090 then yes it is, but everything else is a three horse race. This doesn't mean AMD is saying buy consoles, all this means is they're focusing on the midrange market that I'm 100% confident Nvidia will completely ditch sometime in the future and tell most customers to use Geforce Now if they want midrange prices. To be completely real here, Nvidia only has ray tracing holding them up right now, as soon as the competition catches up they won't have anything to gouge and will be kicking themselves for not really innovating any further.
This is how it's always been for the past 30 years, Nvidia makes a good card and prices it high with 4 or 5 generational updates, Radeon makes a good price/performance midrange card that undercuts Nvidia, everyone wins. The only difference now is Intel has created a very compelling product with their GPUs and I'm pretty confident battle mage will be a big improvement over the current Arc cards and give AMD a run for their money.
Intel is learning from AMD and playing the long game with their hardware, the latest core ultra CPUs are great in Linux vs Windows and will hopefully get better over time, and the battle mage cards will hopefully have day 1 support for Linux and good support for Windows. You simply have to change your expectations here, the market is shifting, Nvidia makes more money hand over fist with AI/ML chips and GFN than they do their consumer graphics cards, they don't have to make their cards cheap anymore to compete, they can price them however they want due to the cost offset of their server market. I personally find the midrange market to be way more compelling these days than overpriced high end Nvidia chips, maybe you should rethink your position as well.
MacOS, no matter what anyone says, has extremely efficient memory management. It's seriously impressive how efficient that OS truly is, and it's no surprise they stuck with 8GB for so long. The thing these clickbait articles don't really bring to light is that the 16GB increase is really for Apple intelligence. If that wasn't a thing these Macs would stick to 8GB.
You know, I was honestly going to give you some credit for trying, but then you edited your post and decided to turn the conversation political. Your entire argument has lost all credibility, these are video games, please try not to take them so seriously, have a nice day.
Still running an m1 Mac mini right now, it's a damn good machine, but the performance gains over the years on the m series chips haven't really forced me to upgrade yet. As for gaming, I just use GeForce now to play my steam library and it's awesome, it's a really great combo. The 8GB of ram is lacking, but I'm using GFN and not pushing it too hard, so I don't notice any meaningful performance problems. I'm also not editing photos or videos, so that probably helps.
There have been more positive reviews so far than negative, and not a single post has shown any proof that EA is manipulating reviews and cherry picking. The only thing we've seen is one guy at fextralife throwing out conspiracy theories about how EA hates him, another guy who's apparently a racist and sexist asshat, and that's pretty much it. Mortismal has even stated he wasn't paid off or anything by EA and would be fined and his account deleted if he was and didn't disclose that fact, which he didn't.
Bioware has made two bad games, Andromeda and Anthem, two. One is objectively not that bad, the other is a game in a genre they've never dipped their toes into, and the biggest issue is that those two releases were back to back, so that apparently means they've gone to shit now and everything else they've created means nothing. It's really sad how petty and ridiculous some people are over bioware. As for the EA hate that's been around forever, but God forbid someone say something positive about a bioware game.
Do you have proof that EA is forcing reviewers to use catchphrases as you've said? I get it, we all have our hate boners, we all have our pet peeves, but damn son....the conspiracy theories and review embargo nonsense is just stupid at this point. Like you said let's just wait for the objective reviews but how about we simply don't write the game off because bioware apparently murdered our puppies.
But you know, according to EA Linux is worse than guys like this deliberately causing disruptions in service to legit players.