this post was submitted on 19 Jul 2024
73 points (95.1% liked)

Sysadmin

7542 readers
2 users here now

A community dedicated to the profession of IT Systems Administration

No generic Lemmy issue posts please! Posts about Lemmy belong in one of these communities:
[email protected]
[email protected]
[email protected]
[email protected]

founded 1 year ago
MODERATORS
 

cross-posted from: https://lemmy.ml/post/18154572

All our servers and company laptops went down at pretty much the same time. Laptops have been bootlooping to blue screen of death. It's all very exciting, personally, as someone not responsible for fixing it.

Apparently caused by a bad CrowdStrike update.

you are viewing a single comment's thread
view the rest of the comments
[–] EpicFailGuy 17 points 2 months ago* (last edited 2 months ago) (3 children)

I work on the field. They pushed an update, their software loads a low altitude driver on the kernel at boot time. The driver is causing servers and workstations to crash. The only fix so far is to reboot in safe mode, restart, uninstall and restart again.

Imagine having to do that by hand, on every windows device in your organization.

[–] [email protected] 4 points 2 months ago

There are people on r/sysadmin that have 50,000 machines to deal with. Also a lot of companies have remote workers

[–] teft 3 points 2 months ago* (last edited 2 months ago) (1 children)

Meh, that’s an easy fix compared to some other BSODs I’ve had to deal with.

The every device part is daunting but again, at least it’s an easy fix.

[–] [email protected] 1 points 2 months ago (1 children)

But have those ever been released as an update?

And with the employee to computer ratio only getting worse, this really highlights a lot of issues in the system

[–] teft 1 points 2 months ago (2 children)

But have those ever been released as an update?

What BSOD? Many times.

[–] [email protected] 1 points 2 months ago (1 children)

That's actually really funny, I'm more of a Linux user so I didn't realise how down bad things are over there

[–] [email protected] 1 points 2 months ago

It really isn't.

Or wasn't until yesterday

[–] [email protected] 1 points 2 months ago

Not really if you test things.

You are testing things right?

[–] [email protected] 2 points 2 months ago (1 children)

I was with a large organization as tech support and upper IT pushed out an update that corrupted everybody's certificates to log into the network. Imagine having to talk 40k users, who most of them whined and bitched to us about having to do all this work to fix the computer, through a removal of the old certificate, reboot, get the new one after logging in with a backup account, rebooting again and verifying that they can log in. Each computer was about 20-40 minutes to get done. We only had about 50 of us working at peak hours. It took about 2 months of non stop calls to get them fixed.

[–] EpicFailGuy 3 points 2 months ago

YIKES ... I got one worse still ... I was NOC at a company where one of my friends from Desktop services made a mistake pushing hard drive encryption and basically corrupted the hard drive of a large number em laptops. It wasn't everyone thank god because they were rolling it out in stages ... but it was THOUSANDS and there was no real way to get it back. Every single one had to get re-imaged.