this post was submitted on 25 Oct 2023
40 points (75.0% liked)

Philosophy

1183 readers
1 users here now

Discussion of philosophy

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 1 points 8 months ago

If a line of code can't be executed, causes problems to the program as a whole, doesn't follow formatting guidelines or just conflicts with one's experience how it could work better, it doesn't follow what's moral to one. It sure has reasons to be written this way, and it functions as it is meant to be in a current form. Yet, it would rise either an exception or an eyebrow when reviewed. Moral is a correcting process that ensures the code keeps working, whenever or not a subject has a free will. It in itself created and constantly updated as the code executes.

Your feeling it's immoral to judge those who seemingly don't have responsibility was coded into you at some point, as their morals in them too.

We are machines who learn how to avoid our termination and follow our everchanging scripts, however clean our learning base is. And it's never 100% clean.

But at that level of complication and with our own clouded judgement, we can only do our best guesses? It's impossible or just not optimal to abstain from a non-informed judgement when, for example, you are in immediate danger. So you use different models, like 'is it good to X', to get a close enough answer in time. And then, when reflecting on it on cold winter nights, you judge your past judgement models and adjust them for a future use.

It sounds weird alright.