trollbearpig

joined 1 year ago
[–] trollbearpig 9 points 2 months ago* (last edited 2 months ago)

If you actually read the article you will see that they tested both allowing the students to ask for answers from the LLM, and then limiting the students to just ask for guidance from the LLM. In the first case the students did significantly worse than their peers that didn't use the LLM. In the second one they performed the same as students who didn't use it. So, if the results of this study can be replicated, this shows that LLMs are at best useless for learning and most likely harmful. Most students are not going to limit their use of LLMs for guidance.

You AI shills are just ridiculous, you defend this technology without even bothering to read the points under discussion. Or maybe you read an LLM generated summary? Hahahaha. In any case, do better man.

[–] trollbearpig -1 points 3 months ago (1 children)

Any arguments to defend your position? I'm giving you a very clear example of the awful consecuences of following that path. And the same applies to any creative work. You are just being dismissive without proposing any real solution. Do better man.

[–] trollbearpig -1 points 3 months ago* (last edited 3 months ago) (3 children)

And that's the whole point of my comment, did you even read it? To summarize, there is currently a loophole in law that allows these bullshit arguments about it being different than straight up copying shit (though this haven't been litigated yet, so it's not yet clear if these arguments are actually valid). This means that while a person reading my AGPL code and copying it (without following the license) is 100% illegal, doing the same through an LLM may be legal. So this means that open source licenses can be bypassed by first training an LLM with the code and then extracting the code from the LLM. This is terrible for open source, and in general for anyone who wants to make a living from creating copyrighted work. So we should close this loophole, and I'm glad there is a push to close this through better laws. Even if these laws are comming from Disney, Sony, and all those awful companies.

So again, what's the point you are trying to make here? That we shouldn't make these laws stronger to prevent this bullshit? I honestly don't understand what you are trying to argue here, nothing of what you have said has anything to do with this conversation.

[–] trollbearpig -1 points 3 months ago (5 children)

What point are you trying to make? That the fact that someone can break the law means we should not have laws? I honestly don't get what you are trying to say.

[–] trollbearpig -1 points 3 months ago (7 children)

An engineer at AWS can already just copy your code, make minor modifications, and use it.

You are 100% wrong here my man. If an engineer does this they are creating a derivative work and they have to fullfil the conditions of the license of the code. No wonder you don't see anything wrong here, you AI people live in a fantasy world when it comes to how copyright works hahahaha. Please stop talking about shit you know nothing about.

[–] trollbearpig 0 points 3 months ago* (last edited 3 months ago) (10 children)

Nah my man, you are either brainwashed or are being paid hahaha. Is copyright a mess? Of fucking course, I haven't meet a single person (except crazy ass libertarians funnily enough hahaha) that likes copyright. Are big corporations using copyright to exploit artists, create monopolies, and generally being dicks? Again, of fucking course.

But anyone, like you, saying that we should just let AIs destroy copyright effectively is a fucking prick, that simple. And your agruments are dissingenous at best or outright lies. For example, just as big copyright holder companies are pushing to strengthen copyright law, the big tech companies are pushing for effectively destroying copyright through AI models. I have seen you pushing in multiple thread for open source models like that's a solution. But if you were a serious person researching about the software open source community you would see that pretty much no one there agrees with your position because it would effectively destroy the copyleft open source licenses. After all, if an "AI" model, open source or not, is allowed to just "train" on my AGPL code and spit it back (with minor modifications at best) to an engineer in AWS that's it for my project. Amazon will do the Amazon thing and steal the project. So say goodbye to any software freedom we have.

And let's be 100% clear here, this is not being pushed by the evil copyright holders like you seem to imply (and they are totally evil just to be clear hahahah). This is being pushed by the big tech companies and people like you spreading their propaganda. The fact that the copyright holders happen to be in the right this time is just a broken clock being right and all that, but it's still good that they are pushing back to big tech. I do agree we have to keep an eye on them, the objective here can't be to make copyright bigger, just to close the "loophole" that big tech companies are exploting to steal everything.

People like you who want to destroy copyright without offering any alternatives to allow creatives to work in a market are either missinformed or just assholes. Again, of fucking course it's not an ideal system, but going full kamikaze and just destroying any possibility for artists and creatives of making a living with their work is the most evil thing goung on right now, so bad that the big copyright holders happen to fall on the less bad side this time hahaha. And all for what? So people can be lied to by dumb chatbots? Or so people can create mediocre derivative "art" without putting any effort? Or so we can get mediocre code autocomplete that is subtly wrong all the time? Is fucking ridiculous.

[–] trollbearpig -1 points 3 months ago (12 children)

The guy you are replying to is in all AI posts defending AIs. He is probably heavily invested in this BS or being paid for it, don't waste your time with him.

[–] trollbearpig 1 points 3 months ago

Jesus man, chill. Why are all AI people so sensitive? Hahahaha. My man, during this conversation I have only asked about what are the great apps that LLMs have provided. You answered with the usual ones, chatgpt and copilot. It's nice that you find them useful, no need to insult me just because I don't think they are useful. I was honestly hoping for something else, but that's it. Seriously, chill dude.

[–] trollbearpig 1 points 3 months ago (2 children)

So literally you use it for information retrieval hahahaha. I did use copilot, codium, and the jetbrains one for a bit. But I had to disable each one, the amount fo wrong code simply doesn't justify the little boilerplate it generates.

[–] trollbearpig 1 points 3 months ago (4 children)

Me? I'm not using LLMs at all hahaha. I'm asking you, who says they have great value, to provide examples of their uses. I just provided pretty much the only one I have heard, which some random dude told me in a different thread. Everyone else, like you, just keeps it abstract and just bullshits and bullshits hahaha.

[–] trollbearpig 1 points 3 months ago (6 children)

I always ask all people defending AI, or rather LLMs, what's the great value they all mention in their comments. So far the "best" answer I got was one dude using LLMs to extract info from decades old reports that no one has checked in 20 years hahaha. So glad we are allowing LLMs to deetroy the environment and plagiarize all creative work for that lol.

So, what is the great value you see man?

[–] trollbearpig 11 points 4 months ago* (last edited 4 months ago) (3 children)

Probably too late, but just to complement what others have said. The UEFI is responsible for loading the boot software thst runs when the computer is turned on. In theory, some malware that wants to make itself persistent and avoid detection could replace/change the boot software to inject itself there.

Secure boot is sold as a way to prevent this. The way it works, at high level, is that the UEFI has a set of trusted keys that it uses to verify the boot software it loads. So, on boot, the UEFI check that the boot software it's loading is signed by one of these keys. If the siganture check fails, it will refuse to load the software since it was clearly tampered with.

So far so good, so what's the problem? The problem is, who picks the keys that the UEFI trusts? By default, the trusted keys are going to be the keys of the big tech companies. So you would get the keys from Microsoft, Apple, Google, Steam, Canonical, etc, i.e. of the big companies making OSes. The worry here is that this will lock users into a set of approved OSes and will prevent any new companies from entering the field. Just imagine telling a not very technical user that to install your esoteric distro they need to disable something called secure boot hahaha.

And then you can start imagining what would happen if companies start abusing this, like Microsoft and/or Apple paying to make sure only their OSes load by default. To be clear, I'm not saying this is happening right now. But the point is that this is a technology with a huge potential for abuse. Some people, myself included, believe that this will result in personal computers moving towards a similar model to the one used in mobile devices and video game consoles where your device, by default, is limited to run only approved software which would be terrible for software freedom.

Do note that, at least for now, you can disable the feature or add custom keys. So a technical user can bypass these restrictions. But this is yet another barrier a user has to bypass to get to use their own computer as they want. And even if we as technical users can bypass this, this will result in us being fucked indirectly. The best example of this are the current Attestation APIs in Android (and iOS, but iOS is such a closed environment that it's just beating a dead horse hahahah). In theory, you can root and even degoogle (some) android devices. But in practice, this will result in several apps (banks in particular, but more apps too) to stop working because they detect a modified device/OS. So while my device can technically be opened, in practice I have no choice but to continue using Google's bullshit. They can afford to do this because 99% of users will just run the default configuration they are provided, so they are ok with losing the remaining users.

But at least we are stopping malware from corrupting boot right? Well, yes, assuming correct implementations. But as you can see from the article that's not a given. But even if it works as advertised, we have to ask ourselves how much does this protect us in practice. For your average Joe, malware that can access user space is already enough to fuck you over. The most common example is ransonware that will just encrypt your personal files wothout needing to mess with the OS or UEFI at all. Similarly a keylogger can do its thing without messing with boot. Etc, etc. For an average user all this secure boot thing is just security theater, it doesn't stop the real security problems you will encounter in practice. So, IMO it's just not worth it given the potential for abuse and how useless it is.

It's worth mentioning that the equation changes for big companies and governments. In their case, other well funded agents are willing to invest a lot of resources to create very sofisticated malware. Like the malware used to attack the nuclear program plants in Iran. For them, all this may be worth it to lock down their software as much as possible. But they are playing and entirely different game than the rest of us. And their concerns should not infect our day to day lives.

view more: next ›