this post was submitted on 05 May 2024
24 points (96.2% liked)
TechTakes
1493 readers
185 users here now
Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.
This is not debate club. Unless it’s amusing debate.
For actually-good tech, you want our NotAwfulTech community
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Orange Site denizen plays Dr. LLM: https://news.ycombinator.com/item?id=40331850
The example "research paper" was some useless fluff about diabetes, based off an existing data set (read: actual work produced by actual humans), and mad-libs.
I'm too sleepy and statistics-impaired to check how nonsensical the regression "analysis" or findings are, so instead let's check out the references (read: the actual humans who were plagarized to make this fluff)!
Reference #5
This incredibly managed to mangle all non-English alphabet names:
I guess AI has an easier time advancing science than producing a PDF with non-ascii text in it
The fact that actual engineers have been trying to educate newcomers on Unicode for at least 20 years and not only is it still pervasively ignored but the hottest, newest, cutting edge AI that Will Change Everything™ with billions of dollars and so many manhours behind it gets absolutely dumbfounded when it sees
é
is the exact combination of funny and sad that will eventually result in me turning into a Butlerian Jihad Joker.As a C++ programmer I've never been so offended by something I so entirely agree with.
hmm. I can guess at a few reasons this could be happening: model coders "normalizing" everything to flat-ascii in training, or similar happening at training stage (because of the previously-referenced RLHF datamills employing only people with specific localized dialects, instead of wider context-local languages), etc.
wonder if this particular thing is a confluence of those, or just one specific set
have you ever met an English-native dev who didn't need to be trained out of the world being 7-bit ascii
I got lucky and largely missed out on having to deal with those, at a guess largely because of location and age. the type I got to deal with instead were the php-/perl-brained "everything is just a string" types
hell, (circa 2010) I had beers with someone once who was really into Tcl and Second Life, and wanted to be uploaded as a digital consciousness. way before I knew about the other nutjobs, but in retrospect I now have a couple other questions I might've wanted to ask at the time...
@dgerard
7 bits were good enough for Jesus.
First efforts at bible digitization seems incredibly poorly documented online, and from a casual inspection in google scholar, not very well referenced. It's a pity it sounds like a fascinating topic, though 7 bits is likely for the first english versions yes (And according to this there are horrid 7-bits encodings for the ancient greek)
My Jesus wanted characters for drawing borders and playing card suits, which is why He handed down to us Code Page 437. Using the upper 128 characters for things like vowels with funny marks on them is catholic heresy (nuts to Latin 1, down with Unicode).