this post was submitted on 14 May 2024
1353 points (99.1% liked)

Programmer Humor

31434 readers
1647 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS
 
top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 286 points 2 months ago (4 children)

Maybe it's time we invent JPUs (json processing units) to equalize the playing field.

[–] seaQueue 218 points 2 months ago (3 children)

The best I can do is an ML model running on an NPU that parses JSON in subtly wrong and impossible to debug ways

[–] Aceticon 59 points 2 months ago (1 children)

Just make it a LJM (Large JSON Model) capable of predicting the next JSON token from the previous JSON tokens and you would have massive savings in file storagre and network traffic from not having to store and transmit full JSON documents all in exchange for an "acceptable" error rate.

[–] seaQueue 14 points 2 months ago

Hardware accelerated JSON Markov chain operations when?

[–] [email protected] 56 points 2 months ago (1 children)

You need to make sure to remove excess whitespace from the JSON to speed up parsing. Have an AI read the JSON as plaintext and convert it to a handwriting-style image, then another one to use OCR to convert it back to text. Trailing whitespace will be removed.

load more comments (1 replies)
load more comments (1 replies)
[–] [email protected] 34 points 2 months ago* (last edited 2 months ago) (2 children)

Latest Nvidia co-processor can perform 60 million curly brace instructions a second.

load more comments (2 replies)
[–] [email protected] 16 points 2 months ago
load more comments (1 replies)
[–] Randelung 161 points 2 months ago (4 children)

Well, do you have dedicated JSON hardware?

[–] [email protected] 104 points 2 months ago (2 children)

Please no, don't subsidize anything Java-Script. It will only make it less efficient.

[–] [email protected] 59 points 2 months ago (5 children)

And thus JsPU was born from Lemmy

[–] [email protected] 21 points 2 months ago (1 children)
load more comments (1 replies)
load more comments (4 replies)
[–] [email protected] 27 points 2 months ago (3 children)

My thoughts on software in general over the past 20 years. So many programs inefficiently written and in 4th level languages just eats up any CPU/memory gain. (Less soap box and more of a curious what if to how fast things would be if we still wrote highly optimized programs)

[–] [email protected] 36 points 2 months ago* (last edited 2 months ago)

Answer: there'd be far less software in the world, it would all be more archaic and less useful, and our phones and laptops would just sit at 2% utilization most of the time.

There's an opportunity cost to everything, including fussing over whether that value can be stored as an int instead of a double to save 8 bits of space. High level languages let developers express their feature and business logic faster, with fewer bugs, and much lower ongoing maintenance costs.

[–] InternetCitizen2 19 points 2 months ago

Even if our apps used resources better the adware will just use the free space.

load more comments (1 replies)
[–] [email protected] 25 points 2 months ago (7 children)
[–] Randelung 25 points 2 months ago (3 children)

The R in ARM and RISC is a lie.

[–] [email protected] 15 points 2 months ago* (last edited 2 months ago)

The website title says "Arm Developer", not "ARM Developer", in a clearly non-acronym way so it's a guide for making prosthetic hardware. Of course you want a cyborg arm to parse JS natively, why else even get one?

load more comments (2 replies)
load more comments (6 replies)
[–] [email protected] 15 points 2 months ago
load more comments (1 replies)
[–] [email protected] 118 points 2 months ago (5 children)

Everybody gangsta still we invent hardware accelerated JSON parsing

[–] Overtheveloper 97 points 2 months ago (2 children)

https://ieeexplore.ieee.org/document/9912040 "Hardware Accelerator for JSON Parsing, Querying and Schema Validation" "we can parse and query JSON data at 106 Gbps"

[–] ByteJunk 29 points 2 months ago (1 children)

I'm so impressed that this is a thing

load more comments (1 replies)
[–] vvvvv 26 points 2 months ago (2 children)

106 Gbps

They get to this result on 0.6 MB of data (paper, page 5)

They even say:

Moreover, there is no need to evaluate our design with datasets larger than the ones we have used; we achieve steady state performance with our datasets

This requires an explanation. I do see the need - if you promise 100Gbps you need to process at least a few Tbs.

[–] neatchee 13 points 2 months ago (3 children)

Imagine you have a car powered by a nuclear reactor with enough fuel to last 100 years and a stable output of energy. Then you put it on a 5 mile road that is comprised of the same 250 small segments in various configurations, but you know for a fact that starts and ends at the same elevation. You also know that this car gains exactly as much performance going downhill as it loses going uphill.

You set the car driving and determine that, it takes 15 minutes to travel 5 miles. You reconfigure the road, same rules, and do it again. Same result, 15 minutes. You do this again and again and again and always get 15 minutes.

Do you need to test the car on a 20 mile road of the same configuration to know that it goes 20mph?

JSON is a text-based, uncompressed format. It has very strict rules and a limited number of data types and structures. Further, it cannot contain computational logic on it's own. The contents can interpreted after being read to extract logic, but the JSON itself cannot change it's own computational complexity. As such, it's simple to express every possible form and complexity a JSON object can take within just 0.6 MB of data. And once they know they can process that file in however-the-fuck-many microseconds, they can extrapolate to Gbps from there

load more comments (3 replies)
load more comments (1 replies)
load more comments (4 replies)
[–] 2deck 103 points 2 months ago (3 children)

Render the json as polygons?

[–] Dasnap 83 points 2 months ago (2 children)

It's time someone wrote a JSON shader.

[–] [email protected] 42 points 2 months ago (1 children)
load more comments (1 replies)
load more comments (1 replies)
[–] Lumisal 17 points 2 months ago (1 children)

That just results in an image of JSON Bourne.

load more comments (1 replies)
[–] BlessedDog 16 points 2 months ago
[–] [email protected] 77 points 2 months ago (1 children)

That is sometime the issue when your code editor is a disguised web browser 😅

load more comments (1 replies)
[–] [email protected] 55 points 2 months ago* (last edited 2 months ago) (1 children)

there are simd accelerated json decoders

[–] manmachine 66 points 2 months ago (2 children)

every day we stray further from god

[–] [email protected] 22 points 2 months ago (4 children)

Don't worry, they still make extensive use of regexes.

load more comments (4 replies)
load more comments (1 replies)
[–] model_tar_gz 44 points 2 months ago (3 children)

Would you rather have 100,000 kg of tasty supreme pizza, or 200 kg of steaming manure?

Choose wisely.

[–] [email protected] 28 points 2 months ago (4 children)

200kg of steaming manure would be pretty sweet if you had a vegetable garden

[–] ByteJunk 13 points 2 months ago (1 children)

Not sure if I'm just missing a reference here, but if you choose the pizza you can have both.

load more comments (1 replies)
load more comments (3 replies)
load more comments (2 replies)
[–] [email protected] 42 points 2 months ago (1 children)

CPU vs GPU tasks I suppose.

[–] Potatos_are_not_friends 41 points 2 months ago (2 children)

GPU, render my 4.2 MB json file!

[–] [email protected] 19 points 2 months ago

I'm afraid I can't do that, Dave

load more comments (1 replies)
[–] [email protected] 36 points 2 months ago (4 children)

I have the same problem with XML too. Notepad++ has a plugin that can format a 50MB XML file in a few seconds. But my current client won't allow plugins installed. So I have to use VS Code, which chokes on anything bigger than what I could do myself manually if I was determined.

[–] seaQueue 12 points 2 months ago (1 children)

Time to train an LLM to format XML and hope for the best

[–] [email protected] 21 points 2 months ago (1 children)

Do we need a "don't parse xml with LLM" copypasta?

[–] [email protected] 14 points 2 months ago (2 children)
load more comments (2 replies)
load more comments (3 replies)
[–] [email protected] 24 points 2 months ago

Someone just needs to make a GPU-accelerated JSON decoder

[–] [email protected] 24 points 2 months ago (1 children)
[–] [email protected] 12 points 2 months ago (9 children)

Except if it's a single line file, only god can help you then. (Or running prettier -w on it before opening it or whatever.)

load more comments (9 replies)
[–] [email protected] 15 points 2 months ago (15 children)
load more comments (14 replies)
load more comments
view more: next ›