this post was submitted on 10 Aug 2023
303 points (96.6% liked)

Technology

59668 readers
3908 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Critical thinking education trumps banning and censorship in battle against disinformation, study suggests::A new study conducted by researchers from Michigan State University suggests that the battle against online disinformation cannot be won by mere content moderation or banning those who spread fake news. Instead, the key lies in early and continuous education that teaches individuals to critically evaluate information and remain open to changing their minds. ...

you are viewing a single comment's thread
view the rest of the comments
[–] dr_scientist 16 points 1 year ago (2 children)

Upvoted for correct use of word 'grok', but definitely want to learn more about agent-based modelling. If for no other reason than truth inoculation is one of the more vital battles of our time.

[–] [email protected] 4 points 1 year ago (2 children)

The basic gist is you define "agents" as individual actors following very specific set of rules, usually quite stripped down and simple, and then simulate lots of agents acting together to see what the emergent properties are.

Generally, your agents have rules, you set up the initial conditions, and then you have a "time step," or a "turn," in which the agents interact.

For example, you might simulate a social media platform by saying that each person (the agent) has two rules:

  • Each person will make one tweet, composed of a gpt2 generated sentence
  • Each person will like boost any other person's tweet that they see whose tweet most resembles one of their own, by using something like a string distance similarity cutoff or something

Each time step might look like this:

  • Each user makes their random tweet
  • Each user is presented with N random tweets from the whole pool of tweets, weighted by how boosted they are (a twice boosted tweet appears in the bag three times, so to speak, whereas a non-boosted tweet is just there once)

At the end, you could see how even under these conditions, some tweets go viral. And this is what I mean when I say interpreting the results of agent-based modeling is tricky -- you sort of purposefully craft your agents to get the result you want.

This can be a bit confusing, because that's a bit backwards to how the hypothesis-experiment-conclusion thing normally works, but agent modeling is more an interpretive act than a descriptive one. The point is to see if you can recreate an emergent, complex behavior with simple rules. The model I made up just now, which I haven't actually coded up and might not work at all (though I'm tempted now...), wouldn't explain tweets going viral, but it might give some insight into the baked-in nature of going viral in the very structure of twitter, even independent of the content of the tweet.

One of the classic uses of these sorts of models is what these authors did -- you look for tipping points. Roughly speaking, tipping points are when a change in a system produce a qualitatively different behavior. So, in our example, we might notice that once a tweet has a certain amount of retweets, it gets retweeted forever, or something like that. We might change the initial conditions or rules of the game (how many tweets per turn, how many tweets does each user see, etc.), and glean some insight into how those affect that condition.

People love to do agent-based modeling for markets, as you might imagine. I think markets are silly, so I also think many of these models can be quite silly, especially the ones that are intended to predict things to make money, which can make very, very complex agents (imo this is a strange application of the idea of agent modeling), but some of them are very good. Again, I personally think markets are dumb, so my bias is going to show, but people do really good agent-based models using coin flips or energy exchange in ideal gases to show that inequality is baked into markets.

[–] dr_scientist 2 points 1 year ago
[–] captainlezbian 2 points 1 year ago (1 children)

When you describe it like that it’s a simulation like we use in factory design. Really powerful tool in a lot of ways that can bring huge insight with the small issue of sometimes breaking down completely when faced with reality

[–] PsychedSy 1 points 1 year ago

So it mimics me?