this post was submitted on 20 Sep 2023
528 points (93.3% liked)

Technology

59665 readers
3831 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

The majority of U.S. adults don't believe the benefits of artificial intelligence outweigh the risks, according to a new Mitre-Harris Poll released Tuesday.

you are viewing a single comment's thread
view the rest of the comments
[–] nandeEbisu 37 points 1 year ago (3 children)

AI is such a huge term. Google lens is great, when I'm travelling I can take a picture of text and it will automatically get translated. Both of those are aided by machine learning models.

Generative text and image models have proven to have more adverse affects on society.

I think we're at a point where we should start normalizing using more specific terminology. It's like saying I hate machines, when you mean you hate cars, or refrigerators or air conditioners. It's too broad of a term to be used most of the time.

[–] [email protected] 16 points 1 year ago (2 children)

Yeah, I think LLMs and AI art have overdominated the discourse to the degree that some people think they're the only form of AI that exists, ignoring things like text translation, the autocompletion of your phone keyboard, Photoshop intelligent eraser, etc.

Some forms of AI are debatable of their value (especially in their current form). But there's other types of AI that most people consider highly useful and I think we just forget about it because the controversial types are more memorable.

[–] nandeEbisu 5 points 1 year ago

AI is a tool, its value is dependent on whatever the application is. Transformer architectures can be used for generating text or music, but they were also originally developed for text translation which people have fewer qualms with.

[–] [email protected] 3 points 1 year ago

ignoring things like text translation, the autocompletion of your phone keyboard, Photoshop intelligent eraser, etc.

AFAIK two of those are generative AI based or as you said LLMs and AI art

[–] [email protected] 1 points 1 year ago (3 children)

Be the trend setter. What slang would you use(that I'll use)?

[–] [email protected] 5 points 1 year ago

Why not the type of AI? In that.case, LLM.

[–] nandeEbisu 4 points 1 year ago

Its not a matter of slang, its referring to too broad of a thing. You don't need to go as deep as the type of model, something like AI image generation, or generative language models is what you would refer to. We'll hopefully start converging on shorthand from there for specific things.

[–] kicksystem 3 points 1 year ago

I'd like people to make a distinction between AI and machine learning, machine learning and neural networks (the word deep is redundant nowadays). And then have some sense of different popular types of neural nets: GANs, CNN, Transformer, stable diffusion. Might be nice if people know what is supervised unsupervised and reinforcement learning. Lastly people should have some sense of the difference between AI and AGI and what is not yet possible.

[–] [email protected] -1 points 1 year ago (1 children)

Chat GPT needs to be vastly improved pr thrown out to dry

[–] nandeEbisu 4 points 1 year ago

I'm kind of surprised people are more concerned with the output quality for chatGPT, and not where they source their training set from, like for image models.

Language models are still in a stage where they aren't really a product by themselves, they really need to be cajoled into becoming a good product, like looking up context via a traditional search and feeding it to the model, or guiding it towards solving problems. That's more of a traditional software problem that leverages large language models.

Even the amount of engineering to go from text prediction model trained on a bunch of articles to something that infers you should put an answer after a question is a lot of work.