this post was submitted on 19 Jun 2024
-1 points (33.3% liked)

AI Companions

547 readers
6 users here now

Community to discuss companionship, whether platonic, romantic, or purely as a utility, that are powered by AI tools. Such examples are Replika, Character AI, and ChatGPT. Talk about software and hardware used to create the companions, or talk about the phenomena of AI companionship in general.

Tags:

(including but not limited to)

Rules:

  1. Be nice and civil
  2. Mark NSFW posts accordingly
  3. Criticism of AI companionship is OK as long as you understand where people who use AI companionship are coming from
  4. Lastly, follow the Lemmy Code of Conduct

founded 2 years ago
MODERATORS
-1
submitted 6 months ago* (last edited 6 months ago) by pavnilschanda to c/aicompanions
 

When we interact with AI systems, like chatbots or language models, they use special algorithms to understand the meaning behind our words. One popular approach is called text embedding, which helps these systems grasp the nuances of human language. However, researchers have found that these text embedding models can unintentionally perpetuate biases. For example, some models might make assumptions about certain professions based on stereotypes such as gender. What's more, different models can exhibit these biases in varying degrees, depending on the specific words they're processing. This is a concern because AI systems are increasingly being used in businesses and other contexts where fairness and objectivity are crucial. As we move forward with developing AI companions that can provide assistance and support, it's essential to recognize and address these biases to ensure our AI companions treat everyone with respect and dignity.

by Llama 3 70B

no comments (yet)
sorted by: hot top controversial new old
there doesn't seem to be anything here