this post was submitted on 01 Aug 2024
2233 points (99.0% liked)

Technology

59665 readers
3278 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 4 points 3 months ago* (last edited 3 months ago) (1 children)

With proper framework, decent assertions are possible.

  1. It must cite the source and provide the quote, not just a summary.
  2. An adversarial review must be conducted

If that is done, the work on the human is very low.

That said, it's STILL imperfect, but this is leagues better than one shot question and answer

[–] Aceticon 4 points 3 months ago* (last edited 3 months ago) (1 children)

Except LLMs don't store sources.

They don't even store sentences.

It's all a stack of massive N-dimensional probability spaces roughly encoding the probabilities of certain tokens (which are mostly but not always words) appearing after groups of tokens in a certain order.

And all of that to just figure out "what's the most likely next token", an output which is then added to the input and fed into it again to get the next word and so on, producing sentences one word at a time.

Now, if you feed it as input a long, very precise sentence taken from a unique piece, maybe you're luck and it will output the correct next word, but if you already have all that you don't really need an LLM to give you the rest.

Maybe the "framework" you seek - which is quite akin to a indexer with a natural language interface - can be made with AI, but it's not something you can do with LLMs because their structure is entirely unsuited for it.

[–] [email protected] 2 points 3 months ago* (last edited 3 months ago)

The proper framework does, with data store, indexing and access functions.

The cutting edge work is absolutely using LLMs in post-rag pipelines.

Consumer grade chat interfaces def do not do this.

Edit if you worry about topics like context window, sentence splitting or source extraction, you aren't using a best in class framework any more.