this post was submitted on 25 Aug 2024
1 points (100.0% liked)

Technology

59587 readers
5370 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] jacksilver@lemmy.world 0 points 3 months ago (1 children)

LLMs have been around since roughly 2016. While scaling the up has improved their performance/capabilities, there are fundamental limitations on the actual approach. Behind the scenes, LLMs (even multimodal ones like gpt4) are trying to predict what is most expected, while that can be powerful it means they can never innovate or be truth systems.

For years we used things like tf-idf to vectorize words, then embeddings, now transformers (supped up embeddings). Each approach has it limits, LLMs are no different. The results we see now are surprisingly good, but don't overcome the baseline limitations in the underlying model.

[–] todd_bonzalez@lemm.ee 0 points 2 months ago (1 children)

The "Attention Is All You Need" paper that birthed modern AI came out in 2017. Before Transformers, "LLMs" were pretty much just Markov chains and statistical language models.

[–] jacksilver@lemmy.world 0 points 2 months ago

You're right, I thought that paper came out in 2016.