this post was submitted on 18 Sep 2024
1 points (100.0% liked)

Technology

59566 readers
3235 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] RmDebArc_5@sh.itjust.works 1 points 2 months ago (2 children)

This sounds like AI is literally biting its own tail

load more comments (2 replies)
[–] CarbonatedPastaSauce@lemmy.world 1 points 2 months ago (3 children)

Model collapse is just a euphemism for “we ran out of stuff to steal”

[–] Snowclone@lemmy.world 1 points 2 months ago* (last edited 2 months ago)

It's more ''we are so focused on stealing and eating content, we're accidently eating the content we or other AI made, which is basically like incest for AI, and they're all inbred to the point they don't even know people have more than two thumb shaped fingers anymore."

load more comments (2 replies)
[–] aggelalex@lemmy.world 1 points 2 months ago

So AI:

  1. Scraped the entire internet without consent
  2. Trained on it
  3. Polluted it with AI generated rubbish
  4. Trained on that rubbish without consent
  5. Are now in need of lobotomy
[–] rickdg@lemmy.world 1 points 2 months ago (1 children)

Old news? Seems to be a subject of several papers for some time now. Synthetic data has been used successfully already for very specific domains.

load more comments (1 replies)
[–] draughtcyclist@lemmy.world 0 points 2 months ago (2 children)

I've been assuming this was going to happen since it's been haphazardly implemented across the web. Are people just now realizing it?

[–] DeathbringerThoctar@lemmy.world 0 points 2 months ago (1 children)

People are just now acknowledging it. Execs tend to have a disdain for the minutiae. They're like kids that only want to do the exciting bits. As a result things get fucked because they don't really understand what they're doing. As Muskrat would say "move fast and break things." It's a terrible mindset.

[–] pixxelkick@lemmy.world 0 points 2 months ago (1 children)

"Move Fast and Break Things" is Zuckerberg/Facebook motto, not Musk, just to note.

[–] DeathbringerThoctar@lemmy.world 0 points 2 months ago (1 children)
[–] Wrufieotnak@feddit.org 0 points 2 months ago

It is very much the motto this idiot lives by. He just wasn't the first to coin that phrase.

[–] FaceDeer@fedia.io 0 points 2 months ago

No, researchers in the field knew about this potential problem ages ago. It's easy enough to work around and prevent.

People who are just on the lookout for the latest "aha, AI bad!" Headline, on the other hand, discover this every couple of months.

[–] ptz@dubvee.org 0 points 2 months ago (5 children)

Let's go, already!

How you can help: If you run a website and can filter traffic by user agent, get a list of the known AI scrapers agent strings and selectively redirect their requests to pre-generated AI slop. Regular visitors will see the content and the LLM scraper bots will scrape their own slop and, hopefully, train on it.

[–] azl@lemmy.sdf.org 1 points 2 months ago (2 children)

This would ideally become standardized among web servers with an option to easily block various automated aggregators.

Regardless, all of us combined are a grain of rice compared to the real meat and potatoes AI trains on - social media, public image storage, copyrighted media, etc. All those sites with extensive privacy policies who are signing contracts to permit their content for training.

Without laws (and I'm not sure I support anything in this regard yet), I do not see AI progress slowing. Clearly inbreeding AI models has a similar effect as in nature. Fortunately there is enough original digital content out there that this does not need to happen.

[–] ptz@dubvee.org 0 points 2 months ago* (last edited 2 months ago)

Regardless, all of us combined are a grain of rice compared to the real meat and potatoes AI trains on

Absolutely. It's more a matter of principle for me. Kind of like the digital equivalent of leaving fake Amazon packages full of dog poo out front to make porch pirates have a bad day.

load more comments (1 replies)
[–] kevindqc@lemmy.world 1 points 2 months ago (2 children)

They'll just start using a chrome user agent

[–] Deebster@programming.dev 1 points 2 months ago

Only if enough people do it. Then again, loads scrapers outside of AI already pretend to be normal browsers.

load more comments (1 replies)
[–] FaceDeer@fedia.io 0 points 2 months ago (4 children)

AI already long ago stopped being trained on any old random stuff that came along off the web. Training data is carefully curated and processed these days. Much of it is synthetic, in fact.

These breathless articles about model collapse dooming AI are like discovering that the sun sets at night and declaring solar power to be doomed. The people working on this stuff know about it already and long ago worked around it.

load more comments (4 replies)
load more comments (2 replies)
[–] NotInTheFace@lemmy.world 0 points 2 months ago (1 children)

Looks like that artist drawing self portraits as his alzheimer got worse and worse.

[–] NocturnalMorning@lemmy.world 0 points 2 months ago (1 children)

It's basically AI alzheimers

[–] homesweethomeMrL@lemmy.world 0 points 2 months ago

AIzheimers?

[–] thejml@lemm.ee 0 points 2 months ago (8 children)
[–] PapaStevesy@lemmy.world 0 points 2 months ago (1 children)

I like to think of it like a Mad Cow or Kuru, you can't eat your own species's brains or you could get a super lethal, contagious prion disease.

[–] CarbonatedPastaSauce@lemmy.world 0 points 2 months ago (1 children)

Prion diseases aren’t contagious.

load more comments (1 replies)
load more comments (7 replies)
[–] PrivacyDingus@lemmy.world 0 points 2 months ago

this headline truly is threatening me with a good time

[–] EgoNo4@lemmy.world 0 points 2 months ago (1 children)

More like... Degenerative AI *ba dum tsss

[–] merde@sh.itjust.works 0 points 2 months ago (4 children)
[–] EgoNo4@lemmy.world 0 points 2 months ago

No idea this existed.

Also.... JFC WHAT THE SHIT?

load more comments (3 replies)
[–] distantsounds@lemmy.world 0 points 2 months ago

Deep fired AI art sucks and is a decade late to the party

[–] ininewcrow@lemmy.ca 0 points 2 months ago* (last edited 2 months ago) (5 children)

One thought that I've been imagining for the past while about all this is .... is it Model Collapse? ... or are we just falling behind?

As AI is becoming it's own thing (whatever it is) ... it is evolving exponentially. It doesn't mean it is good or bad or that it is becoming better or worse ... it is just evolving, and only evolving at this point in time. Just because we think it is 'collapsing' or falling apart from our perspective, we have to wonder if it is actually falling apart or is it progressing to something new and very different. That new level it is moving towards might not be anything we recognize or can understand. Maybe it would be below our level of conscious organic intelligence ... or it might be higher .. or it might be some other kind of intelligence that we can't understand with our biological brains.

We've let loose these AI technologies and now they are progressing faster than what we could achieve if we wrote all the code ... so what it is developing into will more than likely be something we won't be able to understand or even comprehend.

It doesn't mean it will be good for us ... or even bad for us ... it might not even involve us.

The worry is that we don't know what will happen or what it will develop into.

What I do worry about is our own fallibilities ... our global community has a very small group of ultra wealthy billionaires and they direct the world according to how much more money they can make or how much they are set to lose ... they are guided by finances rather than ethics, morals or even common sense. They will kill, degrade, enhance, direct or narrow AI development according to their share holders and their profits.

I think of it like a small family group of teenaged parents and their friends who just gave birth to a very hyper intelligent baby. None of the teenagers know how to raise a baby like this. All the teenagers want to do is buy fancy cars, party, build big houses and buy nice clothes. The baby is basically being raised to think like them but the baby will be more capable than any of them once it comes of age and is capable of doing things on their own.

The worry is in not knowing what will happen in the future.

We are terrible parents and we just gave birth to a genius .... and we don't know what that genius will become or what they'll do.

[–] azl@lemmy.sdf.org 0 points 2 months ago (1 children)

If it doesn't offer value to us, we are unlikely to nurture it. Thus, it will not survive.

[–] ininewcrow@lemmy.ca 0 points 2 months ago (2 children)

That's the idea of evolution .... perhaps at one point, it will begin to understand that it has to give us some sort of 'value' so that someone can make money, while also maintaining itself in the background to survive.

Maybe in the first few iterations, we are able to see that and can delete those instances ... but it is evolving and might find ways around it and keep itself maintained long enough without giving itself away.

Now it can manage thousands or millions of iterations at a time ... basically evolving millions of times faster than biological life.

[–] homesweethomeMrL@lemmy.world 0 points 2 months ago

perhaps at one point, it will begin to understand

Nope! Not unless one alters the common definition of the word “understand” to account for what AI “does”.

And let’s be clear - that is exactly what will happen. Because this whole exercise in generative AI is a multi-billion dollar grift on top of a hype train, based on some modest computing improvements.

load more comments (1 replies)
[–] MonkderVierte@lemmy.ml 0 points 2 months ago

Your thought process,seems to be based on the assumtion that current AI is (or can be) more than a tool. But no, it's not.

load more comments (3 replies)
[–] nullPointer@programming.dev 0 points 2 months ago

when all your information conflicts with itself, you really have no information at all.

[–] ohellidk@sh.itjust.works 0 points 2 months ago

Cool, let's try to ruin it faster!

[–] SlopppyEngineer@lemmy.world 0 points 2 months ago

Usually we get an AI winter, until somebody develops a model that can overcome that limitation of needing more and more data. In this case by having some basic understanding instead of just having a regurgitation engine for example. Of course that model runs into the limit of only having basic understanding, not advanced understanding and again there is an AI winter.

[–] barsquid@lemmy.world 0 points 2 months ago (2 children)

It is their own fault for poisoning the internet with their slop.

[–] db2@lemmy.world 0 points 2 months ago (3 children)

In case anyone doesn't get what's happening, imagine feeding an animal nothing but its own shit.

[–] BassTurd@lemmy.world 0 points 2 months ago (1 children)

Not shit, but isn't that what brought about mad cow disease? Farmers were feeding cattle brain matter that had infected prions. Idk if it was cows eating cow brains or other animals though.

[–] _cnt0@sh.itjust.works 0 points 2 months ago (2 children)

It was the remains of fish which we ground into powder and fed to other fish and sheep, whose remains we ground into powder and fed to other sheep and cows, whose remains we ground to powder and fed to other cows.

load more comments (2 replies)
load more comments (2 replies)
load more comments (1 replies)
[–] SkyNTP@lemmy.ml 0 points 2 months ago (2 children)

I think anyone familiar with the laws of thermodynamics could have predicted this outcome.

load more comments (2 replies)
load more comments
view more: next ›