this post was submitted on 20 Apr 2025
1235 points (95.7% liked)

Fuck AI

2537 readers
1234 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 1 points 6 days ago (1 children)

I think his argument is that the models initially needed lots of data to verify and validate their current operation. Subsequent advances may have allowed those models to be created cleanly, but those advances relied on tainted data, thus making the advances themselves tainted.

It's not true; you can just train a model from the ground up on properly licensed or open data, you don't have to inherit anything. What you're talking about is called finetuning which is where you "re-train" a model to do something specific because it's much cheaper than training from the ground up.

[–] [email protected] 7 points 6 days ago (1 children)

I don't think that's what they are saying. It's not that you can't now, it's that initially people did need to use a lot of data. Then they found tricks to improve training on less, but these tricks came about after people saw what was possible. Since they initially needed such data, their argument goes, and we wouldn't have been able to improve upon the techniques if we didn't know that huge neutral nets trained by lots of data were effective, then subsequent models are tainted by the original sin of requiring all this data.

As I said above, I don't think that subsequent models are necessarily tainted, but I find it hard to argue with the fact that the original models did use data they shouldn't have and that without it we wouldn't be where we are today. Which seems unfair to the uncompensated humans who produced the data set.

[–] [email protected] 2 points 6 days ago (1 children)

I actually think it's very interesting how nobody in this community seems to know or understand how these models work, or even vaguely follow the open source development of them. The first models didn't have this problem, it was when OpenAI realized there was money to be made that they started scraping the internet and training illegally and consequently a billion other startups did the same because that's how silicon valley operates.

This is not an issue of AI being bad, it's an issue of capitalist incentive structures.

[–] [email protected] 0 points 6 days ago

Cool! What's the effective difference for my life that your insistence on nuance has brought? What's the difference between a world where no one should have ai because the entirety of the tech is tainted with abuse and a world where no one should have ai because the entirety of the publicly available tech is tainted with abuse? What should I, a consumer, do? Don't say 1000 hrs of research on every fucking jpg, you know that's not the true answer just from a logistical standpoint