This blogpost makes an astoundingly good case about LLMs I hadn't considered before.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber Already the forums for VOIP software and embedded stuff (Arduino etc) are fully enshittified - they were toxic enough pre-AI) and folk have simply stopped contributing there (I think this happened just *before* LLMs became popular, so the quality of whatever does go to the training sets isn't going to be much good).
I suspect another factor is when people are getting paid for their work *and* depending on their employers upselling SaaS or other commercial services, they are less inclined to share stuff with the competition (I had to figure out my PJSIP trunk and securing it for myself, most of my findings are tooted here on Fedi as I'm not even sure where else to put them)
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber@social.coop exactly. AI can only exist because of human-made content from the Internet, and because humans have been talking with each other over the internet for roughly 30 years at this point, and most of these conversations have been archived somewhere.
So what would happen if this archive disappears? Maybe not the physical medium of the Internet, but the fact of humans speaking with each other over the Internet and recording their conversations, that can disappear. Less and less often people are asking questions of other humans in forums like Reddit or Stack Overflow, and more and more often humans ask questions to AI apps where the conversation cannot be used as more training data as this would create a destructive feedback loop in the signal that the AI is built to predict, which can lead to the AI behaving erratically.
(I originally wrote the above on my blog post from last August, I'm saying this to not plagiarize myself.)
It is kind of like the Internet is a forest of natural content. AI companies have now cut down the forest without planting new trees. They have all the raw material now and are trying (and failing) to profit from it, but in 5 years that information may become out-of-date, will any humans be asking new questions of each other on platforms not mined for content by AI companies? If not, these LLMs will rapidly become more and more innacurate over time.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber Thanks for sharing. To my ears, it rhymes with this response to the Microsoft-copilot-github encroachment: https://githubcopilotinvestigation.com/
-
@cwebber Thanks for sharing. To my ears, it rhymes with this response to the Microsoft-copilot-github encroachment: https://githubcopilotinvestigation.com/
@copystar
I've been thinking about adverse methods for reclaiming those debugging loops into offline models.
I think the process of "distilling" models can probably work as a sort of shunt to bring that back to the communities.It's not cheap, but it's only like $10k. Not the billions they're spending on these models quarterly.
Today you can pay AWS and they'll leave you alone in a room with Claude all day, no questions asked.
That might change in the future.
@cwebber -
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber Sure didn't help that Stack Overflow was killing itself too.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber dear god, i knew SO was taking a beating, but that chart the post linked to was _bleak_.
-
@cwebber Sure didn't help that Stack Overflow was killing itself too.
@mayintoronto @cwebber
True, that decline was on it's way. But at least the information was publicly available.
This is also why I was ranting about it two years ago when people were encouraging others to delete their stackoverflow contributions (as if that meant that the AI companies wouldn't have access to it, SO made a deal for direct share vs scraping). You're just denying it to other human beings at this point. Build a better web, post in personal blogs, community forums, don't lock up knowledge in opaque systems! -
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber Extra painful: bigger tech companies can afford to pay for plans that limit use of context data to train future versions of an LLM service's models so THEIR work is "protected" while their employees consume the commons. But smaller companies and individual users will be giving up their data.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber enclosure strikes again
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber this article have some good prediction on the knowledge silos LLM might be able to create, but it does not address the fact that the business model is not profitable. When the price hike hit to make LLM generating profits, people will have to balance the price of the subscription with the real value provided... just like any purchase. We might loose a few years of knowledge down the LLM silos before the collapse but personally, I think it is ok, we have plenty of good documented 1/2
-
@cwebber this article have some good prediction on the knowledge silos LLM might be able to create, but it does not address the fact that the business model is not profitable. When the price hike hit to make LLM generating profits, people will have to balance the price of the subscription with the real value provided... just like any purchase. We might loose a few years of knowledge down the LLM silos before the collapse but personally, I think it is ok, we have plenty of good documented 1/2
@cwebber tech from 2010 to 2023 that will still be fully usable. People choosing to use said tech will decrease cost and might fast-forward the downfall of LLM everywhere trend we are currently in.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber I keep repeating this in such contexts, apologies if I sound like a broken record: AI is a fascist project.
The purpose isn't merely enclosure of the commons. Making public stuff private is more of a means to an end.
There is centuries of historic precedent that shows that when a state has natural resources, it needs fewer people to extract wealth from that, and so pay for what keeps rulers in power.
If a state has fewer resources, it has to rely on a large, healthy population's...
-
@cwebber I keep repeating this in such contexts, apologies if I sound like a broken record: AI is a fascist project.
The purpose isn't merely enclosure of the commons. Making public stuff private is more of a means to an end.
There is centuries of historic precedent that shows that when a state has natural resources, it needs fewer people to extract wealth from that, and so pay for what keeps rulers in power.
If a state has fewer resources, it has to rely on a large, healthy population's...
@cwebber ... labour. A population that needs to be educated and mobile enough to fulfil their task. Such a population tends to demand more say in the affairs of state.
So natural resources lead to tyrannies, and lack thereof to democracies.
Privatisation of knowledge is a way of creating artificial resources to extract with fewer labourers. Plus, the more that extraction is automated, the smaller the population a ruler needs - or the more precarious their existence.
That is the goal.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber
One thing I've noticed in the same aspect as this, is that I talk much less code with colleagues now and have much less interactions with them for working through problems, and thus limit my exposure to alternate problem solving.When ever I want to discuss a problem, it's more often than not boiled down to some LLM answer, meaning, I might as well 'cut out' the middle and ask the LLM itself if all I get are LLM answers anyway.
That truly sucks.
"Have you asked Claude/Copilot/Chatgpt"..... -
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber of course guys - it was never about the LLM, it was about crowd-sourcing intelligence at an epic-scale. Every piece of code a developer writes and fixes becomes training data. Same with every conversation. I'm surprised people don't see the danger in having one single overlord and gatekeeper of all information in the world. Its crazy.
People seem to have forgotten what the real meaning of democracy and multi-lateralism are.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber this isn’t really new. For all the things on StackOverflow there was a huge domain of knowledge that was just not on there.
For most of my corporate developer life the knowledge/bug fixes would not be found on public forums but in internal collective knowledge, documentation or simply knowing a person in the same field. Most of this was not public domain.
The biggest issue now is that those firms commit their group knowledge to LLMs and we will not get it back.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber you calling it an 'astoundingly good case' makes me feel insightful in a way no LLM has been able to accomplish. I'm going to be insufferably smug for the rest of the day :)
-
@cwebber Already the forums for VOIP software and embedded stuff (Arduino etc) are fully enshittified - they were toxic enough pre-AI) and folk have simply stopped contributing there (I think this happened just *before* LLMs became popular, so the quality of whatever does go to the training sets isn't going to be much good).
I suspect another factor is when people are getting paid for their work *and* depending on their employers upselling SaaS or other commercial services, they are less inclined to share stuff with the competition (I had to figure out my PJSIP trunk and securing it for myself, most of my findings are tooted here on Fedi as I'm not even sure where else to put them)
@vfrmedia make a small static site blog to share your experience? Put so ethibg like Anubis to stop/slow scraping
-
@cwebber ... labour. A population that needs to be educated and mobile enough to fulfil their task. Such a population tends to demand more say in the affairs of state.
So natural resources lead to tyrannies, and lack thereof to democracies.
Privatisation of knowledge is a way of creating artificial resources to extract with fewer labourers. Plus, the more that extraction is automated, the smaller the population a ruler needs - or the more precarious their existence.
That is the goal.
-
@futzle @cwebber
When newbies encounter toxicity for asking their question on a public forum, cannot really blame them for turning to a LLM.
https://youtu.be/N7v0yvdkIHg