This blogpost makes an astoundingly good case about LLMs I hadn't considered before.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber Already the forums for VOIP software and embedded stuff (Arduino etc) are fully enshittified - they were toxic enough pre-AI) and folk have simply stopped contributing there (I think this happened just *before* LLMs became popular, so the quality of whatever does go to the training sets isn't going to be much good).
I suspect another factor is when people are getting paid for their work *and* depending on their employers upselling SaaS or other commercial services, they are less inclined to share stuff with the competition (I had to figure out my PJSIP trunk and securing it for myself, most of my findings are tooted here on Fedi as I'm not even sure where else to put them)
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber@social.coop exactly. AI can only exist because of human-made content from the Internet, and because humans have been talking with each other over the internet for roughly 30 years at this point, and most of these conversations have been archived somewhere.
So what would happen if this archive disappears? Maybe not the physical medium of the Internet, but the fact of humans speaking with each other over the Internet and recording their conversations, that can disappear. Less and less often people are asking questions of other humans in forums like Reddit or Stack Overflow, and more and more often humans ask questions to AI apps where the conversation cannot be used as more training data as this would create a destructive feedback loop in the signal that the AI is built to predict, which can lead to the AI behaving erratically.
(I originally wrote the above on my blog post from last August, I'm saying this to not plagiarize myself.)
It is kind of like the Internet is a forest of natural content. AI companies have now cut down the forest without planting new trees. They have all the raw material now and are trying (and failing) to profit from it, but in 5 years that information may become out-of-date, will any humans be asking new questions of each other on platforms not mined for content by AI companies? If not, these LLMs will rapidly become more and more innacurate over time.
-
This blogpost makes an astoundingly good case about LLMs I hadn't considered before. The collapse of public forums (like Stack Overflow) for programming answers coincides directly with the rise of programmers asking for answers from chatbots *directly*. Those debugging sessions become part of a training set that now *only private LLM corporations have access to*. This is something that "open models" seemingly can't easily fight. https://michiel.buddingh.eu/enclosure-feedback-loop
@cwebber Thanks for sharing. To my ears, it rhymes with this response to the Microsoft-copilot-github encroachment: https://githubcopilotinvestigation.com/