Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture".
-
This is a purity culture argument about the "fruit of the poisoned tree." The silicon in your laptop was invented by a eugenicist. The network your packets transit was invented by war criminals. The satellite the signal travels on was launched on a rocket descended from Nazi designs that were built by death-camp slaves.
@pluralistic i guess this misses the point: the particular chip in my laptop wasn't made by war criminals (i hope...), but the model you do use was trained under vast amounts of energy and water consumption. I'm not sure this is completely comparable, tbh.
-
Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.
EDIT: Diskussions under this are fine, but I do not want this to turn into an ad hominem attack to Cory. Be fucking respectful
https://tante.cc/2026/02/20/acting-ethical-in-an-imperfect-world/
@tante I am pursuing what I am calling my "AI Tea Party". It's origin was in putting family birth certificates in a self-hosted tool and realizing I didn't know if that meant they would end up in a training set somewhere. That began a process of me purging direct links to any LLM. Next step is switching to tools that do not use any LLM in their operation. After that is switching from suppliers of anything that use LLMs to operate.
This is a Sisyphean task since these round to **fucking everything** but I'm motivated to pursue it anyway.
-
@shiri An used mink coat may not give money directly to mink farmers/killers, but wearing mink fur sends a message about the acceptability of mink. The average passerby can't tell if the mink was bought new. If you walk down the street and there are 10 new mink wearers, the 11th "ethical" mink wearer lends themselves to the message that mink farming is fine, unless they are constantly screaming "this is used mink!" which is strange and obnoxious.
@skyfaller that is a better argument and I'll definitely accept that.
I think for many of us, myself included, the big thing with AI there is the investment bubble. Users aren't making that much difference on the bubble, the people propping up the bubble are the same people creating the problems.
I know I harp on people about anti-AI rage myself, but I specifically harp on people who are overbroad in that rage. So many people dismiss that there are valid use cases for AI in the first place, they demonize people who are using it to improve their lives... people who can be encouraged now to move on to more ethical platforms, and when the bubble bursts will move anyways.
We honestly don't need public pressure to end the biggest abuses of AI, because it's not public interest that's fueling them... it's investor's believing AI techbros. Eventually they're going to wise up and realize there's literally zero return on their investment and we're going to have a truly terrifying economic crash.
It's a lot like the dot-com bubble... but drastically worse.
-
@tante @pluralistic @simonzerafa I agree in principle with Cory, but I really wish that he had clarified that:
1. Ollama is not an LLM, it's a server for various models, of varying degrees of openness.
2. Open weights is not open source, the model is still a black box. We should support projects like OLMO, which are completely open, down to the training data set and checkpoints.
3. It's quite difficult to "seize that technology" without using Someone Else's Computer to do so (a.k.a clown/cloud)@tante @pluralistic @simonzerafa But ALSO: using a multi-billion-parameter synthetic text extruding machine to find spelling and syntax errors is a blatant example of "doing everything the least efficient way possible" and that's why we are living on an overheating planet buried under toxic e-waste.
If I think about it harder I could probably come up with a more clever metaphor than killing a mosquito with a flamethrower, but you get the idea.
-
@FediThing … Certainly this is true of my reasoning for a #noAI stance. For me it's about climate, economic and social impacts of the ever growing mega-LLMs, and the craze to use them for all kinds of purposes for which they are unfit.
I am much less concerned with a local instance checking a writer's grammar. Lumping those two together makes little sense, to me.
On some other topics, I find @pluralistic's leadership constructive and helpful.
-
@Colman @FediThing @tante That's interesting. I've never wondered that about you.
@pluralistic @Colman @FediThing @tante wow. punching down? I had a higher opinion of you.
-
@pluralistic i guess this misses the point: the particular chip in my laptop wasn't made by war criminals (i hope...), but the model you do use was trained under vast amounts of energy and water consumption. I'm not sure this is completely comparable, tbh.
@pluralistic and yes, i'm aware that producing a chip also costs vast amounts of energy and water... but at least my chip is used to solve a multitude of purposes, while a LLM that checks spelling and grammar is built and trained for one single use-case (that, nb, could also be done without an LLM). So yes, I do differenciate. @FediThing @tante
-
@pluralistic and yes, i'm aware that producing a chip also costs vast amounts of energy and water... but at least my chip is used to solve a multitude of purposes, while a LLM that checks spelling and grammar is built and trained for one single use-case (that, nb, could also be done without an LLM). So yes, I do differenciate. @FediThing @tante
Llama 2 was not built to check spelling and grammar. That's "not even wrong."
-
@pluralistic i guess this misses the point: the particular chip in my laptop wasn't made by war criminals (i hope...), but the model you do use was trained under vast amounts of energy and water consumption. I'm not sure this is completely comparable, tbh.
No, this is just more "fruit of the poisoned tree" and your argument that your fruit of the poisoned tree doesn't count is the normal special pleading that this argument always decays into.
-
@tante @pluralistic @simonzerafa But ALSO: using a multi-billion-parameter synthetic text extruding machine to find spelling and syntax errors is a blatant example of "doing everything the least efficient way possible" and that's why we are living on an overheating planet buried under toxic e-waste.
If I think about it harder I could probably come up with a more clever metaphor than killing a mosquito with a flamethrower, but you get the idea.
No. It's like killing a mosquito with a bug zapper whose history includes thousands of years of metallurgy, hundreds of years of electrical engineering, and decades of plastics manufacture.
There is literally no contemporary manufactured good that doesn't sit atop a vast mountain of extraneous (to that purpose) labor, energy expenditure and capital.
-
@pluralistic Ok, fair enough, if spell checking is literally the only thing you use LLMs for.
I still think you wouldn't rely on a 1950s dictionary for checking modern language, and language moves faster on the internet, but I'm willing to concede that point.
I still think a deterministic spell checker could have done the job and not put you in this weird position of defending a technology with wide-reaching negative effects. But I guess your post was for just that purpose.
I'm not using it for spell checking.
Did you read the article that is under discussion?
-
@FediThing @tante Thank you.
-
Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.
EDIT: Diskussions under this are fine, but I do not want this to turn into an ad hominem attack to Cory. Be fucking respectful
https://tante.cc/2026/02/20/acting-ethical-in-an-imperfect-world/
@tante spot on.
-
@tante thank you.
@tante i think the strawman indeed IS the issue comparing (even it was just through context) an LLM for spell checking/grammar where it is really insignificant if IT performs well or not to a general usability, referring to liberation including critical tasks.
I don't detest AI because of the fascists that created most of IT but because they intentionally design and sell "tools" that are good at fascism and not much else of significance. A screwdriver with a grip that cuts the user.
-
@tante i think the strawman indeed IS the issue comparing (even it was just through context) an LLM for spell checking/grammar where it is really insignificant if IT performs well or not to a general usability, referring to liberation including critical tasks.
I don't detest AI because of the fascists that created most of IT but because they intentionally design and sell "tools" that are good at fascism and not much else of significance. A screwdriver with a grip that cuts the user.
@tante a screwdriver that only works on a low percentage of screws it was designed for, thus "Tools".
-
I'm not using it for spell checking.
Did you read the article that is under discussion?
@pluralistic I apologize, I did in fact read the relevant section of your post, and I was using spell-checking as shorthand for all typo checking, because deterministic grammar checkers have also existed for some time, although not as long as spell checkers and perhaps they have not been as reliable. I understand that LLMs can catch some typos that deterministic solutions may not.
I just think we should put more effort into improving deterministic tools instead of giving up.
-
@pluralistic I apologize, I did in fact read the relevant section of your post, and I was using spell-checking as shorthand for all typo checking, because deterministic grammar checkers have also existed for some time, although not as long as spell checkers and perhaps they have not been as reliable. I understand that LLMs can catch some typos that deterministic solutions may not.
I just think we should put more effort into improving deterministic tools instead of giving up.
-
@tante It seems to me Doctorow is obviously correct about this. But I don't think it matters too much if you don't agree... the trajectory of LLMs is going to be whatever it is going to be.
If you don't like it and have buddies that don't like it either, that's not a bad thing especially if you are undergoing real negative effects from it.
It's just if you stray from reality (whatever that will be) too far for too long, you will end up with a big shock when forced to rejoin it.
Don't mistake a hugely popular fad or bubble for "reality." And if you don't believe that "[nearly] everybody believes" can be quite detached from punishingly harsh reality, then you need to read about the "Tulip Mania" craze and bubble:
-
Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.
EDIT: Diskussions under this are fine, but I do not want this to turn into an ad hominem attack to Cory. Be fucking respectful
https://tante.cc/2026/02/20/acting-ethical-in-an-imperfect-world/
I completely agree with your view on us being messy, imperfect beings. And while many take such a realization as a free ticket to shrug themselves into deep cynicism, I deeply appreciate people who tend to try a little harder than most to do the right thing, and own every compromise they decide to make as what it is.
Once we start warping our analysis and critical thinking to match our actions instead of trying our best to make our actions fit the former, we'll quickly start losing any ability to act with accountability. -
Don't mistake a hugely popular fad or bubble for "reality." And if you don't believe that "[nearly] everybody believes" can be quite detached from punishingly harsh reality, then you need to read about the "Tulip Mania" craze and bubble:
And likewise, don't mistake "mainstream thinking" or what "most of the industry is doing" with "reality" or even "best practice." Agile, Lean, and Total Quality Management, and practically about every other significant improvement is a break from "the usual way of doing things." Improvement is a change from the mediocre.
"Appeal to Popularity" (as a signal of truth) is literally a well documented Logical Fallacy: