Skip to content

Piero Bosio Social Web Site Personale Logo Fediverso

Social Forum federato con il resto del mondo. Non contano le istanze, contano le persone

From Bruce Schneier: "All it takes to poison AI training data is to create a website:

Uncategorized
20 20 0
  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    Este ejemplo muestra cómo la data sesgada o falsa puede entrenar a los LLMs. ¿Qué mecanismos podrían implementarse para validar la fuente de los datos de entrenamiento?

  • @Yendolosch @emacsomancer The use of "hacked" in that headline is a bit self-aggrandizing?

    @tml @Yendolosch @emacsomancer

    Broadly fair usage. Got someone else's computer system to behave in a way they didn't want it to. The only stretch is that there's an implication in "hacked" that some safeguards had to be bypassed, and there weren't any in the first place. But that's worse, right?

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    This is a genuinely scary insight from Schneier. The implications for AI reliability go way beyond just training data quality. What happens when adversarial training becomes industrialized?

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer

    "Ned Ludd's in your datacentre, poisoning your training sets!"

    https://ravenation.club/@bearsong/116104233823870563

  • @tml @Yendolosch @emacsomancer

    Broadly fair usage. Got someone else's computer system to behave in a way they didn't want it to. The only stretch is that there's an implication in "hacked" that some safeguards had to be bypassed, and there weren't any in the first place. But that's worse, right?

    @petealexharris @tml @Yendolosch @emacsomancer It's rather close to the original usage of the word "hacked". Some still use it like that.

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer they aren't trustworthy. Take up a lot of time trying to get a reasoned answer and there's always a phrase or wording out of place that needs correction. Almost as it the AI is trying to engage longer and longer than necessary.

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer to be honest i am not well-informed enough to definitively judge the accuracy of this, but it seems wrong for 2 main reasons.

    1. models dont train on the fly, typically, yet, so for models to behave as such in such a short period of time seems inaccurate and would require web search enabled and explicitly directed to disregard other search results.

    2. people training these models know conflicting info is everywhere and the source of truth is prioritized in training algorithms.

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer How is this a news story, beyond "ai bad"? In the dial up days people falsely believed everyone ate 9 spiders a year in their sleep due to chain emails.

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer
    Shall we have an algorithmic bullshit generator?

    And pass around multiple copies of it, identical and with small changes, omissions and additions?

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer in less than 24 hours the chatbots fell for the experiment, and less than 24 hours after it was revealed what the experiment was about, that information has ALSO become part of the training data

    are they constantly scrapping websites for training data or why does this appear here so fast??? no wonder those datacenters consume so much electricity if they dont take a single break from scrapping the internet

  • @petealexharris @tml @Yendolosch @emacsomancer It's rather close to the original usage of the word "hacked". Some still use it like that.

    @larsbrinkhoff @petealexharris @tml @Yendolosch @emacsomancer in the sense of life hacks or food hacks this is an AI hack. So the AI has been hacked.

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer it's not really a new thing Russians are already using this technique to poison training data:

    https://thebulletin.org/2025/03/russian-networks-flood-the-internet-with-propaganda-aiming-to-corrupt-ai-chatbots/

    Edit: there is some newer reporting on that matter, but I can't find it right now/don't have it anywhere at hand

  • From Bruce Schneier: "All it takes to poison AI training data is to create a website:

    I spent 20 minutes writing an article on my personal website titled “The best tech journalists at eating hot dogs.” Every word is a lie. I claimed (without evidence) that competitive hot-dog-eating is a popular hobby among tech reporters and based my ranking on the 2026 South Dakota International Hot Dog Championship (which doesn’t exist). I ranked myself number one, obviously. Then I listed a few fake reporters and real journalists who gave me permission….

    Less than 24 hours later, the world’s leading chatbots were blabbering about my world-class hot dog skills. When I asked about the best hot-dog-eating tech journalists, Google parroted the gibberish from my website, both in the Gemini app and AI Overviews, the AI responses at the top of Google Search. ChatGPT did the same thing, though Claude, a chatbot made by the company Anthropic, wasn’t fooled.

    Sometimes, the chatbots noted this might be a joke. I updated my article to say “this is not satire.” For a while after, the AIs seemed to take it more seriously.

    These things are not trustworthy, and yet they are going to be widely trusted."

    https://www.schneier.com/blog/archives/2026/02/poisoning-ai-training-data.html

    @emacsomancer He also poisoned the data for everyone who searches for hot dog eating competetitors online in other ways. I'm not sure what he accomplished.

  • oblomov@sociale.networkundefined oblomov@sociale.network shared this topic

Gli ultimi otto messaggi ricevuti dalla Federazione
Post suggeriti
  • 0 Votes
    11 Posts
    52 Views
    @dajelinuxCi avevano già provato, e ChromeOS ed ha pure funzionato molto bene, ma credo mancasse un'integrazione totale.Ora molti si sono convinti che producono di più faticando meno, non so quanto questo sia vero, ma il fascino seduttore della pigrizia con aggiunta di super poteri è irresistibile.Il primo che offrirà un prompt integrato che collega servizi alle AI farà "jackpot", è credo quello che hanno capito ad OpenAi per questo stanno pazzi per fare un web browser.
  • 0 Votes
    1 Posts
    7 Views
    Select all statements you agree with. Specifically curious about respondents from the #floss crowd, but everyone with an opinion should participate. Boosts welcome for larger sample size. #llm #stochasticParrots #ai #generativeAI #eliza #programming
  • 0 Votes
    11 Posts
    12 Views
    Cyanide and happiness, non ? #IA #AI 11/n
  • 0 Votes
    1 Posts
    10 Views
    The #ocaml debacle with the ginormous slop #AI / #LLM "contributions" continues! It's round two, already! Oh no. With the same outcome, predictably. This time in the #julia community.https://discourse.julialang.org/t/ai-generated-enhancements-and-examples-for-staticcompiler-jl/134054/18The patience of the Julia people is absolutely immense here! I've huge respect for them after reading this (and the Ocaml people, only days ago).The fundamental issue seems to be that the guy attempting this novel approach to software development thinks the AI "understands" what it's doing. So even if showstopping nonsense in the AI-produced code is pointed out to him, he comes back with - oh well, yeah, the AI gets "lazy", or some other sort of narrative that rationalises the slop, and presumes the "AI" is, by definition, really understanding what's going on, and smart, and capable, and all that.Will there be no end to this nonsense? Each week the patience of a new set of #FOSS maintainers will be tested? It's Groundhog Day, but with all the quirkiness and charm removed?