A woman sues her insurance company for terminating her disability benefits.
-
@divVerent @jonoleth If you're aware of any specific Illinois caselaw that's on point here, I'd be interested to hear about it. But if you're just a nonlawyer making stuff up about what you imagine the law to be, please leave me out of the discussion.
@mjd @jonoleth I am not even American. If in your country machines and companies are "persons" and have human rights that have priority over the human rights of _humans_, then your whole country is wrong. What's next, voting rights in federal elections for corporations? Second Amendment for AIs?
But yeah, that might indeed be the case.
In my country it is "societas delinquere non potest". A company _cannot_ be defendant of a crime - only the people actually performing the actions can.
But yeah, done here. Let's see what broken new case law will come from Trumpistan.
-
@divVerent @mjd
OpenAI is interfering in the insurance company's contract? Or, at least, that's what they're suing for -
A woman sues her insurance company for terminating her disability benefits. They reach a settlement and agree that the suit will be dismissed with prejudice.
She decides she doesn't like the settlement and asks her lawyers to reopen the case.They say they can't: it was dismissed, and in the settlement she agreed not to reopen the case.
She asks ChatGPT if her attorneys are lying to her. It says they are. She fires them and continues pro se, advised by ChatGPT.
CharGPT generates legal arguments for reopening the case, which she files, and 21 more motions, a subpoena, and eight other notices and statements, which she files.
The court denies her motion to reopen the case.
Advised by ChatGPT, she files a new suit against the insurance company and submits 44 more motions, memoranda, etc., which include citations to nonexistent cases.
Now the insurance company has sued OpenAI for tortious interference with their settlement contract.
đż
https://storage.courtlistener.com/recap/gov.uscourts.ilnd.496515/gov.uscourts.ilnd.496515.1.0_1.pdf
@mjd
> They reach a settlement and agree that the suit will be dismissed with prejudice.It's done, put a fork in it.
-
@mjd TBH I do not think OpenAI should be responsible. They're just providing a fancy random text generator to the public. And it's outright impossible to teach a random text generator to _not_ output a specific kind of text, as whatever you do, there is a way around it.
The woman should pay all costs, as per the usual "vexatious filings" or "frivolous lawsuits" standards.
Plus, the law in her state against practicing law without a license starts with "No person shall...". ChatGPT isn't a person.
@divVerent It's funny how one can use "just" to *just*ify anything. Let's reduce "they hyped their product as ÂŤthe artificial superintelligence magical clever toolÂť, but didn't even bother adding safety guardrails or disclaimers about output not being legal advice, and now they're getting reamed" as "they're _just_ providing a fancy random text generator to the public."
Or let's not.
-
@diazona I don't think it is a shame that this happened to this woman. It appears that she is a very ordinary type of vexatious litigant, except that she is also being aided by ChatGPT.
@mjd @diazona Though I'm not a lawyer (thank Cthulhu, or belly rubs to it's acolyte Menhit @antipope_cats) I do recall the Scottish courts taking exception to a "vexatious litigant" a while ago. It ended badly for said litigant.
-
@mjd @krupo that MIT article demonstrated 5% of AI implementations are profit making. 95% are loss making. so when the investment goldrush mania ends or winds down, 95% of invested amounts will be wiped out. it's billions so it will have a disruptive and negative economic effect that I think we will likely experience as recession. and then that 5% of profit making implementations will be what carries forward, with further investment being modelled on those (anyone can do that right now).
@falcennial @mjd @krupo
Looking to the foreseeable future, the AI boosters (especially those using AI to write their "opinions") will blame the bubble-burst on the Trump-Bibi War (no, they can't escape the blame) and *it's* crash.
How much are Darien Scheme share certificates worth these days? -
@mjd â41. On October 29, 2025, OPENAI amended the terms and usage policies of ChatGPT to prohibit users from using ChatGPT to provide tailored legal advice. Prior to the October 29, 2025 emendation, ChatGPTâs terms of use did not prohibit users from using ChatGPT to draft legal papers, conduct legal research, provide legal analysis or give legal advice.â
-
@mjd TBH I do not think OpenAI should be responsible. They're just providing a fancy random text generator to the public. And it's outright impossible to teach a random text generator to _not_ output a specific kind of text, as whatever you do, there is a way around it.
The woman should pay all costs, as per the usual "vexatious filings" or "frivolous lawsuits" standards.
Plus, the law in her state against practicing law without a license starts with "No person shall...". ChatGPT isn't a person.
@divVerent @mjd "barratry" - that's a term I remember from the Scottish Legal High Heidjuns dealing with a similar "vexatious litigant". I remember looking it up (but not the precise definition).
-
@divVerent @mjd ChatGPT is not a person, which is why ChatGPT is not being sued. OpenAI sells a tool that gave her legal advice, and they certainly didn't say anywhere that it's actually just a "fancy random text generator"
Wait, what?
They *sell* this shit? And charge money for it?Where the holy cat turds do they find clients? On the Internet?
(No, I've never tried to use an AI.)
-
@divVerent It's funny how one can use "just" to *just*ify anything. Let's reduce "they hyped their product as ÂŤthe artificial superintelligence magical clever toolÂť, but didn't even bother adding safety guardrails or disclaimers about output not being legal advice, and now they're getting reamed" as "they're _just_ providing a fancy random text generator to the public."
Or let's not.
@adriano @mjd There already is a disclaimer:
https://web.archive.org/web/20240722050439/https://openai.com/policies/row-terms-of-use/
> What you cannot do. You may not use our Services for any illegal, harmful, or abusive activity. For example, you may not:
[...]
> Represent that Output was human-generated when it was not.
[...]
> Output may not always be accurate. You should not rely on Output from our Services as a sole source of truth or factual information, or as a substitute for professional advice.
[...]
> You must not use any Output relating to a person for any purpose that could have a legal or material impact on that person, such as making credit, educational, employment, housing, insurance, legal, medical, or other important decisions about them.Pretty much looks like clearly disclaiming to me.
-
A woman sues her insurance company for terminating her disability benefits. They reach a settlement and agree that the suit will be dismissed with prejudice.
She decides she doesn't like the settlement and asks her lawyers to reopen the case.They say they can't: it was dismissed, and in the settlement she agreed not to reopen the case.
She asks ChatGPT if her attorneys are lying to her. It says they are. She fires them and continues pro se, advised by ChatGPT.
CharGPT generates legal arguments for reopening the case, which she files, and 21 more motions, a subpoena, and eight other notices and statements, which she files.
The court denies her motion to reopen the case.
Advised by ChatGPT, she files a new suit against the insurance company and submits 44 more motions, memoranda, etc., which include citations to nonexistent cases.
Now the insurance company has sued OpenAI for tortious interference with their settlement contract.
đż
https://storage.courtlistener.com/recap/gov.uscourts.ilnd.496515/gov.uscourts.ilnd.496515.1.0_1.pdf
HA!!!!!!!
-
@mjd TBH I do not think OpenAI should be responsible. They're just providing a fancy random text generator to the public. And it's outright impossible to teach a random text generator to _not_ output a specific kind of text, as whatever you do, there is a way around it.
The woman should pay all costs, as per the usual "vexatious filings" or "frivolous lawsuits" standards.
Plus, the law in her state against practicing law without a license starts with "No person shall...". ChatGPT isn't a person.
@divVerent @mjd This is simplistic to the point of being false. Long before we had LLMs, we had Clippy, which was smart enough to say âit looks like youâre writing a memo.â OpenAI and its counterparts can unquestionably add a âit looks like youâre seeking legal adviceâ detector to their products. They already, supposedly, try to detect whether their users are attempting self-harm. LLMs evolved from classification software, so this kind of thing is in their roots.
-
@mjd @jonoleth I am not even American. If in your country machines and companies are "persons" and have human rights that have priority over the human rights of _humans_, then your whole country is wrong. What's next, voting rights in federal elections for corporations? Second Amendment for AIs?
But yeah, that might indeed be the case.
In my country it is "societas delinquere non potest". A company _cannot_ be defendant of a crime - only the people actually performing the actions can.
But yeah, done here. Let's see what broken new case law will come from Trumpistan.
"Pretty sure it's common knowledge that LLMs are nothing but random text generators."
Among us? Yes. Among the rest of folks? No, it is not well known at all, most laypeople I talk to believed the hype at face value
-
@adriano @mjd There already is a disclaimer:
https://web.archive.org/web/20240722050439/https://openai.com/policies/row-terms-of-use/
> What you cannot do. You may not use our Services for any illegal, harmful, or abusive activity. For example, you may not:
[...]
> Represent that Output was human-generated when it was not.
[...]
> Output may not always be accurate. You should not rely on Output from our Services as a sole source of truth or factual information, or as a substitute for professional advice.
[...]
> You must not use any Output relating to a person for any purpose that could have a legal or material impact on that person, such as making credit, educational, employment, housing, insurance, legal, medical, or other important decisions about them.Pretty much looks like clearly disclaiming to me.
-
undefined oblomov@sociale.network shared this topic