Lawyer Cited 6 Fake Cases Made Up By ChatGPT; Judge Calls It 'Unprecedented' (arstechnica.com) 48
An anonymous reader quotes a report from Ars Technica: A lawyer is in trouble after admitting he used ChatGPT to help write court filings that cited six nonexistent cases invented by the artificial intelligence tool. Lawyer Steven Schwartz of the firm Levidow, Levidow, & Oberman "greatly regrets having utilized generative artificial intelligence to supplement the legal research performed herein and will never do so in the future without absolute verification of its authenticity," Schwartz wrote in an affidavit (PDF) on May 24 regarding the bogus citations previously submitted in US District Court for the Southern District of New York.
Schwartz wrote that "the use of generative artificial intelligence has evolved within law rms" and that he "consulted the artificial intelligence website ChatGPT in order to supplement the legal research performed." The "citations and opinions in question were provided by ChatGPT which also provided its legal source and assured the reliability of its content," he wrote. Schwartz admitted that he "relied on the legal opinions provided to him by a source that has revealed itself to be unreliable," and stated that it is his fault for not confirming the sources provided by ChatGPT. Schwartz didn't previously consider the possibility that an artificial intelligence tool like ChatGPT could provide false information, even though AI chatbot mistakes have been extensively reported by non-artificial intelligence such as the human journalists employed by reputable news organizations. The lawyer's affidavit said he had "never utilized ChatGPT as a source for conducting legal research prior to this occurrence and therefore was unaware of the possibility that its content could be false."
Federal Judge Kevin Castel is considering punishments for Schwartz and his associates. In an order on Friday, Castel scheduled a June 8 hearing at which Schwartz, fellow attorney Peter LoDuca, and the law firm must show cause for why they should not be sanctioned. "The Court is presented with an unprecedented circumstance," Castel wrote in a previous order on May 4. "A submission filed by plaintiff's counsel in opposition to a motion to dismiss is replete with citations to non-existent cases... Six of the submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations." [...] In the order issued on Friday last week, Castel said that Schwartz may be sanctioned for "the citation of non-existent cases to the Court," "the submission to the Court of copies of non-existent judicial opinions," and "the use of a false and fraudulent notarization." Schwartz may also be referred to an attorney grievance committee for additional punishment. Castel wrote that LoDuca may be sanctioned "for the use of a false and fraudulent notarization in his affidavit filed on April 25, 2023." The law firm could be sanctioned for "the citation of non-existent cases to the Court," "the submission to the Court of copies of non-existent judicial opinions annexed to the Affidavit filed on April 25, 2023," and "the use of a false and fraudulent notarization in the affidavit filed on April 25, 2023."
Schwartz wrote that "the use of generative artificial intelligence has evolved within law rms" and that he "consulted the artificial intelligence website ChatGPT in order to supplement the legal research performed." The "citations and opinions in question were provided by ChatGPT which also provided its legal source and assured the reliability of its content," he wrote. Schwartz admitted that he "relied on the legal opinions provided to him by a source that has revealed itself to be unreliable," and stated that it is his fault for not confirming the sources provided by ChatGPT. Schwartz didn't previously consider the possibility that an artificial intelligence tool like ChatGPT could provide false information, even though AI chatbot mistakes have been extensively reported by non-artificial intelligence such as the human journalists employed by reputable news organizations. The lawyer's affidavit said he had "never utilized ChatGPT as a source for conducting legal research prior to this occurrence and therefore was unaware of the possibility that its content could be false."
Federal Judge Kevin Castel is considering punishments for Schwartz and his associates. In an order on Friday, Castel scheduled a June 8 hearing at which Schwartz, fellow attorney Peter LoDuca, and the law firm must show cause for why they should not be sanctioned. "The Court is presented with an unprecedented circumstance," Castel wrote in a previous order on May 4. "A submission filed by plaintiff's counsel in opposition to a motion to dismiss is replete with citations to non-existent cases... Six of the submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations." [...] In the order issued on Friday last week, Castel said that Schwartz may be sanctioned for "the citation of non-existent cases to the Court," "the submission to the Court of copies of non-existent judicial opinions," and "the use of a false and fraudulent notarization." Schwartz may also be referred to an attorney grievance committee for additional punishment. Castel wrote that LoDuca may be sanctioned "for the use of a false and fraudulent notarization in his affidavit filed on April 25, 2023." The law firm could be sanctioned for "the citation of non-existent cases to the Court," "the submission to the Court of copies of non-existent judicial opinions annexed to the Affidavit filed on April 25, 2023," and "the use of a false and fraudulent notarization in the affidavit filed on April 25, 2023."
At least it's not on the front page anymore (Score:3)
But this story is OLD now.
Re: (Score:2)
At the very least (Score:3)
The guy should be banned from practising law as he's proven he's criminally negligent in his duty to properly represent clients.
Re: (Score:2)
Comment removed (Score:5, Insightful)
Re: (Score:3)
Some tech people here understand the limitations quite well. A lot of others pray to yet another surrogate God found in technology.
The thing about replacing jobs is that a lot of low-level white-collar jobs can be done by a well-spoken moron with some minimal factual knowledge. And that seems to be quite within the range of a slightly augmented "Chat AI". And since these are low-level white-collar jobs, i.e. not producing anything, there will not be any new jobs by additional demand. Any job that requires a
Re: (Score:2)
There's a dupe in your comment :)
Re: (Score:2)
AI obviously isn't going to do all the programming, but it sure is going to change the nature of programming as a job.
Re:At the very least (Score:5, Interesting)
I'm a little less worried about whether AI can do all the programming so much as I'm worried that business folk will *think* it can. That means layoffs for my colleagues or myself, a huge burden on the remainder, and dealing with management "fixing" it by hiring "prompt engineers" and insisting those prompt engineers contribute an obvious amount to the projects.
Business leaders don't have the best track record of making the best decisions in the face of massive amounts of marketing hype telling them the "right" way to go.
Re: (Score:2)
That's why I exercise my prompts on various AI solutions.
If management decides to fire me, they will come back begging within a couple weeks tops, opening up the opportunity for me to charge them external consultant fees.
Re: (Score:2)
Re: At the very least (Score:2)
Where I work an executive decision was made to ban most AI tools because they represent a data exfiltration security risk. We only allow certain ones and strictly govern their use, and we only allow them to be used at all because they help people be more productive. Nobody, at any level, is under any illusions of them replacing any jobs.
Re: (Score:2)
It's enough to look at machine-based language translation solutions.
How many years since they popped up? A decade? More? They still can't translate properly from one language to another.
On the other hand, many companies switched from human translation to automated translation for their websites, with hilarious (and occasionally dangerous) results. So... jobs were lost, or, at least, not gained.
I, for one, started to branch out years ago. And I mean really branch out. Examples include but are not limited to
Re: (Score:2)
Altavista's Babelfish launched about 25 years ago, so quite a bit more than 10 years ago. Google Translate launched 17 years ago.
Re: (Score:2)
"AI will do all the programming" - Slashdotters panic, "They're taking our jerbs!"
"AI wrote bogus thing that someone relied on who wasn't technically skilled and didn't know the limitations" - Slashdotters point and laugh at the "idiot" who doesn't understand "AI" isn't capable of doing what they claim.
If tech people don't understand the limitations, how are non-tech people supposed to understand it? And why are you all calling this person an idiot when you're panicking every time the idea your own job will be replaced comes up?
Both can be true. AI will replace programmers AND a particular instance of AI spews nonsense with regularity. The thing to be mindful of is trendline of technical capability and not get caught up in what today's systems can and can't do.
Nobody knows what will actually happen with this technology. It was never expected in advance something with capabilities of GPT-4 could arise from such a small ANN and most tech d00ds are not at all confused over the question of fallibility of present day chat bots.
Re: At the very least (Score:2)
Is it criminally negligent? I think dis-barring him is a bit too much. I suspect that, like many people, he thought using AI would help with this sort of thing, and I can't blame him for that. The critical mistake was not double-checking the result. Given how this has made the news, I think the bad PR is punishment enough. He won't be making that mistake again and this served as a warning to others that (currently) ChatGPT is not helpful for finding case precedence.
ChatGPT turns out to be especially unhel
Re: (Score:2)
>I think the bad PR is punishment enough.
So that people who aren't aware of his past can end up with a lazy lawyer who takes shortcuts and does sub-standard work?
This is what disbarment should be for - so that potential customers know they're getting someone who has passed the minimum standard for the industry because if they don't they lose their license.
Re: (Score:2)
I see what you're saying, but the question is whether this is sufficiently bad that the guy should lose his career over it, because that's what disbarment would mean. Should he go flip burgers because he made the mistake of trusting ChatGPT for legal precedence references? To me, at least, this doesn't rise to that level. But then again, I'm not a legal scholar or able to disbar anyone, so what the hell do I know?
If this was the tenth time he's done this sort of thing - if they went back and found a hist
Re: (Score:2)
Editors rely on chatbot to generate articles (Score:5, Funny)
Re: (Score:3)
Re: (Score:3)
Don't be silly. ChatGPT knows how to write proper headlines and not duplicate articles. To screw up as badly as Slashdot Editors takes a human touch.
The AI dimension to this isn't special. (Score:2)
On the other hand, the creation of an entire "ecosystem" of fraudulent jurisprudence has the smell of totalitarian history-revision, so the response should be harsh and proactive.
Re:The AI dimension to this isn't special. (Score:4, Insightful)
Usually you have paralegals doing the fact-checking and the finding of the rulings. They are a lot cheaper than lawyers, but they still need to be paid. This guy went cheap without understanding what he was doing. An obvious case of legal malpractice.
It's incredibly rare (Score:1)
I don't imagine it's rare to catch lawyers being negligent with their sources.
Actually, t's incredibly rare.
Because the core of just about any legal argument is, look at what these other cases decided in similar circumstances.
And, you know that whatever judge is looking at this is going to look at every case (well really his clerks or other legal staff) you cite VERY closely, which is easy to do as every lawyer on Earth has access to Westlaw.
You are simply never, ever going to get away with citing a fake ca
Re: (Score:2)
Pathological Liar (Score:2)
Re: Pathological Liar (Score:2)
Re: (Score:2)
Re: Pathological Liar (Score:3)
Re: (Score:2)
Re: (Score:2)
Harshest punishment (Score:3)
Unprecedented (Score:4, Funny)
I know judges like to issue cute rulings, but they could spare us the dad jokes.
They punish themselves (Score:5, Insightful)
What what the normal human pentalty be? (Score:4, Interesting)
Re: What what the normal human pentalty be? (Score:2)
Re: (Score:2)
The same. Lawyer disbarred. Plus the law firm will fire the paralegal. The lawyer has the responsibility.
So, in this case, the Lawyer will be disbarred (same). Plus the law firm will fire ChatGPT :-)
It shouldn't have hired ChatGPT in the first place, though!
a core problem (Score:2)
He should have researched it (Score:3)
Was the judge real or just another chatbot? (Score:2)