A New York lawyer has been ordered to explain why he should not be disciplined after using ChatGPT for legal research.
The original case in question involved a man suing an airline over an alleged personal injury. His legal team, made up of lawyers from the firm Levidow, Levidow & Oberman submitted a court document on his behalf. The document contained a list of legal cases to prove, using precedent, why the case should be considered.
However upon closer inspection, the airline’s lawyers said they could not find a number of the cases mentioned in the brief.
“Six of the submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations,” Judge Castel wrote in an order demanding the man’s legal team explain itself.
More from News
- Bank Of England Holds Interest Rate At 3.75% – What Does This Mean For The UK Economy Today?
- Could The $80 Billion Metaverse Flop Be One Of The Four Horsemen Of The AI Bubble Apocalypse?
- Is The UK Government Finally Listening To Creatives?
- If The SEC Scraps Quarterly Reporting, Will IPOs In The US Become More Attractive Or More Risky?
- What Are Experts Expecting After EU Commission Presents The EU Inc. Legislative Proposal?
- UK Invests £2 Billion Into Quantum Tech: What Does This Mean For The Future Of AI Across Industries?
- How’s The Government Planning To Make The UK The Fastest AI Adopters Among G7 Nations?
- Top Anti-Drone Startups And Companies
Despite ChatGPT having a disclaimer that it can produce inaccurate information, the lawyer who used the tool told the court he was “unaware that its content could be false”.
Mr Schwartz, a lawyer at the firm who prepared the document, said that he “greatly regrets” using the chatbot, which he had never used before.
He has vowed to never use AI to “supplement” his legal research in future “without absolute verification of its authenticity”.
The event comes as a warning that although AI like ChatGPT is a useful tool for information gathering, it should be used with caution, especially in professional environments.
Whilst it can provide human-like responses, it uses the internet as it was in 2021 as its database meaning some of its information can be outdated.