Why ChatGPT (still) needs a construction lawyer

Who is liable when a ‘borrowed’ construction worker gets hurt?

Why ChatGPT (still) needs a construction lawyer

Why ChatGPT (still) needs a construction lawyer


This audio is auto-generated. Please let us know if you have feedback.

This feature is a part of “The Dotted Line” series, which takes an in-depth look at the complex legal landscape of the construction industry. To view the entire series, click here.

Two and a half years ago, Construction Dive asked ChatGPT, OpenAI’s large language model, to generate a construction contract on a design-bid-build job for a 600-unit, mixed-use project in San Jose, California. The document, attorneys said, could be legally enforceable — it addressed scope of work, payment terms, a termination clause and sections on indemnification, insurance and change orders.

But it also came with missing clauses and poor risk management. Those same attorneys cautioned that using artificial intelligence to solely generate a contract was akin to opening Pandora’s Box — the risk was simply not worth it.

Construction attorneys now say the tech is here to stay and that its capabilities have improved as it has worked its way into myriad aspects of daily life. But after two years of improvement, can it now write an ironclad construction contract? More to the point, should construction attorneys turn to it to do so?

A headshot of Megan Shapiro.

Megan Shapiro

Permission granted by Megan Shapiro

 

“Oh my God, no. I would be terrified if I heard of anybody using it to generate a contract,” said Megan Shapiro, a construction attorney and partner at law firm Radoslovich Shapiro in Sacramento, California.

While AI does present uses for construction attorneys, experts cautioned that using it as a wholesale, one-stop-shop for contracts can only lead to trouble.

“It’s not something where you press a couple buttons and you have a contract that’s ready to go,” said Michael Vardaro, managing partner of New York City-based law firm Zetlin & De Chiara.

The hallucination issue

Shapiro didn’t mince words when it came to AI tools’ accuracy. She said it was “shocking” how wrong AI-generated content could be and how often that occurred.

Indeed, so-called hallucinations have been a massive problem for large language models. While that’s been the case ever since generative AI made its mainstream debut in 2022, a bug developers pledged would become less prevalent with use and the tech’s evolution, The New York Times reported in May that for newer AI systems, the hallucination rates reached as high as 79%, depending on the test and tool.

For attorneys, this presents a significant business and ethics problem as they look to balance speed with accuracy. Case in point: the advent of LLMs has given rise to “AI slop” in the legal system, or filings with bogus or misrepresented court cases, the Times reported on Nov. 7. Currently, a database that tracks AI hallucination cases in legal systems across the world has included, at the time of publication, about 600 cases — the majority of them in the U.S.

Due diligence

These issues present reputational and business risks, said Trent Cotney, a partner at law firm Adams & Reese and the head of its construction team. Should an attorney file a brief with AI-fabricated documents or claims, the presiding judge could consider sanctions.

Attorney Trent Cotney

Trent Cotney

Courtesy of Adams and Reese LLP

 

“Not to mention the fact that there’s a reputational hit, and your client obviously is not going to be happy with that either,” Cotney said.

Shapiro said that what keeps her up at night is the thought that her clients could be using AI on their own without discussing it with her. She fears that if someone at a construction firm doesn’t want to pay to call an attorney on that legal issue, then that person may attempt to figure the problem out on their own.

“I think the risk to them is getting substantially worse, because I think consumer confidence in the output is increasing, despite the fact that I think the quality of the output is decreasing,” Shapiro said.

Finding an edge

With diligence and a careful eye, however, attorneys said there are ways AI can help legal professionals with their daily tasks.

A headshot of Michael Vardaro

Michael Vardaro

Permission granted by Zetlin & De Chiara

 

Vardaro, for example, said one of the benefits of AI was its ability to corral documents and seek out sections contained within them, a job that was left to associates years ago.

“Now, AI can kind of do that and give you kind of the hits in that project file within minutes, where that might have taken weeks, if not months, to do,” Vardaro said.



Source link