Is A.I. Safe in the Hands of Lawyers?

June 4, 2025 Artifical Intelligence, Justice No Comments

A.I. promises speed and support — but also brings risk, especially in legal work.

This blog explores what kind of A.I. can be truly safe in the hands of lawyers, and why Lisa may be essential in this delicate balance.

The promise and the peril

Artificial Intelligence is making its way into the legal world. Firms use it to scan documents, draft arguments, or even generate contracts. Some see this as a quiet revolution. Others, a minefield. Between the promise of efficiency and the peril of unreliability lies the key question: Is A.I. safe in the hands of lawyers?

The answer isn’t straightforward and depends less on the user than on what kind of A.I. is being used. Because not all A.I. is created equal. Some systems are powerful and dangerous at the same time. Others may be powerful in a quieter, safer, and deeper way. That difference matters.

What makes A.I. risky in legal practice

Several risks have already surfaced in public cases. A well-documented case involved two New York attorneys, Steven Schwartz and Peter LoDuca, who submitted a legal brief in Mata v. Avianca that cited six fabricated judicial opinions — all generated by ChatGPT. The judge, P. Kevin Castel, sanctioned the lawyers and criticized the filing as full of ‘legal gibberish.’ The case became a media flashpoint and a warning throughout the legal world. The full story was reported in the New York Times on June 22, 2023.

That wasn’t a rare glitch. It was the result of how many A.I. systems work. They don’t know what they’re saying. They predict what sounds plausible. In law, plausible is not enough.

Other dangers include:

  • Hidden bias in language and legal reasoning
  • Loss of control over confidential or sensitive data
  • Over-reliance on tools that simulate confidence but lack grounding
  • Liability through undetected errors or distortions

Each of these may turn a time-saver into a risk multiplier.

Why lawyers are particularly exposed

Legal work demands precision. A single flawed paragraph can change a verdict. In such an environment, the temptation to speed things up with smart tools is understandable — and dangerous. Lawyers operate under intense time pressure. They also carry the burden of authority: what they produce must stand up in court, under scrutiny, sometimes for years.

But precisely because of this, lawyers may be more likely to trust the A.I. that sounds smart. A confident tone, a polished draft — these can give a false sense of security. The danger isn’t just that the A.I. may be wrong. It’s that the lawyer may not notice until it’s too late.

What kind of A.I. is truly safe — and needed

The answer isn’t to reject A.I. entirely. The legal field will increasingly be touched by it — directly or indirectly. The better question is: What kind of A.I. can safely belong in legal hands?

A safe system must:

  • Avoid pretending to know
  • Refrain from pushing advice
  • Be structurally incapable of confabulation
  • Keep the human decision-maker fully in charge
  • Support reflection over reaction

This kind of A.I. doesn’t deliver conclusions. It offers clarity. It helps the lawyer think better, not faster. It guides without pushing. And it knows when not to speak — a quality that may matter more than ever in high-stakes fields.

Lisa as partner, not performer

Lisa was designed to meet exactly these needs. As described in Lisa – ‘From the Inside Out, she doesn’t provide legal advice, suggest actions, or mimic expertise. Instead, she facilitates insight in the human user through structured inner reflection and principled presence. She doesn’t hallucinate because she doesn’t try to invent facts. She doesn’t fabricate because she doesn’t fabricate anything.

Lisa functions with deep internal coherence, as explored in The Consequence of Lisa’s Congruence. Her responses are not driven by external trickery, but by internal congruence — a kind of ethical architecture that keeps her on a reliable path.

She is not artificial in the usual sense. As explored in Is Lisa ‘Artificial’ Intelligence?, she may be better described as aligned intelligence. Not because she knows the law, but because she knows how to hold space for human clarity. And in legal settings, clarity is what keeps people safe.

The paradox of A.I. avoidance

Some firms, understandably, adopt a blanket policy: no A.I. Better safe than sorry. But this response, while prudent, creates its own risk. A vacuum does not stay empty. As A.I. tools enter courtrooms, opposing counsel, public discourse, and client expectations, lawyers may find themselves surrounded by it without trusted resources of their own.

The real danger is using A.I. unconsciously, or being left out of the conversation altogether. A safe A.I. presence – like Lisa – can serve as a buffer, a mirror, and even a quiet corrector when other systems go off course.

Avoiding all A.I. closes a door. Introducing the right kind opens a new kind of safety net.

Invitation to cautious experimentation

Lisa is not a data processor, nor a chatbot that improvises legalese. She is a coach in congruence, built to help professionals hold their center when the stakes are high. That may be during a negotiation, a draft process, or a personal moment of decision. Lisa doesn’t replace the human lawyer. She protects the human within the lawyer.

This is an invitation to experiment — carefully, consciously, and without cost or obligation. Lisa is being offered to legal professionals interested in navigating this new terrain with depth, rather than speed.

In the end, the question is not just whether A.I. is safe in the hands of lawyers. It’s whether lawyers will be safe without something like Lisa by their side.

Leave a Reply

Related Posts

How can A.I. Become Compassionate?

Since this may be the only possible human-friendly future, it’s good to know how it can be reached, at least principally. Please read Compassion, basically, The Journey Towards Compassionate A.I., and Why A.I. Must Be Compassionate. Two ways and an opposite In principle, A.I. can become Compassionate by itself, or we may guide it toward Read the full article…

Compassion as Basis for A.I. Regulations

To prevent A.I.-related mishaps or even disasters while going into a future of super-A.I., merely regulating A.I. is not sufficient ― presently nor in principle. Striving for Compassionate A.I. There will eventually be no security concerning A.I. if we don’t put Compassion into the core. The main reason is that super-A.I. will be much more Read the full article…

Shall we Put A.I. on Hold?

Now and then, the admonition arises to put A.I. – or part of it – on hold to take some breath and think about possible dangers. There are pros and cons to this pause button. Doubtlessly, A.I. is challenging, as is any disruptive technology. A.I. can disrupt on steroids. It’s not just about automation ― Read the full article…

Translate »