📓 Word of the Week: GEO (Generative Engine Optimization)

Generative Engine Optimization empowers modern lawyers with AI-driven legal marketing!

In legal marketing, GEO—Generative Engine Optimization—is the next evolution beyond traditional SEO. GEO focuses on making your content understandable, trustworthy, and quotable by generative AI systems like ChatGPT, Gemini, Copilot, Perplexity, and Google’s AI experiences. 🧠

Traditional SEO was about ranking in a list of blue links. GEO is about becoming the source that AI tools cite when a potential client asks a legal question in natural language. For lawyers, this means writing clear, jurisdiction-specific, client‑focused answers that AI can safely lift into its responses.

Under ABA Model Rule 1.1, technological competence now includes understanding the benefits and risks of AI tools you use in practice and in marketing. 📚 GEO is not optional “extra credit” anymore, it is part of staying reasonably up to date with “the benefits and risks associated with relevant technology.”

From SEO to GEO for Lawyers

SEO still matters. You still need solid titles, meta descriptions, and clear on‑page structure so Google and other search engines can crawl and index your site. What changes with GEO is the audience for your content expands from humans and search bots to large language models that want direct, conversational, and well‑structured answers.

Think of it this way:

  • SEO asks, “How do I rank for ‘divorce lawyer Toronto’?”

  • GEO asks, “How do I become the answer when someone asks, ‘How does divorce work in Ontario and when should I call a lawyer?’ in an AI chat box?” 🇨🇦

  • Effective GEO content for law firms tends to share these traits:

    • Answer‑first summaries at the top of the page.

    • Clear jurisdiction and practice‑area signals.Plain‑English explanations of specific client questions.

    • Updated timestamps and trustworthy citations to statutes, rules, and court sites.

For attorneys with limited or moderate tech skills, this is less about learning code and more about tightening how you explain your work online. GEO rewards the same skills you already use in client communications: clarity, precision, and staying within your lane. ✅

GEO and the ABA Model Rules ⚖️

Ethical AI use strengthens confidentiality, competence, and trust in legal practice!

GEO strategy touches several ABA Model Rules that govern how you use AI and publish legal content:

  • Model Rule 1.1 – Competence. ABA guidance on AI (e.g., Formal Opinion 512) explains that competence includes understanding how AI tools work, their limitations, and their failure modes. If you expect clients to find you through AI answers, you should understand what those systems are likely to say about your practice area and how your content feeds into them.

  • Model Rule 1.6 – Confidentiality. GEO does not require you to feed client facts into AI systems. You can build GEO‑optimized content using hypotheticals and public information. When you do use AI tools to draft or refine content, you must confirm how the tool handles data, whether it trains on your prompts, and whether additional client consent is needed. 🔐

  • Model Rule 1.4 – Communication. When AI tools materially affect how a matter is handled, ABA guidance suggests you may need to discuss that with clients. In marketing, that translates to accurate disclaimers: clearly state that your GEO‑friendly pages are “general information, not legal advice,” and that an AI‑generated summary is no substitute for a direct consultation.

  • Model Rules 7.1–7.3 – Advertising and Solicitation. GEO content must remain truthful, non‑misleading, and consistent with advertising rules. Avoid guarantees, avoid puffery about being “the best,” and ensure that AI‑oriented content still reflects actual experience and jurisdictional limits.

Handled well, GEO can support your ethical duties: it helps you publish accurate, current, and educational information that clients and AI tools can rely on.

Practical GEO Steps for Law Firms

Difference between SEO and GEO shapes modern legal marketing and AI visibility.

Here are concrete ways to start moving from SEO to GEO without overhauling your entire site:

  1. Rewrite key pages with answer‑first structures. Open with a 3–5 sentence plain‑English answer to the main question, then expand with headings and FAQs.

  2. Add jurisdiction markers everywhere it matters. Include the province or state, city, and court level on your practice pages and FAQs.

  3. Build detailed FAQ hubs around real client questions in your niche, using conversational phrasing that mirrors how people talk to AI tools. 💬

  4. Strengthen E‑E‑A‑T signals: list credentials, publications, bar memberships, and awards; link to reputable external sources; keep author bylines current.

  5. Maintain technical SEO basics: fast, mobile‑friendly pages with clear title tags, meta descriptions, headings, and schema markup (e.g., for FAQs and legal services).

  6. Regularly refresh high‑value pages to keep them current with legal changes and to signal freshness to both search engines and AI systems. 🔁

  7. You do not need to do everything at once. Start with one practice area, identify the ten most common questions, and create a GEO‑optimized resource page that you would be comfortable seeing quoted by an AI tool.

WoW: “Telephobia” in Law Practice: How Fear of Phone Calls Hurts Lawyers, Clients, and Cases 📞⚖️

Fear of phone 📞 calls creates anxiety and impacts legal competence. ⚖️

Telephobia is the fear or intense anxiety associated with making or receiving phone calls, and it shows up more often in law practice than many lawyers admit. 😬📱 Telephobia is not a dislike of the telephone as an object; it is a form of social anxiety centered on real‑time verbal communication, fear of judgment, and the pressure to respond quickly without the safety net of drafting and editing. Lawyers who excel in written advocacy can still feel a spike of anxiety when the phone lights up with a client, partner, or opposing counsel. This reluctance to pick up or dial out is not a character flaw; it is a risk factor that can affect competence, communication, and client service.

What Telephobia Looks Like for Lawyers

Telephobia often appears as avoidance rather than obvious panic. Lawyers may let calls go to voicemail, delay returning calls, or delegate phone calls whenever possible. You might recognize behaviors such as over‑reliance on email, extensively scripting what you plan to say before dialing, or replaying conversations in your head for hours after hanging up. These patterns are common in people with phone anxiety and can exist on a spectrum from mild discomfort to significant impairment.

In legal practice, that avoidance has concrete consequences. Time‑sensitive issues sit in the inbox instead of getting resolved in a five‑minute call. Misunderstandings grow because no one is willing to pick up the phone and clarify. Judges and clients may perceive “radio silence” as a lack of diligence, even when the real issue is anxiety about the call itself. Over time, telephobia can contribute to bottlenecks in case management, strained relationships, and missed opportunities to resolve disputes early.

Telephobia, Opposing Counsel, and Professionalism

Telephone conversations with opposing counsel are still one of the most effective tools for narrowing issues, avoiding motion practice, and reaching practical solutions. Many experienced litigators emphasize the value of “picking up the phone” instead of escalating via email volleys. Yet telephobia can make newer or more anxious lawyers dread direct calls with adversaries, especially those who are aggressive, fast‑talking, or prone to “verballing” (misstating or spinning what was said in the conversation).

Avoiding phone contact with opposing counsel can have several impacts:

  • It can prolong discovery disputes that might have been resolved in a short meet‑and‑confer call.

  • It can increase the tone and temperature of written communications because nuance and rapport are missing.

  • It can reduce opportunities to build professional relationships that later help with scheduling, stipulations, or informal resolutions.

On the other hand, telephobia does not mean a lawyer should accept every unscheduled call or tolerate abusive conversations. Thoughtful boundaries are appropriate. Some practitioners manage risk by taking (or perhaps returning) calls only at set times, ensuring a colleague is nearby, or contemporaneously documenting the substance of the call in a follow‑up email. The key is intentional management, not blanket avoidance.

Telephobia and Client Communication Duties

Avoiding phone calls strains client Relations, and professionalism failure.

Telephobia directly intersects with your ethical duty to communicate with clients. ABA Model Rule 1.4 requires lawyers to keep clients reasonably informed and to promptly comply with reasonable requests for information. Modern guidance recognizes that “client communications” include phone calls, emails, and other electronic channels. If anxiety leads to chronic delay in returning calls or to a pattern of pushing every interaction into email when a call would be more effective, the lawyer may be edging toward a communication problem, not just a preference.

Clients often interpret unanswered calls as a sign of indifference. Many clients—especially those under stress—need a live conversation to feel heard and to understand their case strategy. While written follow‑up is essential, a short, empathetic phone call can prevent distrust and complaints. Telephobia can also create inequity: clients who are comfortable with email may get robust contact, while those who rely on the phone feel neglected.

At the same time, ethics authorities acknowledge that lawyers can use multiple communication tools, not just phone calls, as long as communication is prompt, understandable, and appropriate to the client’s needs. For some neurodivergent lawyers or lawyers with genuine anxiety disorders, establishing a communication plan that mixes scheduled calls, video meetings, and structured emails can satisfy both client needs and the lawyer’s mental health needs. Clear expectation‑setting is critical.

Technology Competence and the Phone in a Digital Age

ABA Model Rule 1.1, Comment 8, emphasizes that competence now includes understanding the benefits and risks associated with relevant technology. Many lawyers hear “technology competence” and think about e‑discovery platforms or cybersecurity, not the humble phone. Yet modern telephony—VoIP, softphones, smartphone apps, call‑recording tools, and integrated practice‑management systems—is very much part of that competence landscape.

For lawyers with telephobia, technology can both help and hinder:

  • VoIP and softphone systems can route calls through your laptop, support call notes, and provide voicemail‑to‑email transcripts, which can reduce anxiety about missing key points.

  • Scheduled video or audio calls through secure platforms can feel more controlled, especially when combined with a shared agenda.

  • Over‑reliance on text‑based channels (email, messaging) because they feel safer can, however, undermine the advantages of real‑time voice communication.

Competence does not require you to love the phone. It does require that you understand the tools available, use them to communicate effectively, and avoid letting anxiety silently undercut your ability to serve clients and manage cases.

Practical Strategies to Manage Telephobia in Practice

Telephobia is manageable, and many of the strategies come from established approaches to phone anxiety. The aim is not to turn every lawyer into an extroverted caller. The aim is to reduce the anxiety enough that telephony becomes a functional, ethical communication tool rather than a source of procrastination.

Practical steps include:

  • Use structured call plans. Before a client or opposing‑counsel call, sketch a brief outline: goals, key points, and closing next steps. This reduces the “blank mind” fear and keeps calls efficient.

  • Start with low‑stakes calls. Build tolerance by making brief, simple calls (e.g., scheduling, confirmations) rather than jumping straight into high‑conflict negotiations.

  • Schedule instead of surprise. Use calendar invites or quick emails: “Can we set a 10‑minute call at 2:30 p.m. to discuss X?” Predictability lowers anxiety for both you and the other side.

  • Pair calls with written follow‑up. After important calls, send a confirming email summarizing agreements and action items. This supports clarity, protects the record, and reassures anxious lawyers who worry they misspoke.

  • Leverage firm support. For very difficult conversations, consider having a colleague present (on the call or in the room), both for support and as a witness.

  • Seek professional help when needed. When anxiety is persistent, intense, or interfering with your practice, consulting a mental health professional familiar with social anxiety or telephobia is a sign of professionalism, not weakness.

These techniques align with ethical duties rather than conflict with them. They help ensure prompt, clear communication (Model Rule 1.4) and support technological and practical competence (Model Rule 1.1) in a digital environment.

Telephobia, Wellness, and Culture in the Profession

Avoiding phone calls lead to miscommunication, delays, and frustration!

Finally, telephobia is also a wellness issue. The legal profession already carries high rates of stress, depression, and anxiety. Telephobia can add another layer of dread to a typical workday, as lawyers watch call notifications with a racing pulse. Open conversation about phone anxiety—especially among younger lawyers and those trained in email‑first environments—can normalize the experience and lead to practical accommodations.v

Mentors and firm leaders can help by modeling balanced behavior. That includes choosing calls when they will truly advance the matter, avoiding unnecessary surprise calls that feel performative, and encouraging associates to prepare for and debrief difficult conversations. Thoughtful phone use, supported by technology and grounded in ethics, can turn telephobia from a hidden liability into a manageable professional challenge.

If you or someone you know is suffering from an imminent mental health crisis, call 988 (in the United States) or 911 or equivalent in the relevant jurisdiction!

🚨 ⛑️ 🚨

If you or someone you know is suffering from an imminent mental health crisis, call 988 (in the United States) or 911 or equivalent in the relevant jurisdiction! 🚨 ⛑️ 🚨

Word 📖 of the Week: Why Lawyers Need to Know the Term “Constitutional AI”

“Constitutional AI” is a design framework for artificial intelligence that aims to make AI systems helpful, harmless, and honest by training them to follow a defined set of higher‑level rules, much like a constitution. 🤖📜 For lawyers, this is not abstract theory; it connects directly to duties of technological competence, confidentiality, and supervision under the ABA Model Rules.

Most legal professionals now rely on AI‑enabled tools in research, drafting, e‑discovery, document automation, and client communication. These tools may use generative AI in the background even when the marketing materials do not emphasize “AI.” Constitutional AI gives you a practical way to evaluate those tools: are they structured to avoid hallucinations, protect confidential data, and resist being prompted into unethical behavior.

At a high level, a Constitutional AI system is trained to follow explicit principles, such as “do not fabricate legal citations,” “do not disclose confidential information,” and “do not assist in unlawful conduct.” The model learns to critique and revise its own outputs against those principles. For law firms, that aligns with the core expectations in ABA Model Rule 1.1 (competence) and its Comment 8, which require lawyers to understand the benefits and risks of relevant technology and stay current with changes in how these systems work. ⚖️

Constitutional AI also intersects with ABA Model Rule 1.6 on confidentiality. If an AI tool is not designed with strong guardrails, prompts, and outputs can expose sensitive client information to external systems or vendors. When you evaluate an AI platform, you should ask where data is stored, how prompts are logged, whether training data will include your matters, and whether the provider has implemented “constitutional” safeguards against data leakage and unsafe uses.

Supervision is another critical angle. ABA Formal Opinion 512 and Model Rules 5.1 and 5.3 stress that supervising lawyers must set policies and training for how attorneys and staff use generative AI. Constitutional AI can reduce risk, yet it does not replace supervisory duties. You still must review AI‑generated work product, confirm citations, validate factual assertions, and ensure the output is consistent with Rules 3.1, 3.3, and 8.4(c) on meritorious claims, candor to the tribunal, and avoiding dishonesty or misrepresentation.

For practitioners with limited to moderate tech skills, the key is to treat Constitutional AI as a practical checklist rather than a buzzword. ✅ Ask three questions about any AI tool you use:

  1. Is this AI actually helpful to the client’s matter, or is it just saving time while adding risk.

  2. Could this output harm the client through inaccuracy, bias, or disclosure of confidential data.

  3. Is the AI acting honestly, meaning it is not hallucinating cases or claiming certainty where none exists.

If any answer is “no,” you must pause, verify, and revise before relying on the AI output.

In the AI era, your ethical risk often turns on how you select, supervise, and document the use of AI in your practice. Constitutional AI will not make you bulletproof, but it gives you a structured way to align your technology choices with ABA Model Rules while protecting your clients, your license, and your reputation. 

Word of the Week: Deepfakes: How Lawyers Can Spot Fake Digital Evidence and Avoid ABA Model Rule Violations ⚖️

A Tech-Savvy Lawyer needs to be able to spot Deepfakes Before Courtroom Ethics Violations!

“Deepfakes” are AI‑generated or heavily manipulated audio, video, or images that convincingly depict people saying or doing things that never happened.🧠 They are moving from internet novelty to everyday litigation risk, especially as parties try to slip fabricated “evidence” into the record.📹

Recent cases and commentary show courts will not treat deepfakes as harmless tech problems. Judges have dismissed actions outright and imposed severe sanctions when parties submit AI‑generated or altered media, because such evidence attacks the integrity of the judicial process itself.⚖️ At the same time, courts are wary of lawyers who cry “deepfake” without real support, since baseless challenges can look like gamesmanship rather than genuine concern about authenticity.

For practicing lawyers, deepfakes are first and foremost a professional responsibility issue. ABA Model Rule 1.1 (Competence) now clearly includes a duty to understand the benefits and risks of relevant technology, which includes generative AI tools that create or detect deepfakes. You do not need to be an engineer, but you should recognize common red flags, know when to request native files or metadata, and understand when to bring in a qualified forensic expert.

Deepfakes in Litigation: Detect Fake Evidence, Protect Your License!

Deepfakes also implicate Model Rule 3.3 (Candor to the tribunal) and Model Rule 3.4 (Fairness to opposing party and counsel). If you knowingly offer manipulated media, or ignore obvious signs of fabrication in your client’s “evidence,” you risk presenting false material to the court and obstructing access to truthful proof. Courts have made clear that submitting fake digital evidence can justify terminating sanctions, fee shifting, and referrals for disciplinary action.

Model Rule 8.4(c), which prohibits conduct involving dishonesty, fraud, deceit, or misrepresentation, sits in the background of every deepfake decision. A lawyer who helps create, weaponize, or strategically “look away” from deepfake evidence is not just making a discovery mistake; they may be engaging in professional misconduct. Likewise, a lawyer who recklessly accuses an opponent of using deepfakes without factual grounding risks violating duties of candor and professionalism.

Practically, you can start protecting your clients with a few repeatable steps. Ask early in the case what digital media exists, how it was created, and who controlled the devices or accounts.🔍 Build authentication into your discovery plan, including requests for original files, device logs, and platform records that can help confirm provenance. When the stakes justify it, consult a forensic expert rather than relying on “gut feel” about whether a recording “looks real.”

lawyers need to know Deepfakes, Metadata, and ABA Ethics Rules!

Finally, talk to clients about deepfakes before they become a problem. Explain that altering media or using AI to “clean up” evidence is dangerous, even if they believe they are only fixing quality.📲 Remind them that courts are increasingly sophisticated about AI and that discovery misconduct in this area can destroy otherwise strong cases. Treat deepfakes as another routine topic in your litigation checklist, alongside spoliation and privilege, and you will be better prepared for the next “too good to be true” video that lands in your inbox.

Word of the Week: Vendor Risk Management for Law Firms in 026: Lessons from the Clio–Alexi CRM Fight ⚖️💻

Clio vs. Alexi: CRM Litigation COULD THREATEN Law Firm Data

“Vendor risk management” is no longer an IT buzzword; it is now a core law‑practice skill for any attorney who relies on cloud‑based tools, CRMs, or AI‑driven research platforms.⚙️📊 The Tech‑Savvy Lawyer.Page’s February 2, 2026 editorial on the Clio–Alexi CRM litigation showed how a dispute between legal‑tech companies can reach straight into your client list, calendars, and workflows.⚖️🧾

In that piece, Clio and Alexi’s legal fight over data, AI training, and competition was framed not as “tech drama,” but as a live test of how well your firm understands its dependencies on vendors that control client‑related information.🧠📂 When the platform that hosts your CRM, matter data, or AI research tools becomes embroiled in high‑stakes litigation, your risk profile changes even if you never set foot in that courtroom.⚠️🏛️

Under ABA Model Rule 1.1, competence includes a practical understanding of the technology that underpins your practice, and that now clearly includes vendor risk.📚💡 You do not have to reverse‑engineer APIs, yet you should be able to answer basic questions: Which vendors are mission‑critical, what data do they hold, how would you respond if one faced an injunction, outage, or rushed acquisition.🧩🚨 That is vendor risk management at a level that is realistic for lawyers with limited to moderate tech skills.🙂🧑‍💼

LawyerS NEED TO Build Vendor Risk Plan for Ethical Compliance

Model Rule 1.6 on confidentiality sits at the center of this analysis, because litigation involving a vendor can expose or pressure the systems that hold client information.🔐📁 Our February 2 article emphasized the need to know where your data is hosted, what the contracts say about subpoenas and law‑enforcement requests, and how quickly you can export data if your ethics analysis changes.⏱️📄 Vendor risk management, therefore, includes reviewing terms of service, capturing “current” versions of online agreements, and documenting export rights and notice obligations.📝🧷

Model Rule 5.3 requires reasonable efforts to ensure that non‑lawyer assistance is compatible with your professional duties, and 2026 legal‑tech commentary increasingly treats vendors as supervised extensions of the law office.🧑‍⚖️🤝 CRMs, AI research tools, document‑automation platforms, and e‑billing systems all act as non‑lawyer assistants for ethics purposes, which means you must screen them before adoption, monitor them for material changes, and reassess when events like the Clio–Alexi dispute surface.📡📊

Recent legal‑tech reporting has described 2026 as a reckoning year for vendors, with AI‑driven tools under heavier regulatory and client scrutiny, which makes disciplined vendor risk management a competitive advantage rather than a burden.📈🤖 Practical steps include maintaining a simple vendor inventory, ranking systems by criticality, reviewing cyber and data‑security representations, and identifying a plausible backup provider for each crucial function.📋🛡️

LAWYERS NEED TO SHIELD THEIR CLIENT DATA FROM CRM LITIGATION AS MUCH AS THEY NEED TO PROTECT THEIR EthicS DUTIES!

Vendor risk management, properly understood, turns your technology stack into part of your professional judgment instead of a black box that “IT” owns alone.🧱🧠 For solo and small‑firm lawyers, that shift can feel incremental rather than overwhelming: start by reading the Clio–Alexi editorial, pull your top three vendor contracts, and ask whether they let you protect competence, confidentiality, and continuity if your vendors suddenly become the ones needing legal help.🧑‍⚖️🧰

Word of the week: “Legal AI institutional memory” engages core ethics duties under the ABA Model Rules, so it is not optional “nice to know” tech.⚖️🤖

Institutional Memory Meets the ABA Model Rules

“Legal AI institutional Memory” is AI that remembers how your firm actually practices law, not just what generic precedent says. It captures negotiation history, clause choices, outcomes, and client preferences across matters so each new assignment starts from experience instead of a blank page.

From an ethics perspective, this capability sits directly in the path of ABA Model Rule 1.1 on competence, Rule 1.6 on confidentiality, and Rule 5.3 on responsibilities regarding nonlawyer assistance (which now includes AI systems). Comment 8 to Rule 1.1 stresses that competent representation requires understanding the “benefits and risks associated with relevant technology,” which squarely includes institutional‑memory AI in 2026. Using or rejecting this technology blindly can itself create risk if your peers are using it to deliver more thorough, consistent, and efficient work.🧩

Rule 1.6 requires “reasonable efforts” to prevent unauthorized disclosure or access to information relating to representation. Because institutional memory centralizes past matters and sensitive patterns, it raises the stakes on vendor security, configuration, and firm governance. Rule 5.3 extends supervision duties to “nonlawyer assistance,” which ethics commentators and bar materials now interpret to include AI tools used in client work. In short, if your AI is doing work that would otherwise be done by a human assistant, you must supervise it as such.🛡️

Why Institutional Memory Matters (Competence and Client Service)

Tools like Luminance and Harvey now market institutional‑memory features that retain negotiation patterns, drafting preferences, and matter‑level context across time. They promise faster contract cycles, fewer errors, and better use of a firm’s accumulated know‑how. Used wisely, that aligns with Rule 1.1’s requirement that you bring “thoroughness and preparation” reasonably necessary for the representation, and Comment 8’s directive to keep abreast of relevant technology.

At the same time, ethical competence does not mean turning judgment over to the model. It means understanding how the system makes recommendations, what data it relies on, and how to validate outputs against your playbooks and client instructions. Ethics guidance on generative AI emphasizes that lawyers must review AI‑generated work product, verify sources, and ensure that technology does not substitute for legal judgment. Legal AI institutional memory can enhance competence only if you treat it as an assistant you supervise, not an oracle you obey.⚙️

Legal AI That Remembers Your Practice—Ethics Required, Not Optional

How Legal AI Institutional Memory Works (and Where the Rules Bite)

Institutional‑memory platforms typically:

  • Ingest a corpus of contracts or matters.

  • Track negotiation moves, accepted fall‑backs, and outcomes over time.

  • Expose that knowledge through natural‑language queries and drafting suggestions.

That design engages several ethics touchpoints🫆:

  • Rule 1.1 (Competence): You must understand at a basic level how the AI uses and stores client information, what its limitations are, and when it is appropriate to rely on its suggestions. This may require CLE, vendor training, or collaboration with more technical colleagues until you reach a reasonable level of comfort.

  • Rule 1.6 (Confidentiality): You must ensure that the vendor contract, configuration, and access controls provide “reasonable efforts” to protect confidentiality, including encryption, role‑based access, and breach‑notification obligations. Ethics guidance on cloud and AI use stresses the need to investigate provider security, retention practices, and rights to use or mine your data.

  • Rule 5.3 (Nonlawyer Assistance): Because AI tools are “non‑human assistance,” you must supervise their work as you would a contract review outsourcer, document vendor, or litigation support team. That includes selecting competent providers, giving appropriate instructions, and monitoring outputs for compliance with your ethical obligations.🤖

Governance Checklist: Turning Ethics into Action

For lawyers with limited to moderate tech skills, it helps to translate the ABA Model Rules into a short adoption checklist.✅

When evaluating or deploying legal AI institutional memory, consider:

  1. Define Scope (Rules 1.1 and 1.6): Start with a narrow use case such as NDAs or standard vendor contracts, and specify which documents the system may use to build its memory.

  2. Vet the Vendor (Rules 1.6 and 5.3): Ask about data segregation, encryption, access logs, regional hosting, subcontractors, and incident‑response processes; confirm clear contractual obligations to preserve confidentiality and notify you of incidents.

  3. Configure Access (Rules 1.6 and 5.3): Use role‑based permissions, client or matter scoping, and retention settings that match your existing information‑governance and legal‑hold policies.

  4. Supervise Outputs (Rules 1.1 and 5.3): Require that lawyers review AI suggestions, verify sources, and override recommendations where they conflict with client instructions or risk tolerance.

  5. Educate Your Team (Rule 1.1): Provide short trainings on how the system works, what it remembers, and how the Model Rules apply; document this as part of your technology‑competence efforts.

Educating Your Team Is Core to AI Competence

This approach respects the increasing bar on technological competence while protecting client information and maintaining human oversight.⚖️

This approach respects the increasing bar on technological competence while protecting client information and maintaining human oversight.⚖️

Word of the Week: What is a “Token” in AI parlance?

Lawyers need to know what “tokens” are in ai jargon!

In artificial intelligence, a “token” is a small segment of text—such as a word, subword, or even punctuation—that AI tools like ChatGPT or other large language models (LLMs) use to understand and generate language. In simple terms, tokens are the “building blocks” of communication for AI. When you type a sentence, the system breaks it into tokens so it can analyze meaning, predict context, and produce a relevant response.

For example, the sentence “The court issued its opinion.” might be split into six tokens: “The,” “court,” “issued,” “its,” “opinion,” and “.” By interpreting how those tokens relate, the AI produces natural and coherent language that feels human-like.

This concept matters to law firms and practitioners because AI systems often measure capacity and billing by token count, not by word count. AI-powered tools used for document review, legal research, and e-discovery commonly calculate both usage and cost based on the number of tokens processed. Naturally, longer or more complex documents consume more tokens and therefore cost more to analyze. As a result, a lawyer’s AI platform may also be limited in how much discovery material it can process at once, depending on the platform’s token capacity.

lawyers have an ethical duty to know how tokens apply when using ai in their legal work!~

But there’s a second, more important dimension to tokens: ethics and professional responsibility. The ABA Model Rules of Professional Conduct—particularly Rules 1.1 (Competence), 1.6 (Confidentiality of Information), and 5.3 (Responsibilities Regarding Nonlawyer Assistance)—apply directly when lawyers use AI tools that process client data.

  • Rule 1.1 requires technological competence. Attorneys must understand how their chosen AI tools function, at least enough to evaluate token-based costs, data use, and limitations.

  • Rule 1.6 restricts how client confidential information may be shared or stored. Submitting text to an AI system means tokens representing that text may travel through third-party servers or APIs. Lawyers must confirm the AI tool’s data handling complies with client confidentiality obligations.

  • Rule 5.3 extends similar oversight duties when relying on vendors that provide AI-based services. Understanding what happens to client data at the token level helps attorneys fulfill those responsibilities.

a “token” is a small segment of text.

In short, tokens are not just technical units. They represent the very language of client matters, billing data, and confidential work. Understanding tokens helps lawyers ensure efficient billing, maintain confidentiality, and stay compliant with professional ethics rules while embracing modern legal technology.

Tokens may be tiny units of text—but for lawyers, they’re big steps toward ethical, informed, and confident use of AI in practice. ⚖️💡

Word of the Week: "Constitutional AI" for Lawyers - What It Is, Why It Matters for ABA Rules, and How Solo & Small Firms Should Use It!

Constitutional AI’s ‘helpful, harmless, honest’ standard is a solid starting point for lawyers evaluating AI platforms.

The term “Constitutional AI” appeared this week in a Tech Savvy Lawyer post about the MTC/PornHub breach as a cybersecurity wake‑up call for lawyers 🚨. That article used it to highlight how AI systems (like those law firms now rely on) must be built and governed by clear, ethical rules — much like a constitution — to protect client data and uphold professional duties. This week’s Word of the Week unpacks what Constitutional AI really means and explains why it matters deeply for solo, small, and mid‑size law firms.

🔍 What is Constitutional AI?

Constitutional AI is a method for training large language models so they follow a written set of high‑level principles, called a “constitution” 📜. Those principles are designed to make the AI helpful, honest, and harmless in its responses.

As Claude AI from Anthropic explains:
Constitutional AI refers to a set of techniques developed by researchers at Anthropic to align AI systems like myself with human values and make us helpful, harmless, and honest. The key ideas behind Constitutional AI are aligning an AI’s behavior with a ‘constitution’ defined by human principles, using techniques like self‑supervision and adversarial training, developing constrained optimization techniques, and designing training data and model architecture to encode beneficial behaviors.” — Claude AI, Anthropic (July 7th, 2023).

In practice, Constitutional AI uses the model itself to critique and revise its own outputs against that constitution. For example, the model might be told: “Do not generate illegal, dangerous, or unethical content,” “Be honest about what you don’t know,” and “Protect user privacy.” It then evaluates its own answers against those rules before giving a final response.

Think of it like a junior associate who’s been given a firm’s internal ethics manual and told: “Before you send that memo, check it against these rules.” Constitutional AI does that same kind of self‑checking, but at machine speed.

🤝 How Constitutional AI Relates to Lawyers

For lawyers, Constitutional AI is important because it directly shapes how AI tools behave when handling legal work 📚. Many legal AI tools are built on models that use Constitutional AI techniques, so understanding this concept helps lawyers:

  • Judge whether an AI assistant is likely to hallucinate, leak sensitive info, or give ethically problematic advice.

  • Choose tools whose underlying AI is designed to be more transparent, less biased, and more aligned with professional norms.

  • Better supervise AI use in the firm, which is a core ethical duty under the ABA Model Rules.

Solo and small firms, in particular, often rely on off‑the‑shelf AI tools (like chatbots or document assistants). Knowing that a tool is built on Constitutional AI principles can give more confidence that it’s designed to avoid harmful outputs and respect confidentiality.

⚖️ Why It Matters for ABA Model Rules

For solo and small firms, asking whether an AI platform aligns with Constitutional AI’s standards is a practical first step in choosing a trustworthy tool.

The ABA’s Formal Opinion 512 on generative AI makes clear that lawyers remain responsible for all work done with AI, even if an AI tool helped draft it 📝. Constitutional AI is relevant here because it’s one way that AI developers try to build in ethical guardrails that align with lawyers' obligations.

Key connections to the Model Rules:

  • Rule 1.1 (Competence): Lawyers must understand the benefits and risks of the technology they use. Knowing that a tool uses Constitutional AI helps assess whether it’s reasonably reliable for tasks like research, drafting, or summarizing.

  • Rule 1.6 (Confidentiality): Constitutional AI models are designed to refuse to disclose sensitive information and to avoid memorizing or leaking private data. This supports the lawyer’s duty to make “reasonable efforts” to protect client confidences.

  • Rule 5.1 / 5.3 (Supervision): Managing partners and supervising attorneys must ensure that AI tools used by staff are consistent with ethical rules. A tool built on Constitutional AI principles is more likely to support, rather than undermine, those supervisory duties.

  • Rule 3.3 (Candor to the Tribunal): Constitutional AI models are trained to admit uncertainty and avoid fabricating facts or cases, which helps reduce the risk of submitting false or misleading information to a court.

In short, Constitutional AI doesn’t relieve lawyers of their ethical duties, but it can make AI tools safer and more trustworthy when used under proper supervision.

🛡️ The “Helpful, Harmless, and Honest” Principle

The three pillars of Constitutional AI — helpful, harmless, and honest — are especially relevant for lawyers:

  • Helpful: The AI should provide useful, relevant information that advances the client’s matter, without unnecessary or irrelevant content.

  • Harmless: The AI should avoid generating illegal, dangerous, or unethical content, and should respect privacy and confidentiality.

  • Honest: The AI should admit when it doesn’t know something, avoid fabricating facts or cases, and not misrepresent its capabilities.

For law firms, this “helpful, harmless, and honest” standard is a useful mental checklist when using AI:

  • Is this AI output actually helpful to the client’s case?

  • Could this output harm the client (e.g., by leaking confidential info or suggesting an unethical strategy)?

  • Is the AI being honest (e.g., not hallucinating case law or pretending to know facts it can’t know)?

If the answer to any of those questions is “no,” the AI output should not be used without significant human review and correction.

🛠️ Practical Takeaways for Law Firms

For solo, small, and mid‑size firms, here’s how to put this into practice:

Lawyers need to screen AI tools and ensure they are aligned with ABA Model Rules.

  1. Know your tools. When evaluating a legal AI product, ask whether it’s built on a Constitutional AI–style model (e.g., Claude). That tells you it’s designed with explicit ethical constraints.

  2. Treat AI as a supervised assistant. Never let AI make final decisions or file work without a lawyer’s review. Constitutional AI reduces risk, but it doesn’t eliminate the need for human judgment.

  3. Train your team. Make sure everyone in the firm understands that AI outputs must be checked for accuracy, confidentiality, and ethical compliance — especially when using third‑party tools.

  4. Update your engagement letters and policies. Disclose to clients when AI is used in their matters, and explain how the firm supervises it. This supports transparency under Rule 1.4 and Rule 1.6.

  5. Focus on “helpful, honest, harmless.” Use Constitutional AI as a mental checklist: Is this AI being helpful to the client? Is it honest about its limits? Is it harmless (no bias, no privacy leaks)? If not, don’t rely on it.

Words of the Week: “ANTHROPIC” VS. “AGENTIC”: UNDERSTANDING THE DISTINCTION IN LEGAL TECHNOLOGY 🔍

lawyers need to know the difference anthropic v. agentic

The terms "Anthropic" and "agentic" circulate frequently in legal technology discussions. They sound similar. They appear in the same articles. Yet they represent fundamentally different concepts. Understanding the distinction matters deeply for legal practitioners seeking to leverage artificial intelligence effectively.

Anthropic is a company—specifically, an AI safety-focused organization that develops large language models, most notably Claude. Think of Anthropic as a technology provider. The company pioneered "Constitutional AI," a training methodology that embeds explicit principles into AI systems to guide their behavior toward helpfulness, harmlessness, and honesty. When you use Claude for legal research or document drafting, you are using a product built by Anthropic.

Agentic describes a category of AI system architecture and capability—not a company or product. Agentic systems operate autonomously, plan multi-step tasks, make decisions dynamically, and execute workflows with minimal human intervention. An agentic system can break down complex assignments, gather information, refine outputs, and adjust its approach based on changing circumstances. It exercises judgment about which tools to deploy and when to escalate matters to human oversight.

"Constitutional AI" is an ai training methodology promoting helpfulness, harmlessness, and honesty in ai programing

The relationship between these concepts becomes clearer through a practical scenario. Imagine you task an AI system with analyzing merger agreements from a target company. A non-agentic approach requires you to provide explicit instructions for each step: search the database, extract key clauses, compare terms against templates, and prepare a summary. You guide the process throughout. An agentic approach allows you to assign a goal—Review these contracts, flag risks, and prepare a risk summary—and the AI system formulates its own research plan, prioritizes which documents to examine first, identifies gaps requiring additional information, and works through the analysis independently, pausing only when human judgment becomes necessary.

Anthropic builds AI models capable of agentic behavior. Claude, Anthropic's flagship model, can function as an agentic system when configured appropriately. However, Anthropic's models can also operate in simpler, non-agentic modes. You might use Claude to answer a direct question or draft a memo without any agentic capability coming into play. The capability exists within Anthropic's models, but agentic functionality remains optional depending on your implementation.

They work together as follows: Anthropic provides the underlying AI model and the training methodology emphasizing constitutional principles. That foundation becomes the engine powering agentic systems. The Constitutional AI approach matters specifically for agentic applications because autonomous systems require robust safeguards. As AI systems operate more independently, explicit principles embedded during training help ensure they remain aligned with human values and institutional requirements. Legal professionals cannot simply deploy an autonomous AI agent without trust in its underlying decision-making framework.

Agentic vs. Anthropic: Know the Difference. Shape the Future of Law!

For legal practitioners, the distinction carries practical implications. You evaluate Anthropic as a vendor when selecting which AI provider's tools to adopt. You evaluate agentic architecture when deciding whether your specific use case requires autonomous task execution or whether simpler, more directed AI assistance suffices. Many legal workflows benefit from direct AI support without requiring full autonomy. Others—such as high-volume contract analysis during due diligence—leverage agentic capabilities to move work forward rapidly.

Both elements represent genuine advances in legal technology. Recognizing the difference positions you to make informed decisions about tool adoption and appropriate implementation for your practice. ✅

📖 WORD OF THE WEEK YEAR🥳:  Verification: The 2025 Word of the Year for Legal Technology ⚖️💻

all lawyers need to remember to check ai-generated legal citations

After reviewing a year's worth of content from The Tech-Savvy Lawyer.Page blog and podcast, one word emerged to me as the defining concept for 2025: Verification. This term captures the essential duty that separates competent legal practice from dangerous shortcuts in the age of artificial intelligence.

Throughout 2025, The Tech-Savvy Lawyer consistently emphasized verification across multiple contexts. The blog covered proper redaction techniques following the Jeffrey Epstein files disaster. The podcast explored hidden AI in everyday legal tools. Every discussion returned to one central theme: lawyers must verify everything. 🔍

Verification means more than just checking your work. The concept encompasses multiple layers of professional responsibility. Attorneys must verify AI-generated legal research to prevent hallucinations. Courts have sanctioned lawyers who submitted fictitious case citations created by generative AI tools. One study found error rates of 33% in Westlaw AI and 17% in Lexis+ AI. Note the study's foundation is from May 2024, but a 2025 update confirms these findings remain current—the risk of not checking has not gone away. "Verification" cannot be ignored.

The duty extends beyond research. Lawyers must verify that redactions actually remove confidential information rather than simply hiding it under black boxes. The DOJ's failed redaction of the Epstein files demonstrated what happens when attorneys skip proper verification steps. Tech-savvy readers simply copied text from beneath the visual overlays. ⚠️

use of ai-generated legal work requires “verification”, “Verification”, “Verification”!

ABA Model Rule 1.1 requires technological competence. Comment 8 specifically mandates that lawyers understand "the benefits and risks associated with relevant technology." Verification sits at the heart of this competence requirement. Attorneys cannot claim ignorance about AI features embedded in Microsoft 365, Zoom, Adobe, or legal research platforms. Each tool processes client data differently. Each requires verification of settings, outputs, and data handling practices. 🛡️

The verification duty also applies to cybersecurity. Zero Trust Architecture operates on the principle "never trust, always verify." This security model requires continuous verification of user identity, device health, and access context. Law firms can no longer trust that users inside their network perimeter are authorized. Remote work and cloud-based systems demand constant verification.

Hidden AI poses another verification challenge. Software updates automatically activate AI features in familiar tools. These invisible assistants process confidential client data by default. Lawyers must verify which AI systems operate in their technology stack. They must verify data retention policies. They must verify that AI processing does not waive attorney-client privilege. 🤖

ABA Formal Opinion 512 eliminates the "I didn't know" defense. Lawyers bear responsibility for understanding how their tools use AI. Rule 5.3 requires attorneys to supervise software with the same care they supervise human staff members. Verification transforms from a good practice into an ethical mandate.

verify your ai-generated work like your bar license depends on it!

The year 2025 taught legal professionals that technology competence means verification competence. Attorneys must verify redactions work properly. They must verify AI outputs for accuracy. They must verify security settings protect confidential information. They must verify that hidden AI complies with ethical obligations. ✅

Verification protects clients, preserves attorney licenses, and maintains the integrity of legal practice. As The Tech-Savvy Lawyer demonstrated throughout 2025, every technological advancement creates new verification responsibilities. Attorneys who master verification will thrive in the AI era. Those who skip verification steps risk sanctions, malpractice claims, and disciplinary action.

The legal profession's 2025 Word of the Year is verification. Master it or risk everything. 💼⚖️