This week, I have been experimenting with Deep Research, the AI agent OpenAI released on Sunday that it says is capable of completing multi-step research tasks and synthesizing large amounts of online information. Not to be confused with the controversial Chinese AI product DeepSeek), Deep Research is said to be particularly useful for people in fields such as finance, science and law.
Already this week, I published two of these experiments. In the first, I used it to analyze the legality of President Trump’s pause of federal grants. In about 10 minutes, it produced a 9,000 word detailed memorandum, concluding that the pause “appears to rest on shaky legal ground.”
Next, I used it to research and recommend the best law practice management suite for a four-lawyer firm. It produced a fairly detailed response, including two charts comparing features, pricing, usability, security, support and user satisfaction.
For today’s task, I asked it to create a report detailing every legal ethics opinion pertaining to generative AI. Here was my exact prompt:
“Create a report detailing every legal ethics opinion from every national, state, local and specialty bar association or lawyer licensing body pertaining to the ethics of lawyers’ use of generative artificial intelligence.”
It responded to my prompt with several questions about the scope of the research I’d requested, such as whether it should focus only on formal ethics opinions or also include informal guidance. After I answered its questions, it produced the report published below. After it produced the report, I asked it to also summarize the findings in a chart, which is what you see immediately below.
I have not verified that this is a complete list. If anyone knows where I can find a complete list to compare against, please let me know.
That said, I was again impressed by its ability to conduct comprehensive research across multiple sources and generate a report. The entire task took it 15 minutes.
Issuing Body | Opinion Title/Number | Date Issued | Key Themes | Disclosure Required | Billing Guidance |
---|---|---|---|---|---|
ABA | Formal Opinion 512 | July 2024 | Competence, Confidentiality, Supervision, Candor, Fees | Case-dependent | AI efficiency should reduce fees |
California | Practical Guidance | November 2023 | Confidentiality, Competence, AI Disclosure, Supervision | Not mandatory, but recommended | Efficiency gains must be fair to clients |
Florida | Advisory Opinion 24-1 | January 2024 | Confidentiality, Supervision, Fees, AI Transparency | Client informed consent advised | Must not charge for AI time savings |
New York State Bar | Task Force Report | April 2024 | Competence, Bias, AI Oversight, Client Communication | Recommended in certain cases | Clients should benefit from AI-driven efficiency |
New York City Bar | Formal Opinion 2024-5 | August 2024 | Confidentiality, Competence, Supervision, AI Disclosure | Required for external AI use | No overcharging for AI use |
New Jersey Supreme Court | Preliminary Guidelines | January 2024 | Competence, Candor, Supervision, AI Verification | Not always, but recommended | No billing for time not actually spent |
Pennsylvania & Philadelphia Bars | Joint Opinion 2024-200 | June 2024 | Competence, Candor, AI Hallucinations, Billing Ethics | Yes, in some AI use cases | Cannot bill AI time as manual work |
Kentucky | Ethics Opinion KBA E-457 | March 2024 | Competence, AI Use in Billing, Supervision, Client Consent | Depends on AI’s impact on case | Cannot bill for AI learning time |
North Carolina | Formal Ethics Opinion 2024-1 | November 2024 | Competence, Supervision, Confidentiality, AI Oversight | Not always, but must protect confidentiality | AI-based efficiency should lower costs |
Texas | Proposed Opinion 2024-6 | November 2024 | Competence, Confidentiality, AI Trustworthiness, Supervision | Not explicitly required | Fair pricing required for AI use |
Virginia | AI Guidance Update | August 2024 | Confidentiality, Billing, Supervision, AI Court Compliance | Not mandated but recommended | AI costs must align with ethical billing |
D.C. Bar | Ethics Opinion 388 | September 2024 | Competence, AI Verification, Supervision, Client Files | Required in specific situations | No excess fees for AI use |
USPTO | Practice Guidance (2023–2024) | April 2024 | Candor, Confidentiality, AI Use in Legal Submissions | Court compliance required | Legal AI use cannot inflate costs |
National Bar Associations
American Bar Association – Formal Opinion 512 (July 2024)
The ABA Standing Committee on Ethics and Professional Responsibility issued Formal Opinion 512, “Generative Artificial Intelligence Tools,” on July 29, 2024. americanbar.org americanbar.org. This is the ABA’s first ethics guidance focused on generative AI use by lawyers. It instructs attorneys to “fully consider their applicable ethical obligations, including their duties to provide competent legal representation, to protect client information, to communicate with clients, to supervise their employees and agents, to advance only meritorious claims and contentions, to ensure candor toward the tribunal, and to charge reasonable fees.” jenkinslaw.org In short, existing ABA Model Rules apply to AI just as they do to any technology.
Key concerns and recommendations: The opinion emphasizes that lawyers must maintain technological competence – understanding the benefits and risks of AI tools they use jenkinslaw.org. It notes the duty of confidentiality (Model Rule 1.6) requires caution when inputting client data into AI tools; lawyers should ensure no confidential information is revealed without informed client consent jenkinslaw.org. Lawyers should also evaluate whether to inform or obtain consent from clients about AI use, especially if using it in ways that affect the representation jenkinslaw.org. AI outputs must be independently verified for accuracy to fulfill duties of candor and avoid filing false or frivolous material (Rules 3.3, 3.1) jenkinslaw.org. The ABA highlights that “hallucinations” (convincing but false outputs) are a major pitfall americanbar.org. Supervision duties (Rules 5.1 and 5.3) mean lawyers must oversee both subordinate lawyers and nonlawyers and the AI tools they use jenkinslaw.org. The opinion also warns that fees must be reasonable – if AI improves efficiency, lawyers should not overbill for time not actually spent kaiserlaw.com. Overall, Formal Op. 512 provides a comprehensive framework mapping generative AI use to existing ethics rules americanbar.org americanbar.org.
(See ABA Formal Op. 512 jenkinslaw.org for full text.)
State Bar Associations and Regulatory Bodies
California – “Practical Guidance” by COPRAC (November 2023)
The State Bar of California took early action by issuing “Practical Guidance for the Use of Generative AI in the Practice of Law,” approved by the Bar’s Board of Trustees on Nov. 16, 2023
calbar.ca.gov jdsupra.com. Rather than a formal opinion, it is a guidance document (in chart format) developed by the Committee on Professional Responsibility and Conduct (COPRAC). It applies California’s Rules of Professional Conduct to generative AI scenarios.
Key points: California’s guidance stresses confidentiality – attorneys “must not input any confidential client information” into AI tools that lack adequate protections calbar.ca.gov. Lawyers should vet an AI vendor’s security and data use policies, and anonymize or refrain from sharing sensitive data unless certain it will be protected calbar.ca.gov calbar.ca.gov. The duty of competence and diligence requires understanding how the AI works and its limitations jdsupra.com. Lawyers should review AI outputs for accuracy and bias, and “AI should never replace a lawyer’s professional judgment.” jdsupra.com If AI assists with research or drafting, the attorney must critically review the results. The guidance also addresses supervision: firms should train and supervise lawyers and staff in proper AI use jdsupra.com. Communication with clients may entail disclosing AI use in some cases – e.g. if it materially affects the representation – but California did not mandate disclosure in all instances jdsupra.com. Finally, the guidance notes candor: the duty of candor to tribunals means attorneys must check AI-generated citations and facts to avoid false statements in court jdsupra.com. Overall, California’s approach is to treat AI as another technology that must be used consistent with existing rules on competence, confidentiality, supervision, etc., providing “guiding principles rather than best practices” calbar.ca.gov.
(Source: State Bar of CA Generative AI Guidance jdsupra.com jdsupra.com.)
Florida – Advisory Opinion 24-1 (January 2024)
The Florida Bar issued Proposed Advisory Opinion 24-1 in late 2023, which was adopted by the Bar’s Board of Governors in January 2024 floridabar.org floridabar.org. Titled “Lawyers’ Use of Generative AI,” this formal ethics opinion gives a green light to using generative AI “to the extent that the lawyer can reasonably guarantee compliance with the lawyer’s ethical obligations.” floridabar.org It identifies four focus areas: confidentiality, oversight, fees, and advertising hinshawlaw.com hinshawlaw.com.
Key points: Confidentiality: Florida stresses that protecting client confidentiality (Rule 4-1.6) is paramount. Lawyers should take “reasonable steps to prevent inadvertent or unauthorized disclosure” of client info by an AI system jdsupra.com. The opinion “advisable to obtain a client’s informed consent before using a third-party AI that would disclose confidential information.”
jdsupra.com This aligns with prior cloud-computing opinions. Oversight: Generative AI must be treated like a non-lawyer assistant – the lawyer must supervise and vet its work jdsupra.com. The opinion warns that lawyers relying on AI face “the same perils as relying on an overconfident nonlawyer assistant” floridabar.org. Attorneys must review AI outputs (research, drafts, etc.) for accuracy and legal soundness before use floridabar.org. Notably, after the infamous Mata v. Avianca incident of fake cases, Florida emphasizes candor: no frivolous or false material from AI should be submitted floridabar.org. Fees: Improved efficiency from AI cannot be used to charge inflated fees. A lawyer “can ethically only charge a client for actual costs incurred” – time saved by AI should not be billed as if the lawyer did the work jdsupra.com. If a lawyer will charge for using an AI tool (as a cost), the client must be informed in writing jdsupra.com. And training time – a lawyer’s time learning an AI tool – cannot be billed to the client jdsupra.com. Advertising: If lawyers advertise their use of AI, they must not be false or misleading. Florida specifically notes that if using a chatbot to interact with potential clients, those users must be told they are interacting with an AI, not a human lawyer jdsupra.com. Any claims about an AI’s capabilities must be objectively verifiable (no puffery that your AI is “better” than others without proof) floridabar.org floridabar.org. In sum, Florida concludes: “a lawyer may ethically utilize generative AI, but only to the extent the lawyer can reasonably guarantee compliance with duties of confidentiality, candor, avoiding frivolous claims, truthfulness, reasonable fees, and proper advertising.” floridabar.org.
(Sources: Florida Bar Op. 24-1 floridabar.org jdsupra.com.)
New York State Bar Association – Task Force Report (April 2024)
The New York State Bar Association (NYSBA) did not issue a formal ethics opinion via its ethics committee, but its Task Force on Artificial Intelligence produced a comprehensive 85-page report adopted by the House of Delegates on April 6, 2024 floridabar.org floridabar.org. This report includes a chapter on the “Ethical Impact” of AI on law practice floridabar.org, effectively providing guidance to NY lawyers. It mirrors many concerns seen in formal opinions elsewhere.
Key points: The NYSBA report underscores competence and cautions against “techno-solutionism.” It notes that “a refusal to use technology that makes legal work more accurate and efficient may be considered a refusal to provide competent representation” nysba.org nysba.org – implying lawyers should stay current with helpful AI tools. At the same time, it warns attorneys not to blindly trust AI as a silver bullet. The report coins “techno-solutionism” as the overbelief that new tech (like gen AI) can solve all problems, reminding lawyers that human verification is still required nysba.org nysba.org. The infamous Avianca case is cited to illustrate the need to verify AI outputs and supervise the “nonlawyer” tool (AI) under Rule 5.3 nysba.org. The report addresses the duty of confidentiality & privacy in depth: Lawyers must ensure client information isn’t inadvertently shared or used to train public AI models nysba.org nysba.org. It suggests that if AI tools store or learn from inputs, that raises confidentiality concerns nysba.org. Client consent or use of secure “closed” AI systems may be needed to protect privileged data. The report also covers supervision (Rule 5.3) – attorneys should supervise AI use similarly to how they supervise human assistants nysba.org. It touches on bias and fairness, noting generative AI trained on biased data could perpetuate discrimination, which lawyers must guard against lawnext.com. Interestingly, the NYSBA guidance also links AI use to reasonable fees: it suggests effective use of AI can factor into whether a fee is reasonable jdsupra.com jdsupra.com (e.g. inefficiently refusing to use available AI might waste client money, whereas using AI and still charging full hours might be unreasonable). In sum, New York’s bar leaders affirm that ethical duties of competence, confidentiality, and supervision fully apply to AI. They encourage using AI’s benefits to improve service, but caution against its risks and urge ongoing attorney oversight floridabar.org floridabar.org.
(Sources: NYSBA Task Force Report nysba.org nysba.org.)
New York City Bar Association – Formal Opinion 2024-5 (August 2024)
The New York City Bar Association Committee on Professional Ethics issued Formal Ethics Opinion 2024-5 on August 7, 2024 nydailyrecord.com nydailyrecord.com. This opinion, in a user-friendly chart format, provides practical guidelines for NYC lawyers on generative AI. The Committee explicitly aimed to give “guardrails and not hard-and-fast restrictions” in this evolving area nydailyrecord.com.
Key points: Confidentiality: The NYC Bar draws a distinction between “closed” AI systems (e.g. an in-house or vendor tool that does not share data externally) and public AI services like ChatGPT. If using an AI that stores or shares inputs outside the firm, client informed consent is required before inputting any confidential information nydailyrecord.com. Even with closed/internal AI, lawyers must maintain internal confidentiality protections. The opinion warns lawyers to review AI Terms of Use regularly to ensure the provider isn’t using or exposing client data without consent nydailyrecord.com. Competence: Echoing others, NYC advises that lawyers “understand to a reasonable degree how the technology works, its limitations, and the applicable Terms of Use” before using generative AI nydailyrecord.com. Attorneys should avoid delegating their professional judgment to AI; any AI output is just a starting point or draft nydailyrecord.com. Lawyers must ensure outputs are accurate and tailored to the client’s needs – essentially, verify everything and edit AI-generated material so that it truly serves the client’s interests nydailyrecord.com. Supervision: Firms should implement policies and training for lawyers and staff on acceptable AI use nydailyrecord.com. The Committee notes that client intake chatbots (if used on a firm’s website, for example) require special oversight to avoid inadvertently forming attorney-client relationships or giving legal advice without proper vetting nydailyrecord.com. In other words, a chatbot interacting with the public should be carefully monitored by lawyers to ensure it doesn’t mislead users about its nature or create unintended obligations nydailyrecord.com. The NYC Bar’s guidance aligns with California’s in format and substance, reinforcing that the core duties of confidentiality, competence (tech proficiency), and supervision all apply when lawyers use generative AI tools nydailyrecord.com nydailyrecord.com.
(Source: NYC Bar Formal Op. 2024-5nydailyrecord.com nydailyrecord.com.)
New Jersey Supreme Court – Preliminary Guidelines (January 2024)
In New Jersey, the state’s highest court itself weighed in. On January 24, 2024, the New Jersey Supreme Court’s Committee on AI and the Courts issued “Preliminary Guidelines on the Use of AI by New Jersey Lawyers,” which were published as a Notice to the Bar njcourts.gov njcourts.gov. These guidelines, effective immediately, aim to help NJ lawyers comply with existing Rules of Professional Conduct when using generative AI njcourts.gov.
Key points: The Court made clear that AI does not change lawyers’ fundamental duties. Any use of AI “must be employed with the same commitment to diligence, confidentiality, honesty, and client advocacy as traditional methods of practice.” njcourts.gov In other words, tech advances do not dilute responsibilities. The NJ guidelines highlight accuracy and truthfulness: lawyers have an ethical duty to ensure their work is accurate, so they must always check AI-generated content for “hallucinations” or errors before relying on it jdsupra.com. Submitting false or fake information generated by AI would violate rules against misrepresentations to the court. The guidelines reiterate candor to tribunals – attorneys must not present AI-produced output containing fabricated cases or facts (the Mata/Avianca situation is alluded to)jdsupra.com. Regarding communication and client consent, NJ took a measured approach: There is “no per se requirement to inform a client” about every AI use, unless not telling the client would prevent the client from making informed decisions about the representation jdsupra.com. For example, if AI is used in a trivial manner (typo correction, formatting), disclosure isn’t required; but if it’s used in substantive tasks that affect the case, lawyers should consider informing the client, especially if there’s heightened risk. Confidentiality: Lawyers must ensure any AI tool is secure to avoid inadvertent disclosures of client info jdsupra.com. This echoes the duty to use “reasonable efforts” to safeguard confidential data (RPC 1.6). No misconduct: The Court reminds that all rules on attorney misconduct (dishonesty, fraud, bias, etc.) apply in AI usage jdsupra.com. For instance, using AI in a way that produces discriminatory outcomes or that frustrates justice would breach Rule 8.4. Supervision: Law firms must supervise how their lawyers and staff use AI jdsupra.com – establishing internal policies to ensure ethical use. Overall, New Jersey’s top court signaled that it embraces innovation (noting AI’s potential benefits) but insists lawyers “balance the benefits of innovation while safeguarding against misuse.” njcourts.gov
(Sources: NJ Supreme Court Guidelines jdsupra.com jdsupra.com.)
Pennsylvania & Philadelphia Bars – Joint Opinion 2024-200 (June 2024)
The Pennsylvania Bar Association (PBA) and Philadelphia Bar Association jointly issued Formal Opinion 2024-200 in mid-2024 lawnext.com lawnext.com. This collaborative opinion (“Joint Formal Op. 2024-200”) provides ethical guidance for Pennsylvania lawyers using generative AI. It repeatedly emphasizes that the same rules apply to AI as to any technology lawnext.com.
Key points: The joint opinion places heavy emphasis on competence (Rule 1.1). It famously states “Lawyers must be proficient in using technological tools to the same extent they are in traditional methods” lawnext.com. In other words, attorneys should treat AI as part of the competence duty – understanding e-discovery software, legal research databases, and now generative AI, is part of being a competent lawyer lawnext.com. The opinion acknowledges generative AI’s unique risk: it can hallucinate (generate false citations or facts) lawnext.com. Thus, due diligence is required – lawyers must verify all AI outputs, especially legal research results and citations lawnext.com lawnext.com. The opinion bluntly warns that if you ask AI for cases and “then file them in court without even bothering to read or Shepardize them, that is stupid.” lawnext.com (The opinion uses more polite language, but this captures the spirit.) It highlights bias as well: AI may carry implicit biases from training data, so lawyers should be alert to any discriminatory or skewed content in AI output lawnext.com. The Pennsylvania/Philly opinion also advises lawyers to communicate with clients about AI use. Specifically, lawyers should be transparent and “provide clear, transparent explanations” of how AI is being used in the case lawnext.com lawnext.com. In some situations, obtaining client consent before using certain AI tools is recommended lawnext.com lawnext.com – e.g., if the tool will handle confidential information or significantly shape the legal work. The opinion lays out “12 Points of Responsibility” for using gen AI lawnext.com lawnext.com, which include many of the above: ensure truthfulness and accuracy of AI-derived content, double-check citations, maintain confidentiality (ensure AI vendors keep data secure) lawnext.com, check for conflicts (make sure use of AI doesn’t introduce any conflict of interest) lawnext.com, and transparency with clients, courts, and colleagues about AI use and its limitations lawnext.com. It also addresses proper billing practices: lawyers shouldn’t overcharge when AI boosts efficiency lawnext.com. If AI saves time, the lawyer should not bill as if they did it manually – they may bill for the actual time or consider value-based fees, but padding hours violates the rule on reasonable fees lawnext.com. Overall, the Pennsylvania and Philly bars take the stance that embracing AI is fine — even beneficial — as long as lawyers “remain fully accountable for the results,” use AI carefully, and don’t neglect any ethical duty in the process lawnext.com lawnext.com.
(Sources: Joint PBA/Phila. Opinion 2024-200 summarized by Ambrogi lawnext.com lawnext.com.)
Kentucky – Ethics Opinion KBA E-457 (March 2024)
The Kentucky Bar Association issued Ethics Opinion KBA E-457, “The Ethical Use of Artificial Intelligence in the Practice of Law,” on March 15, 2024 cdn.ymaws.com. This formal opinion (finalized after a comment period in mid-2024) provides a nuanced roadmap for Kentucky lawyers. It not only answers basic questions but also offers broader insight, reflecting the work of a KBA Task Force on AI techlawcrossroads.com.
Key points: Competence: Like other jurisdictions, Kentucky affirms that keeping abreast of technology (including AI) is a mandatory aspect of competence techlawcrossroads.com techlawcrossroads.com. Kentucky’s Rule 1.1 Comment 6 (equivalent to ABA Comment 8) says lawyers “should keep abreast of … the benefits and risks associated with relevant technology.” The opinion stresses this is not optional: “It’s not a ‘should’; it’s a must.” techlawcrossroads.com Lawyers cannot ethically ignore AI’s existence or potential in law practice techlawcrossroads.com techlawcrossroads.com (implying that failing to understand how AI might improve service could itself be a lapse in competence). Disclosure to clients: Kentucky takes a practical stance that there is “no duty to disclose to the client the ‘rote’ use of AI generated research,” absent special circumstances techlawcrossroads.com. If an attorney is just using AI as a tool (like one might use Westlaw or a spell-checker), they generally need not inform the client. However, there are important exceptions – if the client has specifically limited use of AI, or if use of AI presents significant risk or would require client consent under the rules, then disclosure is needed techlawcrossroads.com. Lawyers should discuss risks and benefits of AI with clients if client consent is required for its use (for example, if AI will process confidential data, informed consent may be wise) techlawcrossroads.com. Fees: KBA E-457 is very direct about fees and AI. If AI significantly reduces the time spent on a matter, the lawyer may need to reduce their fees accordingly techlawcrossroads.com. A lawyer cannot charge a client as if a task took 5 hours if AI allowed it to be done in 1 hour – that would make the fee unreasonable. The opinion also says a lawyer can only charge a client for the expense of using AI (e.g., the cost of a paid AI service) if the client agrees to that fee in writing techlawcrossroads.com. Otherwise, passing along AI tool costs may be impermissible. In short, AI’s efficiencies should benefit clients, not become a hidden profit center. Confidentiality: Lawyers have a “continuing duty to safeguard client information if they use AI,” and must comply with all applicable court rules on AI use techlawcrossroads.com. This means vetting AI providers’ security and ensuring no confidential data is exposed. Kentucky echoes that attorneys must understand the terms and operation of any third-party AI system they use techlawcrossroads.com. They should know how the AI service stores and uses data. Court rules compliance: Notably, the opinion reminds lawyers to follow any court-imposed rules about AI (for instance, if a court requires disclosure of AI-drafted filings, the lawyer must do so) cdn.ymaws.com. Firm policies and training: KBA E-457 advises law firms to create informed policies on AI use and to supervise those they manage in following these policies techlawcrossroads.com. In summary, Kentucky’s opinion encourages lawyers to embrace AI’s potential but to do so carefully: stay competent with the technology, be transparent when needed, adjust fees fairly, protect confidentiality, and always maintain ultimate responsibility for the work. It concludes that Kentucky lawyers “cannot run from or ignore AI.” techlawcrossroads.com
(Source: KBA E-457 (2024) via TechLaw Crossroads summary techlawcrossroads.com techlawcrossroads.com.)
North Carolina – Formal Ethics Opinion 2024-1 (November 2024)
The North Carolina State Bar adopted 2024 Formal Ethics Opinion 1, “Use of Artificial Intelligence in a Law Practice,” on November 1, 2024 ncbar.gov ncbar.gov. This opinion squarely addresses whether and how NC lawyers can use AI tools consistent with their ethical duties.
Key points: The NC State Bar gives a cautious “Yes” to using AI, under specific conditions: “Yes, provided the lawyer uses any AI program, tool, or resource competently, securely to protect client confidentiality, and with proper supervision when relying on the AI’s work product.” ncbar.gov. That single sentence captures the three pillars of NC’s guidance: competence, confidentiality, and supervision. NC acknowledges that nothing in the Rules explicitly prohibits AI use ncbar.gov, so it comes down to applying existing rules. Competence: Lawyers must understand the technology sufficiently to use it effectively and safely ncbar.gov. Rule 1.1 and its Comment in NC (which, like the ABA, includes tech competence) require lawyers to know what they don’t know – if a lawyer isn’t competent with an AI tool, they must get up to speed or refrain. NC emphasizes that using AI is often the lawyer’s own decision but it must be made prudently, considering factors like the tool’s reliability and cost-benefit for the client ncbar.gov ncbar.gov. Confidentiality & Security: Rule 1.6(c) in North Carolina obligates lawyers to take reasonable efforts to prevent unauthorized disclosure of client info. So, before using any cloud-based or third-party AI, the lawyer must ensure it is “sufficiently secure and compatible with the lawyer’s confidentiality obligations.” ncbar.gov ncbar.gov. The opinion suggests attorneys evaluate providers like they would any vendor handling client data – e.g., examine terms of service, data storage policies, etc., similar to prior NC guidance on cloud computing ncbar.gov ncbar.gov. If the AI is “self-learning” (using inputs to improve itself), lawyers should be wary that client data might later resurface to others ncbar.gov. NC stops short of mandating client consent for AI use, but it implies that if an AI tool can’t be used consistent with confidentiality, then either don’t use it or get client permission. Supervision and Independent Judgment: NC treats AI output like work by a nonlawyer assistant. Under Rule 5.3, lawyers must supervise the use of AI tools and “exercise independent professional judgment in determining how (or if) to use the product of an AI tool” for a client ncbar.gov ncbar.gov. This means a lawyer cannot blindly accept an AI’s result – they must review and verify it before relying on it. If an AI drafts a contract or brief, the lawyer is responsible for editing and ensuring it’s correct and appropriate. NC explicitly analogizes AI to both other software and to nonlawyer staff: AI is “between” a software tool and a nonlawyer assistant in how we think of it ncbar.gov. Thus, the lawyer must both know how to use the software and supervise its output as if it were a junior employee’s work. Bottom line: NC FO 2024-1 concludes that a lawyer may use AI in practice – for tasks like document review, legal research, drafting, etc. – as long as the lawyer remains fully responsible for the outcome ncbar.gov ncbar.gov. The opinion purposefully doesn’t dictate when AI is appropriate or not, recognizing the technology is evolving ncbar.gov. But it clearly states that if a lawyer decides to employ AI, they are “fully responsible” for its use and must ensure it is competent use, confidential use, and supervised use ncbar.gov ncbar.gov.
(Source: NC 2024 FEO-1ncbar.gov ncbar.gov.)
Texas – Proposed Opinion 2024-6 (Draft, November 2024)
The State Bar of Texas Professional Ethics Committee has circulated a Proposed Ethics Opinion No. 2024-6 (posted for public comment on Nov. 19, 2024) regarding lawyers’ use of generative AI texasbar.com. (As of this writing, it is a draft opinion awaiting final adoption.) This Texas draft provides a “high-level overview” of ethical issues raised by AI, requested by a Bar task force on AI texasbar.com.
Key points (draft): The proposed Texas opinion covers familiar ground. It notes the duty of competence (Rule 1.01) extends to understanding relevant technology texasbar.com. Texas specifically cites its prior ethics opinions on cloud computing and metadata, which required lawyers to have a “reasonable and current understanding” of those technologies texasbar.com texasbar.com. By analogy, any Texas lawyer using generative AI “must have a reasonable and current understanding of the technology” and its capabilities and limits texasbar.com. In practical terms, this means lawyers should educate themselves on how tools like ChatGPT actually work (e.g. that they predict text rather than retrieve vetted sources) and what their known pitfalls are texasbar.com. The draft opinion spends time describing Mata v. Avianca to illustrate the dangers of not understanding AI’s lack of a reliable legal database texasbar.com texasbar.com. On confidentiality (Rule 1.05 in Texas), the opinion again builds on prior guidance: lawyers must safeguard client information when using any third-party service texasbar.com texasbar.com. It suggests precautions similar to those for cloud storage: “acquire a general understanding of how the technology works; review (and potentially renegotiate) the Terms of Service; [ensure] the provider will keep data confidential; and stay vigilant about data security.” texasbar.com. (These examples are drawn from Texas Ethics Op. 680 on cloud computing, which the AI opinion heavily references.) If an AI tool cannot be used in a way that protects confidential info, the lawyer should not use it for those purposes. The Texas draft also flags duty to avoid frivolous submissions (Rule 3.01) and duty of candor to tribunal (Rule 3.03) as directly relevant texasbar.com. Using AI doesn’t excuse a lawyer from these obligations – citing fake cases or making false statements is no less an ethical violation because an AI generated them. Lawyers must thoroughly vet AI-generated legal research and content to ensure it’s grounded in real law and facts texasbar.com texasbar.com. The opinion essentially says: if you choose to use AI, you must double-check its work just as you would a junior lawyer’s memo or a nonlawyer assistant’s draft. Supervision (Rules 5.01, 5.03): Supervising partners should have firm-wide measures so that any use of AI by their team is ethical texasbar.com texasbar.com. This could mean creating policies on approved AI tools and requiring verification of AI outputs. In summary, the Texas proposed opinion doesn’t ban generative AI; it provides a “snapshot” of issues and reinforces that core duties of competence, confidentiality, candor, and supervision must guide any use of AI in practice texasbar.com texasbar.com. (The committee acknowledges the AI landscape is rapidly changing, so they focused on broad principles rather than specifics that might soon be outdated texasbar.com.) Once finalized, Texas’s opinion will likely align with the consensus: lawyers can harness AI’s benefits if they remain careful and accountable.
(Source: Texas Proposed Op. 2024-6 texasbar.com texasbar.com.)
Virginia State Bar – AI Guidance Update (August 2024)
In 2024 the Virginia State Bar released a short set of guidelines on generative AI as an update on its website (around August 2024) nydailyrecord.com. This concise guidance stands out for its practicality and flexibility. Rather than an extensive opinion, Virginia issued overarching advice that can adapt as AI technology evolves nydailyrecord.com.
Key points: Virginia first emphasizes that lawyers’ basic ethical responsibilities “have not changed” due to AI, and that generative AI presents issues “fundamentally similar” to those with other technology or with supervising people nydailyrecord.com. This frames the guidance: existing rules suffice. On confidentiality, the Bar advises lawyers to vet how AI providers handle data just as they would with any vendor nydailyrecord.com nydailyrecord.com . Legal-specific AI products (designed for lawyers, with better data security) may offer more protection, but even then attorneys “must make reasonable efforts to assess” the security and “whether and under what circumstances” confidential info could be exposed nydailyrecord.com. In other words, even if using an AI tool marketed as secure for lawyers, you should confirm that it truly keeps your client’s data confidential (no sharing or training on it without consent) nydailyrecord.com nydailyrecord.com. Virginia notably aligns with most jurisdictions (and diverges from a stricter ABA stance) regarding client consent: “there is no per se requirement to inform a client about the use of generative AI in their matter” nydailyrecord.com. Unless something about the AI use would necessitate client disclosure (e.g., an agreement with the client, or an unusual risk like using a very public AI for sensitive info), lawyers generally need not obtain consent for routine AI use nydailyrecord.com. This is consistent with the idea that using AI can be like using any software tool behind the scenes. Next, supervision and verification: The bar stresses that lawyers must review all AI outputs as they would work done by a junior attorney or nonlawyer assistant nydailyrecord.com nydailyrecord.com. Specifically, “verify that any citations are accurate (and real)” and generally ensure the AI’s work product is correct nydailyrecord.com. This duty extends to supervising others in the firm – if a paralegal or associate uses AI, the responsible lawyer must ensure they are doing so properly nydailyrecord.com. On fees and billing, Virginia takes a clear stance: a lawyer may not bill a client for time not actually spent due to AI efficiency gains nydailyrecord.com. “A lawyer may not charge an hourly fee in excess of the time actually spent … and may not bill for time saved by using generative AI.” nydailyrecord.com If AI cuts a research task from 5 hours to 1, you can’t still charge 5 hours. The Bar suggests considering alternative fee arrangements to account for AI’s value, instead of hourly billing windfalls nydailyrecord.com. As for passing along AI tool costs: the Bar says you can’t charge the client for your AI subscription or usage unless it’s a reasonable charge and permitted by the fee agreement nydailyrecord.com. Finally, Virginia reminds lawyers to stay aware of any court rules about AI. Some courts (even outside Virginia) have begun requiring attorneys to certify that filings were checked for AI-generated falsehoods, or even prohibiting AI-drafted documents absent verification. Virginia’s guidance highlights that lawyers must comply with any such disclosure or anti-AI rules in whatever jurisdiction they are in nydailyrecord.com nydailyrecord.com. Overall, the Virginia State Bar’s message is: use common sense and existing rules. Be transparent when needed, protect confidentiality, supervise and double-check AI outputs, bill fairly, and follow any new court requirements nydailyrecord.com nydailyrecord.com. This short-form guidance was praised for being “streamlined” and adaptable as AI tools continue to change nydailyrecord.com.
(Source: Virginia State Bar AI Guidance via N.Y. Daily Record nydailyrecord.com nydailyrecord.com.)
District of Columbia Bar – Ethics Opinion 388 (September 2024)
The D.C. Bar issued Ethics Opinion 388: “Attorneys’ Use of Generative AI in Client Matters” in 2024 (the second half of the year) kaiserlaw.com. This opinion closely analyzes the ethical implications of lawyers using gen AI, using the well-known Mata v. Avianca incident as a teaching example kaiserlaw.com kaiserlaw.com . It then organizes guidance under specific D.C. Rules of Professional Conduct.
Key points: The opinion breaks its analysis into categories of duties kaiserlaw.com kaiserlaw.com:
- Competence (Rule 1.1): D.C. reiterates that tech competence is part of a lawyer’s duty. Attorneys must “keep abreast of … practice [changes], including the benefits and risks of relevant technology.” kaiserlaw.com Before using AI, lawyers should understand how it works, what it does, and its potential dangers kaiserlaw.com kaiserlaw.com. The opinion vividly quotes a description of AI as “an omniscient, eager-to-please intern who sometimes lies to you.” kaiserlaw.com kaiserlaw.com In practical terms, D.C. lawyers must know that AI output can be very convincing but incorrect. The Mata/Avianca saga – where a lawyer unknowingly relied on a tool that “sometimes lies” – underscores the need for knowledge and caution dcbar.org dcbar.org.
- Confidentiality (Rule 1.6): D.C.’s Rule 1.6(f) specifically requires lawyers to prevent unauthorized use of client info by third-party service providers kaiserlaw.com kaiserlaw.com. This applies to AI providers. Lawyers are instructed to ask themselves: “Will information I provide [to the AI] be visible to the AI provider or others? Will my input affect future answers for other users (potentially revealing my data)?” kaiserlaw.com kaiserlaw.com. If using an AI tool that sends data to an external server, the lawyer must ensure that data is protected. D.C. likely would advise using privacy-protective settings or choosing tools that allow opt-outs of data sharing, or obtaining client consent if needed. Essentially, treat AI like any outside vendor under Rule 5.3/1.6: do due diligence to ensure confidentiality is preserved kaiserlaw.com kaiserlaw.com.
- Supervision (Rules 5.1 & 5.3): A lawyer must supervise both other lawyers and nonlawyers in the firm regarding AI use kaiserlaw.com kaiserlaw.com. This may entail firm policies: e.g., vetting which AI tools are approved and training staff to verify AI output for accuracy kaiserlaw.com kaiserlaw.com. If a subordinate attorney or paralegal uses AI, the supervising attorney should reasonably ensure they are doing so in compliance with all ethical duties (and correcting any mistakes). The opinion views AI as an extension of one’s team – requiring oversight.
- Candor to Tribunal & Fairness (Rules 3.3 and 3.4): Simply put, a lawyer cannot make false statements to a court or submit false evidence kaiserlaw.com kaiserlaw.com. D.C. notes the existing comment to Rule 3.3 already forbids knowingly misrepresenting legal authority. Opinion 388 makes clear this includes presenting AI-fabricated cases or quotes as if they were real kaiserlaw.com kaiserlaw.com. Even if the lawyer didn’t intend to lie, relying on AI without checking and thereby filing fake citations could violate the duty of candor (at least negligently, if not knowingly). The lesson: no courtroom use of AI content without verification. Also, under fairness to opposing party (3.4), one must not use AI to manipulate evidence or discovery unfairly.
- Fees (Rule 1.5): The D.C. Bar echoed the consensus on billing: if you charge hourly, you “may never charge a client for time not expended.” kaiserlaw.com Increased efficiency through AI cannot be used as an opportunity to overcharge. They cite a 1996 D.C. opinion which said that a lawyer who is more efficient than expected (perhaps through technology or expertise) can’t then bill extra hours that weren’t worked kaiserlaw.com kaiserlaw.com. The same principle applies now: time saved by AI is the client’s benefit, not the lawyer’s windfall. So if AI drafts a contract in 1 hour whereas manual drafting would take 5, the lawyer cannot bill 5 hours – only the 1 hour actually spent (or use a flat fee structure that the client agrees on, but not lie about hours).
- Client Files (Rule 1.16(d)): Interestingly, D.C. Opinion 388 touches on whether AI interactions should be retained as part of the client file upon termination kaiserlaw.com kaiserlaw.com. D.C. law requires returning the “entire file” to a client, including internal notes, unless they are purely administrative. The opinion suggests lawyers should consider saving important AI prompts or outputs used in the representation as part of the file material that may need to be provided to the client kaiserlaw.com kaiserlaw.com. For example, if an attorney used an AI tool to generate a research memo or a draft letter that was then edited and sent to a client, the initial AI-generated text might be analogous to a draft or research note. This is a new facet many haven’t considered: how to handle AI-generated work product in terms of file retention.
In conclusion, D.C.’s Ethics Opinion 388 aligns with other jurisdictions while adding thoughtful details. It “recognizes AI may eventually greatly benefit the legal industry,” but in the meantime insists that lawyers “must be vigilant” kaiserlaw.com. The overarching theme is captured in the NPR quote: treat AI like an intern who needs close supervision kaiserlaw.com. Do not assume the AI is correct; double-check everything, maintain confidentiality, and use the tool wisely and transparently. D.C. lawyers were effectively told that generative AI is permissible to use, but only in a manner that fully preserves all ethical obligations as enumerated above kaiserlaw.com.
(Sources: D.C. Ethics Op. 388 via Kaiser summary kaiserlaw.com kaiserlaw.com.)
Specialty Bar and Licensing Bodies
U.S. Patent and Trademark Office (USPTO) – Practice Guidance (2023–2024)
Beyond state bars, at least one lawyer licensing body has addressed AI: the USPTO, which regulates patent and trademark attorneys. In 2023 and 2024, the USPTO issued guidance on the use of AI by practitioners in proceedings before the Office. On April 10, 2024, the USPTO published a notice (and a Federal Register guidance document) concerning “the use of AI tools by parties and practitioners” before the USPTO uspto.gov uspto.gov. This followed an earlier internal guidance on Feb 6, 2024 for USPTO administrative tribunals uspto.gov.
Key points: The USPTO made clear that existing duties in its rules (37 C.F.R. and USPTO ethics rules) “apply regardless of how a submission is generated.” uspto.gov In other words, whether a patent application or brief is written by a human or with AI assistance, the attorney is fully responsible for compliance with all requirements. The guidance reminds practitioners of pertinent rules and “helps inform … the risks associated with AI” while giving suggestions to mitigate them uspto.gov. For example, patent attorneys have a duty of candor and truthfulness in dealings with the Office; using AI that produces inaccurate statements could violate that duty if not corrected. USPTO Director Kathi Vidal emphasized “the integrity of our proceedings” must be protected and that the USPTO encourages “safe and responsible use of AI” to benefit efficiency uspto.gov. But critically, lawyers and agents must ensure AI is not misused or left unchecked. The USPTO guidance likely points to rules akin to Fed. R. Civ. P. 11: patent practitioners must make a reasonable inquiry that submissions (claims, arguments, prior art citations, etc.) are not frivolous or false, even if AI was used as a tool. It also addresses confidentiality and data security concerns: patent lawyers often handle sensitive technical data, so if they use AI for drafting or searching prior art, they must ensure they aren’t inadvertently disclosing invention details. The USPTO suggested mitigation steps such as: carefully choosing AI tools (perhaps ones that run locally or have strong confidentiality promises), verifying outputs (especially legal conclusions or prior art relevance), and staying updated as laws/regulations evolve in this area uspto.gov uspto.gov. In sum, the USPTO’s stance is aligned with the bar associations’: AI can expand access and efficiency, but practitioners must use it responsibly. They explicitly note that AI’s use “does not change” the attorney’s obligations to avoid delay, avoid unnecessary cost, and uphold the quality of submissions uspto.gov. The patent bar was cautioned by the USPTO, much as litigators were by the courts, that any mistakes made by AI will be treated as the practitioner’s mistakes. The Office will continue to “listen to stakeholders” and may update policies as needed uspto.gov, but for now practitioners should follow this guidance and existing rules.
(Source: USPTO Director’s announcement uspto.gov uspto.gov.)
Other Specialty Groups
Other specialty lawyer groups and bar associations have engaged in policy discussions about AI (for example, the American Immigration Lawyers Association and various sections of the ABA have offered CLE courses or informal tips on AI use). While these may not be formal ethics opinions, they echo the themes above: maintain client confidentiality, verify AI output, and remember that technology doesn’t diminish a lawyer’s own duties.
In summary, across national, state, and local bodies in the U.S., a clear consensus has emerged: Lawyers may use generative AI tools in their practice, but they must do so cautiously and in full compliance with their ethical obligations. Key recommendations include obtaining client consent if confidential data will be involved jdsupra.com nydailyrecord.com, understanding the technology’s limits (no blind trust in AI) nysba.org kaiserlaw.com, thoroughly vetting and supervising AI outputs ncbar.gov kaiserlaw.com, and ensuring that AI-driven efficiency benefits the client (through accurate work and fair fees) lawnext.com kaiserlaw.com. All the formal opinions – from the ABA to state bars like California, Florida, New York, Pennsylvania, Kentucky, North Carolina, Virginia, D.C., and others – converge on the message that the lawyer is ultimately responsible for everything their generative AI tool does or produces. Generative AI can assist with research, drafting, and more, but it remains “a tool that assists but does not replace legal expertise and analysis.” lawnext.com. As the Pennsylvania opinion neatly put it, in more colloquial terms: don’t be stupid – a lawyer cannot abdicate common sense and professional judgment to an AI lawnext.com. By following these ethics guidelines, lawyers can harness AI’s benefits (greater efficiency and capability) while upholding their duties to clients, courts, and the justice system.
Sources: Formal ethics opinions and guidance from the ABA and numerous bar associations, including ABA Formal Op. 512 jenkinslaw.org, State Bar of California guidance jdsupra.com, Florida Bar Op. 24-1 jdsupra.com, New Jersey Supreme Court AI Guidelines jdsupra.com, New York City Bar Op. 2024-5 nydailyrecord.com, Pennsylvania Bar & Philadelphia Bar Joint Op. lawnext.com, Kentucky Bar Op. E-457 techlawcrossroads.com, North Carolina Formal Op. 2024-1 ncbar.gov, D.C. Bar Op. 388 kaiserlaw.com, and USPTO practitioner guidance uspto.gov. Each of these sources provides detailed discussion of ethical concerns and best practices for using generative AI in law.