Categories
AI + Access to Justice Current Projects

Good/Bad AI Legal Help at Trust and Safety Conference

This week, Margaret Hagan presented at theĀ Trust and Safety Research Conference, that brings together academics, tech professionals, regulators, nonprofits, and philanthropies to work on making the Internet a more safe, user-friendly place.

Margaret presented interim results of the Labā€™s expert and user studies of AIā€™s performance at answering everyday legal questions, like around evictions and other landlord-tenant problems.

Some of the topics for discussion in the audience and panel on the Future of Search:

  • How can regulators, frontline domain experts (like legal aid lawyers and court professionals), and tech companies better work together to spot harmful content, set tailored policies, and ensure better outcomes for users?
  • Should tech companiesā€™ and governmentsā€™ policies towards ā€œwhat is the best way/amountā€ information for a user differ in different domains? Like perhaps for legal help queries, is it better to encourage more straightforward, simple, directive & authoritative info ā€” or more complex, detailed information that encourages more curiosity and exploration?
  • How do we more proactively spot the harms and risks that might come from new & novel tech systems, that might be quite different than previous search engines or other tech systems?
  • How can we hold tech companiesĀ accountable to make more accurate tech systems, without chilling them out of a certain domain (like legal or health), where they donā€™t want to provide any substantial information for fear of liability?
Categories
AI + Access to Justice Class Blog Current Projects Design Research

Interviewing Legal Experts on the Quality of AI Answers

This month, our team commenced interviews with landlord-tenant subject matter experts, including court help staff, legal aid attorneys, and hotline operators. These experts are comparing and rating various AI responses to commonly asked landlord-tenant questions that individuals may get when they go online to find help.

Learned Hands Battle Mode

Our team has developed a new ‘Battle Mode’ of our rating/classification platform Learned Hands. In a Battle Mode game on Learned Hands, experts compare two distinct AI answers to the same user’s query and determine which one is superior. Additionally, we have the experts speak aloud as they are playing, asking that they articulate their reasoning. This allows us to gain insights into why a particular response is deemed good or bad, helpful or harmful.

Our group will be publishing a report that evaluates the performance of various AI models in answering everyday landlord-tenant questions. Our goal is to establish a standardized approach for auditing and benchmarking AI’s evolving ability to address people’s legal inquiries. This standardized approach will be applicable to major AI platforms, as well as local chatbots and tools developed by individual groups and startups. By doing so, we hope to refine our methods for conducting audits and benchmarks, ensuring that we can accurately assess AI’s capabilities in answering people’s legal questions.

Instead of speculating about potential pitfalls, we aim to hear directly from on-the-ground experts about how these AI answers might help or harm a tenant who has gone onto the Internet to problem-solve. This means regular, qualitative sessions with housing attorneys and service providers, to have them closely review what AI is telling people when asked for information on a landlord-tenant problem. These experts have real-world experience in how people use (or donā€™t) the information they get online, from friends, or from other experts — and how it plays out for their benefit or their detriment. 

We also believe that regular review by experts can help us spot concerning trends as early as possible. AI answers might change in the coming months & years. We want to keep an eye on the evolving trends in how large tech companiesā€™ AI platforms respond to peopleā€™s legal help problem queries, and have front-line experts flag where there might be a big harm or benefit that has policy consequences.

Stay tuned for the results of our expert-led rating games and feedback sessions!

If you are a legal expert in landlord-tenant law, please sign up to be one of our expert interviewees below.

https://airtable.com/embed/appMxYCJsZZuScuTN/pago0ZNPguYKo46X8/form

Categories
AI + Access to Justice Current Projects

Autumn 24 AI for Legal Help

Our team is excited to announce the new, 2024-25 version of our ongoing class, AI for Legal Help. This school year, we’re moving from background user and expert research towards AI R&D and pilot development.

Can AI increase access to justice, by helping people resolve their legal problems in more accessible, equitable, and effective ways? What are the risks that AI poses for people seeking legal guidance, that technical and policy guardrails should mitigate?

In this course,Ā students will design and develop new demonstration AI projects and pilot plans, combining human-centered design, tech & data work, and law & policy knowledge.Ā 

Students will work on interdisciplinary teams, each partnered with frontline legal aid and court groups interested in using AI to improve their public services. Student teams will help their partners scope specific AI projects, spot and mitigate risks, train a model, test its performance, and think through a plan to safely pilot the AI.Ā 

By the end of the class, students and their partners willĀ co-design new tech pilots to help people dealing with legal problems like evictions, reentry from the criminal justice system, debt collection, and more.

Students will get experience in human-centered AI development, and critical thinking about if and how technology projects can be used in helping the public with a high-stakes legal problem. Along with their AI pilot, teams will establish important guidelines to ensure that new AI projects are centered on the needs of people, and developed with a careful eye towards ethical and legal principles.

Join our policy lab team to do R&D to define the future of AI for legal help.Apply for the class at the SLS Policy Lab link here.

Categories
AI + Access to Justice Current Projects

AI+A2J Research x Practice Seminar

The Legal Design Lab is proud to announce a new monthly online, public seminar on AI & Access to Justice: Research x Practice.

At this seminar, we’ll be bringing together leading academic researchers with practitioners and policymakers, who are all working on how to make the justice system more people-centered, innovative, and accessible through AI. Each seminar will feature a presentation from either an academic or practitioner who is working in this area & has been gathering data on what they’re learning. The presentations could be academic studies about user needs or the performance of technology, or less academic program evaluations or case studies from the field.

We look forward to building a community where researchers and practitioners in the justice space can make connections, build new collaborations, and advance the field of access to justice.

Sign up for the AI&A2J Research x Practice seminar, every first Friday of the month on Zoom.

Categories
AI + Access to Justice Current Projects

AI & Legal Help at Codex FutureLaw

At the April 2024 Stanford Codex FutureLaw Conference, our team at Legal Design Lab both presented the research findings about users’ and subject matter experts’ approaches to AI for legal help, and to lead a half-day interdisciplinary workshop on what future directions are possible in this space.

Many of the audience members in both sessions were technologists interested in the legal space, who are not necessarily familiar with the problems and opportunities for legal aid groups, courts, and people with civil legal problems. Our goal was to help them understand the “access to justice” space and spot opportunities to which their development & research work could relate.

Some of the ideas that emerged in our hands-on workshop included the following possible AI + A2J innovations:

AI to Scan Scary Legal Documents

Several groups identified that AI could help a person, who has received an intimidating legal document — a notice, a rap sheet, an immigration letter, a summons and complaint, a judge’s order, a discovery request, etc. AI could let them take a picture of the document, synthesize the information, present it back with a summary of what it’s about, what important action items are, and how to get started on dealing with it.

It could make this document interactive through FAQs, service referrals, or a chatbot that lets a person understand and respond to it. It could help people take action on these important but off-putting documents, rather than avoid them.

Using AI for Better Gatekeeping of Eviction Notices & Lawsuits

One group proposed that a future AI-powered system could screen possible eviction notices or lawsuit filings, to check if the landlord or property manager has fulfilled all obligations and m

  • Landlords must upload notices.
  • AI tools review the notice: is it valid? have they done all they can to comply with legal and policy requirements? is there any chance to promote cooperative dispute resolution at this early stage?
  • If the AI lives at the court clerk level, it might help court staff better detect errors, deficiencies, and other problems that better help them allocate limited human review.

AI to empower people without lawyers to respond to a lawsuit

In addition, AI could help the respondent (tenant) prepare their side, helping them to present evidence, prep court documents, understand court hearing expectations, and draft letters or forms to send.

Future AI tools could help them understand their case, make decisions, and get work product created with little burden.

With a topic like child support modification, AI could help a person negotiate a resolution with the other party, or do a trial run to see how a possible negotiation might go. It could also change their tone, to take a highly emotional negotiation request and transform it to be more likely to get a positive, cooperative reply from the other party.

AI to make Legal Help Info More Accessible

Another group proposed that AI could be integrated into legal aid, law library, and court help centers to:

  • Better create and maintain inter-organization referrals, so there are warm handoffs and not confusing roundabouts when people seek help
  • Clearer, better maintained, more organized websites for a jurisdiction, with the best quality resources curated and staged for easy navigation
  • Multi-modal presentations, to make information available in different visual presentations and languages
  • Providing more information in speech-to-text format, conversational chats, and across different dialects. This was especially highlighted in immigration legal services.

AI to upskill students & pro bono clinics

Several groups talked about AI for training and providing expert guidance to staff, law students, and pro bono volunteers to improve their capacity to serve members of the public.

AI tools could be used in simulations to better educate people in a new legal practice area, and also supplement their knowledge when providing services. Expert practitioners can supply knowledge to the tools, that can then be used by novice practitioners so that they can provide higher-quality services more efficiently in pro bono or law student clinics.

AI could also be used in community centers or other places where community justice workers operate, to get higher quality legal help to people who don’t have access to lawyers or who do not want to use lawyers.

AI to improve legal aid lawyers’ capacity

Several groups proposed AI that could be used behind-the-scenes by expert legal aid or court help lawyers. They could use AI to automate, draft, or speed up the work that they’re already doing. This could include:

  • Improving intake, screening, routing, and summaries of possible incoming cases
  • Drafting first versions of briefs, forms, affidavits, requests, motions, and other legal writing
  • Documenting their entire workflow & finding where AI can fit in.

Cross-Cutting action items for AI+ A2J

Across the many conversations, some common tasks emerged that cross different stakeholders and topics.

Reliable AI Benchmarks:

We as a justice community need to establish solid benchmarks to test AI effectiveness. We can use these benchmarks to focus on relevant metrics.

In addition, we need to regularly report on and track AI performance at different A2J tasks.

This can help us create feedback loops for continuous improvement.

Data Handling and Feedback:

The community needs reliable strategies and rules for how to do AI work that respects obligations for confidentiality and privacy.

Can there be more synthetic datasets that still represent what’s happening in legal aid and court practice, so they don’t need to share actual client information to train models?

Can there be better Personally Identifiable Information (PII) redaction for data sharing?

Who can offer guidance on what kinds of data practices are ethical and responsible?

Low-Code AI Systems:

The justice community is never going to have large tech, data, or AI working groups within their legal aid or court organization. They are going to need low-code solutions that will let them deploy AI systems, fine-tune them, and maintain them without a huge technical requirement.

Overall, the presentation, Q&A, and workshop all pointed to enthusiasm for responsible innovation in the AI+A2J space. Tech developers, legal experts, and strategists are excited about the opportunity to improve access to justice through AI-driven solutions, and enhance efficiency and effectiveness in legal aid. With more brainstormed ideas for solutions in this space, now it is time to move towards R&D incubation that can help us understand what is feasible and valuable in practice.

Categories
AI + Access to Justice Current Projects

Law360 Article on Legal Design Lab’s AI-Justice Work

In early May 2024, the Stanford Legal Design Lab’s work was profiled in the Law360 publication.

The article summarizes the Legal Design Lab’s work, partnerships & human-centered design approach to tackle legal challenges & develop new technologies.

The article covers our recent user & legal help provider research, our initial phase of groundwork research, and our new phase of R&D to see if we can develop legal AI solutions in partnership with frontline providers.

Finally, the article touches on our research on quality metrics & our upcoming AI platform audit.

Categories
AI + Access to Justice Current Projects

3 Shifts for AI in the Justice System: LSC 50th Anniversary presentation

In mid-April, Margaret Hagan presented on the Lab’s research and development efforts around AI and access to justice at the Legal Services Corporation 50th anniversary forum. This large gathering of legal aid executive directors, national justice leaders, members of Congress, philanthropists, and corporate leaders celebrated the work of LSC & profiled future directions of legal services.

Margaret was on a panel along with legal aid leader Sateesh Nori, Suffolk Law School Dean Andy Perlman, and former LSC president James Sandman.

She presented 3 big takeaways for the audience, about how to approach if and how AI should be used to close on the justice gap — especially to move beyond gut reactions & anecdotes that tend towards too much optimism or skepticism. Based on the lab’s research and design activities she proposed 3 big shifts for civil justice leaders towards generative AI.

Shift 1: Towards Techno-Realism

This shift away from hardline camps about too much optimism or pessimism about AI’s potential futures can lead us to more empirical, detailed work. Where are the specific tasks where AI can be helpful? Can we demonstrate with lab studies and controlled pilots exactly if AI can perform better than humans at these specific tasks — with equal or higher quality, and efficiency? This move towards applied research can lead towards more responsible innovation, rather than rushing towards AI applications too quickly or chilling the innovation space pre-emptively.

Shift 2: From Reactive to Proactive leadership

The second shift is how lawyers and justice professionals approach the world of AI. Will they be reactive to what technologists put out to the public, trying to create the right mix of norms, lawsuits, and regulations that can try to push AI towards being safe enough, and also quality enough for legal use cases?

Instead, they can be proactive. They can be running R&D cohorts to see what AI is good at, what risks and harms emerge in these test applications, and then work with AI companies and regulators to better encourage the AI strengths and mitigate its risks. This means joining together with technologists (especially those at universities and benefit corporations) to do hands-on, exploratory demonstration project development to better inform investments, regulation, and other policy-making on AI for justice use cases.

Shift 3: Local Pilots to Coordinated Network

The final shift is about how innovators work. Legal aid groups or court staff could launch AI pilots on their own, building out a new application or bot for their local jurisdiction, and then share it at upcoming conferences to let others know about it. Or, from the beginning, they could be crafting their technical system, UX design, vendor relationships, data management, and safety evaluations in concert with others around the country who are working on similar efforts. Even if the ultimate application is run and managed locally, much of the infrastructure can be shared in national cohorts. These national cohorts can also help gather data, experiences, risk/harm incidents, and other important information that can help guide task forces, attorneys general, tech companies, and others setting the policies for legal help AI in the future.

See more of the presentation in the slides below.

Categories
AI + Access to Justice Current Projects

User interviews on AI & Access to Justice

As we continue to run interviews with people from across the country about their possible use of AI for legal help tasks, we wanted to share out what we’re learning about people’s thoughts about AI.Please see the full interactive Data Dashboard of interview results here.

Below, find images of the data dashboard. Follow the link above to interact more with the data.

We will also maintain a central page of user research findings on AI & Access to Justice here.

Below, find the results of our interviews as of early 2024.

We asked people to self-assess their ability to solve legal problems and to use the Internet to solve life problems.

We also asked them how often they use the Internet.

Finally, we asked them about their past use of generative AI tools like ChatGPT, Bing/CoPilot, or Bard/Gemini.

Trust & Value of AI to Participants

We asked people at the beginning of the interview how much they would trust what AI would tell them for a legal problem.

We asked them the same question after they tried out an AI tool for a fictional legal problem of getting an eviction notice from their landlord.

We also asked them how helpful the AI was in dealing with the fictional problem, and how likely they would be to use this in the future for similar problems.

Preferences for possible AI tool features

We presented a variety of possible interface & policy changes, that could be made to an AI platform.

We asked the participants to rank the utility of these different possible changes.

Categories
AI + Access to Justice Current Projects

AI as the next frontier of making justice accessible

Last week, Margaret had the privilege of presenting on the lab’s work on AI and Innovation at the American Academy of Arts and Sciences in Cambridge, Massachusetts.

As a part of the larger conference of Making Justice Accessible, her work was featured on the panel about new solutions to improve the civil justice system through technology.

She discussed how the Lab’s current research and development work around AI has grown out of a larger question about helping people who are increasingly going online to find legal help.

The AI work is an outgrowth of previous work on

  • improving legal help websites,
  • auditing and improving search engines’ treatment of legal queries,
  • working on new ways to present information in more visual and plain language ways, and
  • building cohorts of providers across regions to have more standardized and discoverable help online.

This panel also included presentations on other, linked efforts to use technology to improve civil justice, including Georgetown’s Judicial Innovation Fellowship program, Stanford’s Filing Fairness Project, and Suffolk LIT Lab’s document assembly and efiling efforts.

Categories
AI + Access to Justice Current Projects

AI & Justice Workers

At the Arizona State University/American Bar Foundation conference on the Future of Justice Work, Margaret Hagan spoke on if and how generative AI might be part of new service and business models to serve people with legal problems.

Many in the audience are already developing new staffing & service models, that combine traditional lawyer-provided services with help provided by community-justice workers.

In the conference’s final session, the panelists discussed how technology — particularly the new generative AI models — might also figure into new initiatives to better reach & serve people struggling with eviction orders, bad living conditions, domestic violence, debt collection, custody problems, and more.

Margaret presented a brief summary of the Legal Design Lab’s work on user research into what people need & want from legal AI, how they currently use AI tools, what justice professionals are brainstorming as possible AI-powered justice work, and metrics and benchmark protocols to evaluate the AI.

Possible AI-powered justice work zones

This clear listing of the tasks that go into “legal work” and “legal services” that we need to do in AI – -is similar to what people working on new community justice worker models are also doing.

Breaking legal work apart into these tasks can help us think systematically about new, stratified models of delivering services.

  • Inside of these zones of work, what are the specific tasks that exist (that lawyers and legal org staff currently do, or should be doing)?
  • Who can and should be best doing this task?
    • Only Seasoned Lawyers: Which of the tasks can only be done by expert lawyers, with JDs, bar admissions, and multiple years practicing in a given problem area & on this task?
    • Medium-to-Novice Lawyers: Which of the tasks can be done by medium-to-novice lawyers, with JDs, bar admission, but little to no practice experience in this problem area or on this task (like pro bono volunteers, or new lawyers)?
    • Seasoned Justice Workers: Which of the tasks can be done by people who are paralegals, advocates, volunteers, social workers, and other community justice workers who have multiple years working this problem area & doing this kind of task?
    • Medium-to-Novice Justice Workers: Which of the tasks can be done by community justice workers who are new to this problem area & task?
    • Tech + Lawyer/Justice Worker: Which of these tasks can be done by technology (initial draft/work product) then reviewed by a lawyer or justice worker?
    • Technology: Which of these tasks can be done technology without human review?

Ideally, our justice community will have more of these discussions about the future of providing services with smart, safe models that can improve capacity & people’s outcomes.