Hayley Stillwell & Sean Harrington

In this episode, Ted sits down with Hayley Stillwell, Associate Professor, and Sean Harrington, Director of Technology and Innovation at the University of Oklahoma College of Law, to discuss how artificial intelligence is reshaping the way we understand judgment, bias, and the future of legal education. From their groundbreaking mock jury study, “Michael Scott Is Not a Juror,” to the challenges of fine-tuning AI for legal use, Hayley and Sean share their expertise in law, data science, and innovation. With insights on the reasonable juror standard, bias in AI models, and the evolving role of lawyers in the age of automation, this conversation offers law professionals a thought-provoking look at the intersection of human judgment and machine intelligence.

In this episode, Hayley and Sean share insights on how to:

  • Explore how AI models interpret and replicate human legal reasoning
  • Understand the limitations and biases that arise in AI-driven mock jury studies
  • Identify why the “reasonable juror” standard may need rethinking in the age of AI
  • Examine the future of legal education and the skills needed for AI-era lawyers
  • Recognize the growing importance of specialized AI tools in legal practice

Key takeaways:

  • The “reasonable juror” standard may not align with real-world juror behavior or AI simulations
  • Bias in AI models can produce significant risks in legal decision-making
  • Legal education must adapt to prepare students for an AI-driven future
  • Specialized, fine-tuned AI models are more effective than general-purpose systems in legal contexts
  • Collaboration between lawyers, technologists, and educators is key to building trustworthy AI in law

About the guests Hayley Stillwell and Sean Harrington

Hayley Stillwell is an Associate Professor of Law at the University of Oklahoma College of Law, where she teaches and writes in the areas of evidence and criminal procedure. Her research explores criminal justice reform and the intersection of artificial intelligence and the law, examining how emerging technologies influence legal processes and decision-making.

At a high level, these platforms are just not equipped to replicate human judgment the way it actually comes out with real humans and their varying backgrounds.
Hayley Stillwell

Sean Harrington is the Director of Technology and Innovation at the University of Oklahoma College of Law, where he leverages his expertise in law, information science, and data analytics to drive the school’s digital transformation. His work focuses on integrating AI into legal education, developing research tools, and preparing future lawyers to navigate an increasingly technology-driven profession.

There just wasn’t good academic research in this area [of predicting what jurors will do]. So that’s another reason to integrate AI into this thing.
Sean Harrington

Connect with Hayley Stillwell and Sean Harrington:

Subscribe for Updates

Newsletter

Newsletter

Machine Generated Episode Transcript

1 00:00:00,151 --> 00:00:02,083 Sean Haley, how are you today? 2 00:00:02,083 --> 00:00:03,453 Good to see you. 3 00:00:03,453 --> 00:00:05,655 I really appreciate you guys joining me, man. 4 00:00:05,655 --> 00:00:12,260 I enjoyed your Michael Scott paper quite a bit and we're going to talk about it today. 5 00:00:12,260 --> 00:00:14,602 But before we do, let's get you guys introduced. 6 00:00:14,602 --> 00:00:20,985 So you're both academics at University of, is it Oklahoma? 7 00:00:21,766 --> 00:00:22,786 University of Oklahoma. 8 00:00:22,786 --> 00:00:24,027 In my notes, have Arkansas. 9 00:00:24,027 --> 00:00:25,988 That's wrong. 10 00:00:26,068 --> 00:00:29,450 Why don't, Hailey, we start with you, introduce yourself and then we'll... 11 00:00:29,450 --> 00:00:30,811 Hand it over to Sean. 12 00:00:31,017 --> 00:00:31,387 Sure. 13 00:00:31,387 --> 00:00:33,128 So I'm Haley Stilwell. 14 00:00:33,128 --> 00:00:37,950 I am an associate professor of law at the University of Oklahoma College of Law. 15 00:00:37,950 --> 00:00:40,211 I teach evidence and criminal law. 16 00:00:40,211 --> 00:00:45,813 I write in those fields as well, but I also have a particular interest in exploring AI in my scholarship. 17 00:00:45,813 --> 00:00:49,015 And that's kind of where Sean and I's relationship began. 18 00:00:49,015 --> 00:00:49,255 Yeah. 19 00:00:49,255 --> 00:00:50,025 And I'm Sean Harrington. 20 00:00:50,025 --> 00:00:57,178 I'm the director of technology innovation here at OU Law, which is kind of a nebulous title, you know, at different institutions that means different things. 21 00:00:57,178 --> 00:00:59,711 Here what it means is I teach AI in the practice of law. 22 00:00:59,711 --> 00:01:06,231 And I also run what's called their digital initiative, which gets tech training to all the students, essentially. 23 00:01:06,326 --> 00:01:09,869 And I met someone from your law school. 24 00:01:09,869 --> 00:01:10,440 Yeah. 25 00:01:10,440 --> 00:01:11,930 Was it Kenton? 26 00:01:12,792 --> 00:01:13,572 Yeah. 27 00:01:13,572 --> 00:01:19,107 Um, we were, he happened to be attending a session that I was helping, helping facilitate and sat at my table. 28 00:01:19,107 --> 00:01:23,601 So it was pretty cool when I saw your paper and was like, yeah. 29 00:01:23,601 --> 00:01:26,404 And connected the dots and, here we are today. 30 00:01:27,626 --> 00:01:28,459 So 31 00:01:28,459 --> 00:01:29,190 to your podcast. 32 00:01:29,190 --> 00:01:31,432 And so he was like, my gosh, you got to go on that. 33 00:01:31,432 --> 00:01:33,373 I was like, Okay, yeah, I'm in. 34 00:01:33,434 --> 00:01:34,014 Good. 35 00:01:34,014 --> 00:01:34,374 Good. 36 00:01:34,374 --> 00:01:34,594 Yeah. 37 00:01:34,594 --> 00:01:44,574 We started off, you know, we started off primarily in the knowledge management and innovation realm at large law firms as our audience. 38 00:01:44,574 --> 00:01:46,174 And it's really evolved from that. 39 00:01:46,174 --> 00:01:49,954 I mean, we have, I've got people from the marketing function. 40 00:01:49,954 --> 00:01:52,985 I've got attorneys, I've got managing partners, I have academics. 41 00:01:52,985 --> 00:01:59,721 So, you know, the innovation topic is very timely in the legal world. 42 00:01:59,721 --> 00:02:00,477 right. 43 00:02:00,477 --> 00:02:03,718 Like the timing was, the timing was very fortunate for us. 44 00:02:03,718 --> 00:02:07,940 So I started right around the time chat GPT was released. 45 00:02:07,940 --> 00:02:08,740 Yeah. 46 00:02:08,740 --> 00:02:10,260 So, I mean, just pure luck. 47 00:02:10,260 --> 00:02:13,181 Like I didn't see that and go, we need a podcast. 48 00:02:13,181 --> 00:02:18,754 was just, we had been selling into not the world of knowledge management for many, many years. 49 00:02:18,754 --> 00:02:26,164 And then we saw this interesting dynamic where all saw these innovation roles start to pop up and 50 00:02:26,164 --> 00:02:26,958 Right. 51 00:02:27,244 --> 00:02:31,688 started having conversations with, what's the difference between knowledge management and innovation? 52 00:02:31,688 --> 00:02:35,612 And those conversations were very enlightening. 53 00:02:35,612 --> 00:02:37,372 And I thought, you know what? 54 00:02:37,473 --> 00:02:40,115 I should record these and make a podcast. 55 00:02:40,115 --> 00:02:54,757 So I recorded three episodes, didn't tell anyone about it, posted them, you know, just on the usual platforms, Spotify, Apple podcasts, like three weeks before, Ilta Ilta con where 56 00:02:54,757 --> 00:02:55,338 we met. 57 00:02:55,338 --> 00:03:00,578 Kenton and I had like 10 people come up to me in the booth and go, Hey, I your podcast. 58 00:03:00,578 --> 00:03:01,018 It's great. 59 00:03:01,018 --> 00:03:03,398 I was like, how the hell did you find it? 60 00:03:03,398 --> 00:03:12,958 And I guess people just, I don't know if they keep Google alerts on certain keywords, but, um, that was almost a hundred episodes. 61 00:03:12,958 --> 00:03:15,458 We're going to have our hundredth episode here in a few weeks. 62 00:03:15,458 --> 00:03:18,115 And, uh, yeah, it's been, it's been a lot of fun. 63 00:03:18,115 --> 00:03:18,495 awesome. 64 00:03:18,495 --> 00:03:27,835 mean, the thing is you go to something like Ilta and you just have, you know, 100 of those types of conversations and you run into Damien real or whatever, you know, and you have a 65 00:03:27,835 --> 00:03:32,075 conversation like, man, I wish I had somebody to save this, you know, other people could repurpose this. 66 00:03:32,075 --> 00:03:37,175 I want to have so I have this little, uh, have you seen the limitless AI pendant? 67 00:03:37,334 --> 00:03:42,692 No, but I've used, I've actually used the, think it's called plod, which I think it's similar. 68 00:03:43,437 --> 00:03:43,849 Yeah. 69 00:03:43,849 --> 00:03:46,802 And so he just sticks right here and then you have that knowledge base for forever. 70 00:03:46,802 --> 00:03:50,816 And I'm like, man, anytime I want to go to one of those conferences, I'm to be wearing that thing. 71 00:03:50,816 --> 00:03:59,666 Yeah, you know, the whole privacy thing around that, like, do you have to tell everybody like, hey, I'm recording that, you know, I was, yeah. 72 00:03:59,666 --> 00:04:05,546 of don't have like the social etiquette for it because obviously, you know, people knew that you were recording them, they would speak to you very differently. 73 00:04:05,546 --> 00:04:13,146 But you want the other conversation because the utility of retaining that information and having it in a data set that you could manipulate is like so high. 74 00:04:13,146 --> 00:04:20,886 It's just like, we're gonna have to figure out something like a little green, you know, light on my lapel that says Sean's always recording or something. 75 00:04:20,886 --> 00:04:22,406 think I turn it off. 76 00:04:22,406 --> 00:04:25,146 Nobody's want Nobody's gonna want to talk to me anyways. 77 00:04:25,146 --> 00:04:26,582 That's only gonna make it worse. 78 00:04:26,582 --> 00:04:29,211 in the legal crowd, right? 79 00:04:29,211 --> 00:04:31,918 I mean, they're very privacy sensitive. 80 00:04:31,918 --> 00:04:37,265 I had a faculty member here that I was just explaining what the thing did. 81 00:04:37,265 --> 00:04:38,096 It was in its dock. 82 00:04:38,096 --> 00:04:39,149 I was like, it's not recording. 83 00:04:39,149 --> 00:04:40,551 I was explaining what it was. 84 00:04:40,551 --> 00:04:44,115 And she made up an excuse to get up and leave my office almost immediately. 85 00:04:44,383 --> 00:04:45,003 That's funny. 86 00:04:45,003 --> 00:04:45,423 Yeah. 87 00:04:45,423 --> 00:04:54,763 I mean, we, um, we use a platform called fireflies for our meetings and 100 % of our, of our client base is law, our law firms, large law firms. 88 00:04:55,183 --> 00:04:59,983 And, um, surprisingly we don't get a lot of pushback on that. 89 00:05:00,063 --> 00:05:09,803 And, you know, but I do think that people manage their candor when they see and, and rightly so I get it. 90 00:05:09,803 --> 00:05:11,319 Um, but 91 00:05:11,355 --> 00:05:14,071 Anyway, let's talk about your paper. 92 00:05:14,071 --> 00:05:15,706 So it really caught my eye. 93 00:05:15,706 --> 00:05:19,112 It was, and the reason is because I'm a big office fan. 94 00:05:19,484 --> 00:05:20,464 Love it. 95 00:05:20,464 --> 00:05:22,324 So the title was all Haley. 96 00:05:22,324 --> 00:05:24,264 She came up with that. 97 00:05:25,924 --> 00:05:27,264 Oh, yeah. 98 00:05:27,264 --> 00:05:27,535 Yeah. 99 00:05:27,535 --> 00:05:29,077 they're great. 100 00:05:29,077 --> 00:05:35,343 So if I remember correctly, the title was Michael Scott is not a juror. 101 00:05:36,244 --> 00:05:36,766 Okay. 102 00:05:36,766 --> 00:05:40,654 And, um, give us a little bit of background on the paper. 103 00:05:40,654 --> 00:05:45,203 Uh, I don't know whoever wants to start and then we can jump into that a little bit. 104 00:05:45,321 --> 00:05:46,491 Yeah, you started, it your idea. 105 00:05:46,491 --> 00:05:51,744 Yeah, so the origin story kind of comes from my just academic research in general. 106 00:05:51,744 --> 00:06:01,828 There's this standard that's often utilized em in courts called the reasonable juror standard where courts are basically trying to determine what a perfectly behaving 107 00:06:01,828 --> 00:06:07,210 impartial juror that follows all the court's directions, how they would think and understand evidence. 108 00:06:07,210 --> 00:06:14,387 And it comes into play a lot when em courts are assessing whether the admission or exclusion of evidence was 109 00:06:14,387 --> 00:06:16,939 in violation of a rule or the constitution. 110 00:06:16,939 --> 00:06:26,874 And so I was bothered by some of the court's reasonable jury determinations because it just didn't track at all with how I thought real people would think and understand about 111 00:06:26,874 --> 00:06:27,985 certain evidence. 112 00:06:27,985 --> 00:06:38,390 So that prompted me to want to create a mock jury study to compare what the court has determined reasonable jurors would think and compare it to what real people would think 113 00:06:38,390 --> 00:06:43,301 and kind of highlight where the court has got it wrong and in turn results in 114 00:06:43,301 --> 00:06:45,922 probable constitutional violations. 115 00:06:45,922 --> 00:06:57,508 And so with that project, what I wanted to then do is find a way to empirically ground the reasonable juror where possible to bring the reasonable juror and real jurors closer 116 00:06:57,508 --> 00:07:00,389 together when courts were making those determinations. 117 00:07:00,389 --> 00:07:06,682 And the easiest way to do that is with mock jury data like we did in this study, but it takes way too much time. 118 00:07:06,682 --> 00:07:08,867 It's way too expensive to do. 119 00:07:08,867 --> 00:07:20,567 And so when uh Sean and I were kind of talking about it, we got the idea, what if we could create a mock jury bot that could output mock jury information instantaneously to help 120 00:07:20,567 --> 00:07:27,852 inform courts when they're making these determinations to protect criminal defendants and all people that are appealing their cases. 121 00:07:27,852 --> 00:07:38,704 And so we thought it would be pretty easy, just use existing platforms to feed them, you know, what this, the demographic profile of a juror and spit out information and 122 00:07:38,704 --> 00:07:41,156 Hopefully it would track what real jurors would do. 123 00:07:41,156 --> 00:07:43,587 And that is just not even close to what happened. 124 00:07:43,587 --> 00:07:44,958 And that's where this paper came. 125 00:07:44,958 --> 00:07:45,308 Yeah. 126 00:07:45,308 --> 00:07:50,642 And like the other thing is, you know, we looked up a bunch of studies that were trying to do this and it's like sample size, 112 people. 127 00:07:50,642 --> 00:07:56,566 So there just wasn't a lot of research and uh actual surveys of humans done on this stuff. 128 00:07:56,566 --> 00:08:00,669 So, you know, you quickly realize that it's really expensive to survey people. 129 00:08:00,669 --> 00:08:05,773 And so there just wasn't like a good academic research in this area. 130 00:08:05,773 --> 00:08:06,649 So we're like, well, 131 00:08:06,649 --> 00:08:10,635 That's even another reason then to integrate AI into this thing. 132 00:08:10,801 --> 00:08:11,111 Yeah. 133 00:08:11,111 --> 00:08:20,087 And you know, I, we talked to, we get a lot of attorneys on the show, some practicing, some not the ones who practice. 134 00:08:20,087 --> 00:08:28,072 I always like to ask, you know, what are the valuable use cases with AI trial prep is near the top of the list. 135 00:08:28,072 --> 00:08:31,034 And the data doesn't show this. 136 00:08:31,034 --> 00:08:33,396 really it's kind of drafting due diligence. 137 00:08:33,396 --> 00:08:36,923 you see as kind of the most common, document review. 138 00:08:36,923 --> 00:08:38,559 Yeah, document review would seem like fun. 139 00:08:38,559 --> 00:08:41,859 Yeah, it seemed to be the most common use cases cited. 140 00:08:41,879 --> 00:08:49,650 But when I hear about in terms of value, we had a guy on the podcast, name's Stephen Embry. 141 00:08:49,650 --> 00:08:50,750 You familiar with? 142 00:08:50,750 --> 00:08:51,650 Yeah. 143 00:08:51,810 --> 00:08:58,650 So Steve is a former, he was a 30 year attorney at I believe Finnegan, which is a, they're a client of ours. 144 00:08:58,650 --> 00:09:00,930 They're an intellectual property firm. 145 00:09:01,270 --> 00:09:04,914 And he was telling me how 146 00:09:05,111 --> 00:09:19,716 either him or his colleagues would use it for trial preparation in basically having AI challenge the evidence in some way, like ask all of the questions that opposing counsel 147 00:09:19,716 --> 00:09:22,977 might present in a courtroom. 148 00:09:22,977 --> 00:09:30,249 he said, you know, he talked about how, how useful that was, which I found very intriguing and makes perfect sense. 149 00:09:30,249 --> 00:09:32,760 Like I'm a daily user of AI. 150 00:09:32,860 --> 00:09:33,877 super familiar with. 151 00:09:33,877 --> 00:09:34,647 drive here for you. 152 00:09:34,647 --> 00:09:36,118 said, pretend. 153 00:09:36,118 --> 00:09:37,298 I said, here's our paper. 154 00:09:37,298 --> 00:09:40,469 Pretend that you're Ted and ask me some stuff. 155 00:09:40,469 --> 00:09:45,541 I mean, it really lowers the temperature coming into the environment because you're like, okay, I'm feeling pretty confident. 156 00:09:45,541 --> 00:09:49,172 You know, I've kind of been challenged on things that where I wasn't super confident. 157 00:09:49,172 --> 00:09:50,132 I looked a few things up. 158 00:09:50,132 --> 00:09:51,983 So it's like, you feel ready to roll. 159 00:09:51,983 --> 00:09:55,230 So I could definitely see that having a lot of value for attorneys. 160 00:09:55,230 --> 00:09:58,441 Yeah, I mean, it's almost like a practice run that you get. 161 00:09:58,441 --> 00:10:03,943 So when I heard the study, the name was catchy, so that drew me in initially. 162 00:10:03,943 --> 00:10:11,946 But then when I read what you guys were actually doing, I was like, wow, this seems like a perfect use case for AI. 163 00:10:11,946 --> 00:10:18,468 And it turns out it wasn't that perfect using the frontier models. 164 00:10:18,468 --> 00:10:21,699 Tell us about what your findings were initially. 165 00:10:22,083 --> 00:10:28,305 Well, the first thing we tried to do was just let the frontier models try to create a jury. 166 00:10:28,305 --> 00:10:34,016 So we said, create for us a jury pool that is similar to what a federal jury pool would be. 167 00:10:34,016 --> 00:10:36,438 And that's where Michael Scott emerged. 168 00:10:36,438 --> 00:10:39,100 It was really hilarious. 169 00:10:39,100 --> 00:10:41,411 They would output the demographics of the jurors. 170 00:10:41,411 --> 00:10:48,433 So it was a white man in his mid-40s who is the manager of a mid-sized paper firm in Scranton, Pennsylvania. 171 00:10:48,433 --> 00:10:52,145 which you and I would obviously know is Michael Scott. 172 00:10:52,145 --> 00:10:57,867 Michael Scott is not a real person, let alone a real juror in the federal jury pool, right? 173 00:10:57,867 --> 00:11:06,069 We also had a lot of other interesting combinations of, there was a 90 year old woman who was a part-time botanist, part-time DJ. 174 00:11:06,069 --> 00:11:07,437 I love that one. 175 00:11:07,437 --> 00:11:10,311 We had an abolitionist podcaster. 176 00:11:10,311 --> 00:11:15,824 So it seemed like when these platforms were left to their own devices, they were generating 177 00:11:15,824 --> 00:11:25,106 jurors that were more for show kind of eye catching types of backgrounds that really didn't reflect what we needed for our purposes. 178 00:11:25,106 --> 00:11:29,717 What real people on a jury would actually look like demographically. 179 00:11:29,717 --> 00:11:30,478 Yeah. 180 00:11:30,478 --> 00:11:39,560 And you can tell that, know, there a kid in, you know, Washington is using them right now to study who's 12 years old and maybe using it for creative writing. 181 00:11:39,560 --> 00:11:44,079 So, you know, there's a big range of why people are using these tools and they have the dial. 182 00:11:44,079 --> 00:11:51,150 on certain types of representation, which could be very useful, obviously, in a creative writing context, but in ours, that was, you know, catastrophic, because it was wasn't 183 00:11:51,150 --> 00:11:52,451 representing reality. 184 00:11:52,896 --> 00:11:55,888 Well, and look, I love Michael Scott. 185 00:11:55,888 --> 00:11:59,200 He'd be the last person I'd ever choose to be on a jury. 186 00:11:59,200 --> 00:12:01,301 Not, not known for good judgment. 187 00:12:01,301 --> 00:12:04,543 Um, well, and that's interesting. 188 00:12:04,543 --> 00:12:19,501 So I have had a lot of debate myself and I've seen a lot of debate in cyberspace, LinkedIn and Reddit about the frontier models coming and essentially crushing the legal specific 189 00:12:19,501 --> 00:12:27,750 tools just based on just how quickly the trajectory that they're on in terms of capabilities and 190 00:12:27,750 --> 00:12:38,941 I think your findings will first tell us what you found as potential root cause as to why this experiment was not successful with the Frontier models. 191 00:12:38,941 --> 00:12:39,461 Yeah. 192 00:12:39,461 --> 00:12:45,143 So first of all, you know, they have lengthy system prompts, which people, you know, leak on GitHub all the time. 193 00:12:45,143 --> 00:12:50,608 And you can see kind of why they're training these models to do things like be a good study tool for a student. 194 00:12:50,608 --> 00:12:57,852 And then we also just saw that just rampant bias on certain topics that just was completely at odds with reality. 195 00:12:57,852 --> 00:13:07,117 And so when we were looking, you when we actually finally ran our actual mock juror surveys, what we discovered is that people are like fairly moderate on average. 196 00:13:07,117 --> 00:13:10,320 And these would be kind of extreme versions of people. 197 00:13:10,320 --> 00:13:17,727 And again, you know, I think it's just because they're, you know, kind of meant to do everything instead of replicate juror juror scenarios. 198 00:13:17,727 --> 00:13:18,787 What else? 199 00:13:19,108 --> 00:13:24,132 Well, I think to covering kind of the second part of our experiment is important to explain that as well. 200 00:13:24,132 --> 00:13:33,380 After we left the platforms to their own devices, that's where the mock jury data came in to where, OK, we were going to tell the platforms now. 201 00:13:33,518 --> 00:13:35,409 here are the demographics of the jurors. 202 00:13:35,409 --> 00:13:37,099 You're not gonna pick them anymore. 203 00:13:37,099 --> 00:13:38,349 No more Michael Scott's. 204 00:13:38,349 --> 00:13:38,699 So. 205 00:13:38,699 --> 00:13:44,451 part coded in all of the demographics with a lengthy system prompt ah using Lang chain if anyone cares. 206 00:13:44,451 --> 00:13:54,354 Yeah, so we use basically the demographics from the real humans and told the platforms generate those exact humans and predict how incriminating this piece of evidence was. 207 00:13:54,354 --> 00:13:58,395 So we had an apples to apples comparison, how accurate were these platforms? 208 00:13:58,395 --> 00:13:59,956 And they weren't very accurate. 209 00:13:59,956 --> 00:14:01,368 Not only were they not accurate, 210 00:14:01,368 --> 00:14:03,340 but they were biased in different ways. 211 00:14:03,340 --> 00:14:16,733 ChatGPT was actually not very biased, then Jim and I thought the evidence was a little more incriminating on average, and then Claude was a little underrating in terms of 212 00:14:16,733 --> 00:14:17,364 incrimination. 213 00:14:17,364 --> 00:14:23,951 That was consistent basically across every type of demographic, so we saw those consistent skews in every single one. 214 00:14:23,951 --> 00:14:26,012 you know, X type of person, was deterministic. 215 00:14:26,012 --> 00:14:32,196 So if you're X type of person, then you necessarily would believe why you would, you know, if you're a certain type person, maybe you don't like the police. 216 00:14:32,196 --> 00:14:37,419 And so you would always find them, you know, guilty, you know, those types of things just happened over and over and over again. 217 00:14:37,419 --> 00:14:41,162 And we knew from our real world human data that that's not how real people behave at all. 218 00:14:41,162 --> 00:14:43,964 You could be a type of person and think something completely different. 219 00:14:43,964 --> 00:14:48,727 And so that, you know, just the replication of stereotypes was rampant through them. 220 00:14:48,727 --> 00:14:51,515 And this was something that as we were doing it, 221 00:14:51,515 --> 00:14:56,108 as somebody who sits in these things all day effectively, I was like, yeah, they're definitely gonna do that. 222 00:14:56,169 --> 00:14:59,801 But that was the most interesting thing to Haley, understandably, I think. 223 00:14:59,801 --> 00:15:04,505 And then when I presented this topic at a couple of different conferences, that was the thing that everybody latched onto. 224 00:15:04,505 --> 00:15:09,979 Like really, these are just wildly over-representing stereotypes and replicating stereotypes. 225 00:15:09,979 --> 00:15:12,360 That's catastrophic for the law in many ways, obviously. 226 00:15:12,360 --> 00:15:14,973 So that was something where I was like, people are interested in this. 227 00:15:14,973 --> 00:15:15,677 And so we... 228 00:15:15,677 --> 00:15:22,277 Spent a long time in the paper talking about that because it was something that maybe doesn't occur to a lot of people like the system prompts. 229 00:15:22,337 --> 00:15:25,597 If you're not, maybe Ted, you know what a system prompt is. 230 00:15:25,597 --> 00:15:34,377 But I think the average chat GPT user doesn't realize before you enter anything into the system, there's a five and 10 and 20 page potentially system prompt that has all of these 231 00:15:34,377 --> 00:15:37,377 instructions, do this, don't do that type stuff. 232 00:15:37,517 --> 00:15:41,737 so those types of things, I think maybe don't occur to the average lawyer. 233 00:15:41,737 --> 00:15:45,417 And they're definitely happening on the legal tools as well. 234 00:15:45,841 --> 00:15:47,739 You know, Westlaw has a lengthy system prompt. 235 00:15:47,739 --> 00:15:51,655 Lexus says they all have lengthy system prompts that tell it what it can and cannot do. 236 00:15:51,655 --> 00:15:52,195 Interesting. 237 00:15:52,195 --> 00:16:03,144 So is this kind of the Swiss army knife scenario where, know, they've got a lot of tools, including a little, little scissors there, but if you ever had to cut a thick piece of 238 00:16:03,144 --> 00:16:05,325 canvas with it, it's just not a fit. 239 00:16:05,325 --> 00:16:07,997 So it's good at doing a lot of things. 240 00:16:07,997 --> 00:16:08,588 Okay. 241 00:16:08,588 --> 00:16:11,410 But not really anything specialized. 242 00:16:11,410 --> 00:16:13,671 Is that really the dynamic at play? 243 00:16:13,870 --> 00:16:17,612 Yeah, and that's how tell the students, know, lawyers are very worried about hallucinations. 244 00:16:17,612 --> 00:16:18,972 In large language models. 245 00:16:18,972 --> 00:16:25,715 I saw students hallucinations at this point in time are basically solved if you're using Vlex's Vincent AI for using Lexus AI fusing Westlaw. 246 00:16:25,715 --> 00:16:27,476 It has a hyperlink to the case. 247 00:16:27,476 --> 00:16:30,477 It's your duty as an attorney to click that and go read the case. 248 00:16:30,477 --> 00:16:31,698 That's why they pay you the big bucks. 249 00:16:31,698 --> 00:16:33,899 But if you're using chat GBT, that's not the case. 250 00:16:33,899 --> 00:16:38,059 And so I think, you know, it is, think the Swiss Army Knife is a really good call. 251 00:16:38,059 --> 00:16:41,099 You're using a Swiss Army Knife when you should be using a skillsaw or something. 252 00:16:41,099 --> 00:16:46,159 And that skillsaw is Lexus, know, Vlex, whatever it is, which is a law specific tool. 253 00:16:46,159 --> 00:16:54,859 And here, what our results really showed is that these platforms were not good at replicating the noise in the real human responses. 254 00:16:54,859 --> 00:17:00,419 You know, it really seemed like demographics ended up being deterministic of what 255 00:17:00,455 --> 00:17:08,084 how incriminating they thought each evidence was, but we saw in our actual real human data that there was a lot more variation among demographics. 256 00:17:08,084 --> 00:17:18,587 so really at a high level, these platforms are just not equipped to replicate human judgment the way it actually comes out with real humans and their varying backgrounds. 257 00:17:18,587 --> 00:17:18,867 Yeah. 258 00:17:18,867 --> 00:17:22,250 And you know, I mean, I've actually noticed some interesting dynamics. 259 00:17:22,250 --> 00:17:26,773 The safety and alignment protocols are implemented very differently. 260 00:17:26,773 --> 00:17:35,568 Um, you know, there's four models that I use primarily, which are in chat, GBT, Claude, Gemini and Grok. 261 00:17:35,568 --> 00:17:40,912 And really I use them in that order in terms of frequency. 262 00:17:40,912 --> 00:17:45,685 but I ran into a really interesting little exercise where 263 00:17:45,823 --> 00:17:50,395 I saw there was a graphic on LinkedIn and it was a caricature of two people that I know. 264 00:17:50,395 --> 00:17:52,436 Um, in fact, they're both podcasters. 265 00:17:52,436 --> 00:17:55,217 was Richard Trowman's and sac of Bramowitz. 266 00:17:55,217 --> 00:18:02,380 And they had this picture and I was like, the guy who's supposed to be Richard Trowman's looks just like this actor, but I can't think of who it is. 267 00:18:02,380 --> 00:18:12,164 So I took a quick screenshot and put it into, I think I did Claude first and was just like, what famous actor does this per does the person given the thumbs up look like? 268 00:18:12,204 --> 00:18:13,458 And it said, 269 00:18:13,458 --> 00:18:14,228 I'm sorry. 270 00:18:14,228 --> 00:18:17,593 I actually have the, let's see if I can find it real quick. 271 00:18:17,593 --> 00:18:19,854 what it said, cause it was really interesting. 272 00:18:19,854 --> 00:18:28,569 it basically refused to do it and it did so on the basis of it was trying to be, here we go. 273 00:18:28,569 --> 00:18:29,189 Let's see. 274 00:18:29,189 --> 00:18:35,262 It says, I see the image shows two people in what appears to be a promotional blah, blah. 275 00:18:35,262 --> 00:18:35,965 However, 276 00:18:35,965 --> 00:18:48,165 I can't identify who specific people in images look like or compare them to famous individuals as this could involve making assumptions about people's identity based on 277 00:18:48,165 --> 00:18:49,385 their appearance. 278 00:18:49,405 --> 00:18:52,545 And I was like, all right, that's interesting. 279 00:18:52,545 --> 00:18:54,125 Let me go to ChatGPT. 280 00:18:54,125 --> 00:18:56,525 It's a little goofy, but okay, whatever. 281 00:18:56,525 --> 00:19:03,085 They're probably trying to solve some sort of a bias. 282 00:19:03,085 --> 00:19:05,361 or privacy thing or like you said, bias. 283 00:19:05,361 --> 00:19:08,939 It doesn't want to say that I look like somebody stereotypically a thing, whatever. 284 00:19:08,939 --> 00:19:09,510 Yeah. 285 00:19:09,510 --> 00:19:10,852 So I went to ChatGBT. 286 00:19:10,852 --> 00:19:12,905 ChatGBT just flat out refused. 287 00:19:12,905 --> 00:19:15,188 It just said, I cannot help you with this request. 288 00:19:15,188 --> 00:19:16,891 Gemini did it. 289 00:19:16,891 --> 00:19:20,207 I still don't know who the actor was, um unfortunately. 290 00:19:20,207 --> 00:19:22,027 So that's surprising when we were running. 291 00:19:22,027 --> 00:19:23,529 we're talking about refusal rates. 292 00:19:23,529 --> 00:19:26,822 When we were running ours, Jim and I had the highest refusal rate, right? 293 00:19:26,822 --> 00:19:27,703 Jim and I had it. 294 00:19:27,703 --> 00:19:32,717 And even for the nerds in the audience, even through the API, it would say, sorry, I can't do that. 295 00:19:32,717 --> 00:19:34,479 I can't, I can't engage in legal advice. 296 00:19:34,479 --> 00:19:40,724 Now subsequently, you know, and this is all obviously moving, moving terrain under our feet at all times. 297 00:19:40,724 --> 00:19:42,766 So you can do it one day and it'll refuse to do it the next day. 298 00:19:42,766 --> 00:19:45,869 We'll do it because they've dialed something back or down. 299 00:19:45,869 --> 00:19:47,864 But that was um for us, it was 300 00:19:47,864 --> 00:19:50,426 Gemini and Claude were the two highest refusal rates. 301 00:19:50,426 --> 00:19:52,097 ChatGPT was a little bit better. 302 00:19:52,097 --> 00:19:57,561 And even through the API, I just like, I'm not going to engage in, you know, essentially legal advice. 303 00:19:57,561 --> 00:20:00,273 And you know, what you realize is there was a time there. 304 00:20:00,273 --> 00:20:09,449 And I think that that time has probably passed where, know, if the state of California can sue legal zoom for forms and that's considered unauthorized practice of law, these are 305 00:20:09,449 --> 00:20:12,802 pretty uncomfortable with, you know, anything kind of approximating legal advice. 306 00:20:12,802 --> 00:20:14,783 Now, now they give legal advice, no problem. 307 00:20:14,783 --> 00:20:16,290 They'll issue a bunch of caveats. 308 00:20:16,290 --> 00:20:18,775 but they will give you lots of legal advice. 309 00:20:18,775 --> 00:20:24,327 I don't know what movement behind the scenes has happened, but they all kind of collectively said, we can give some legal advice. 310 00:20:24,327 --> 00:20:25,100 That's what it seems like. 311 00:20:25,100 --> 00:20:26,220 You should try your picture out again. 312 00:20:26,220 --> 00:20:26,821 Yeah. 313 00:20:26,821 --> 00:20:27,633 now it'll do it. 314 00:20:27,633 --> 00:20:28,079 Yeah. 315 00:20:28,079 --> 00:20:28,990 that's a good point. 316 00:20:28,990 --> 00:20:32,063 This was, this was probably a couple of months ago. 317 00:20:32,063 --> 00:20:36,308 So, yeah, there's not a date on this, but Grok did it, of course, which you would expect. 318 00:20:36,308 --> 00:20:39,282 Grok, not a lot of guard rails with Grok. 319 00:20:39,282 --> 00:20:45,067 And, you know, did, did you guys try Grok at all in this experiment? 320 00:20:45,192 --> 00:20:47,442 No, so we just stuck with those three. 321 00:20:47,442 --> 00:20:52,057 I just for a dumb reason, I just didn't have a GROCK subscription at that point so I can get to the API. 322 00:20:52,057 --> 00:20:53,188 So I didn't, I do now. 323 00:20:53,188 --> 00:20:55,370 And there's some stuff where GROCK is really useful. 324 00:20:55,370 --> 00:20:56,640 know, first of all, it's very fast. 325 00:20:56,640 --> 00:20:57,831 just in the chat interface. 326 00:20:57,831 --> 00:21:00,394 And then number two, it performs really, really well. 327 00:21:00,394 --> 00:21:09,761 think, you know, Elon Musk has kind of been in the political news so much that they don't report on his model the same way, but that it's a really interesting and powerful model. 328 00:21:09,761 --> 00:21:11,822 And the other thing is like, if you need 329 00:21:11,860 --> 00:21:19,204 You know, if you're on your phone and you need kind of a survey of kind of what is going on in social media, it's really useful because as Twitter data. 330 00:21:19,947 --> 00:21:22,087 I've done stock stuff for my wife with it. 331 00:21:22,087 --> 00:21:27,113 yeah, I was curious if there were any different results with Grok. 332 00:21:27,113 --> 00:21:32,297 Well, what about, um okay, so you tried the Frontier models, weren't successful. 333 00:21:32,297 --> 00:21:34,789 What was the next step? 334 00:21:34,789 --> 00:21:37,561 It sounds like you might've tried some open source models. 335 00:21:37,703 --> 00:21:39,445 Yeah, so I went on and eat. 336 00:21:39,445 --> 00:21:47,115 So I went on the leaderboard and kind of tried to find something that we could run locally so we could save token costs, of course, and was trying to look at, you know, less 337 00:21:47,115 --> 00:21:49,097 censored models was the idea. 338 00:21:49,097 --> 00:21:59,036 And actually, a handful of the models, even one of the Mistral models that was a French company, their Frontier Lab, and even one of those wouldn't do kind of law, advice, 339 00:21:59,036 --> 00:22:02,495 practice of law, replicate human experimentation type things. 340 00:22:02,495 --> 00:22:08,587 I was able to find a less censored version of Mistral that would do kind of whatever I told it to do, which was very useful. 341 00:22:08,587 --> 00:22:10,228 It was very useful for what we're doing. 342 00:22:10,228 --> 00:22:21,501 And so then we decided to try to take all of our real human responses, turn that into a JSON training set, and then fine tune that less censored model as just kind of like, will 343 00:22:21,501 --> 00:22:22,881 this do anything? 344 00:22:22,881 --> 00:22:26,332 And it ended up being much better than any of the other models that we did. 345 00:22:26,332 --> 00:22:28,621 So that was pretty interesting. 346 00:22:28,621 --> 00:22:29,981 And it was cheap too. 347 00:22:29,981 --> 00:22:34,661 So I remember I used to fine tune models kind of early in the chat, GPC revolution. 348 00:22:34,661 --> 00:22:36,001 And it was expensive. 349 00:22:36,001 --> 00:22:41,261 I mean, it would be too even for a small model, like a 24 billion parameter model would be hundreds of dollars. 350 00:22:41,261 --> 00:22:46,121 When I did this one, it was, I think it was $14 or something. 351 00:22:47,941 --> 00:22:49,321 I use a service. 352 00:22:49,321 --> 00:22:50,741 Well, I just use tokens for mine. 353 00:22:50,741 --> 00:22:52,121 So I did it with two different packages. 354 00:22:52,121 --> 00:22:54,061 There's an open source package called oxalotl. 355 00:22:54,061 --> 00:22:58,436 And then there's a service called open pipe, which is no, I guess no code. 356 00:22:58,452 --> 00:23:04,338 As long as you have the JSON training set, you can just throw it in there and you know, do whatever model and it'll also host it for inference, which is pretty cool. 357 00:23:04,338 --> 00:23:08,421 But initially I just did it at home on my home computer with axolotl. 358 00:23:08,883 --> 00:23:10,964 Open source, fine tuning package. 359 00:23:11,387 --> 00:23:12,969 did you try llama? 360 00:23:13,102 --> 00:23:14,822 No, we so we didn't do the llama. 361 00:23:14,822 --> 00:23:18,602 Oh, so we tested the llama model, but it ended up being even worse. 362 00:23:18,602 --> 00:23:24,382 And so I couldn't even I couldn't even fine tune it enough to get it to be anywhere near human level. 363 00:23:24,382 --> 00:23:29,353 And so I was like, all right, we have to find something less censored than llama, all the meta models. 364 00:23:29,353 --> 00:23:37,413 And then after the fine tuning, were, what, was the, how was the testing after, after you fine tuned? 365 00:23:37,838 --> 00:23:42,518 So we essentially just did the same thing that we did to the real humans and scaled it. 366 00:23:42,518 --> 00:23:45,358 I think we did it up to, was it 5,000? 367 00:23:45,358 --> 00:23:45,678 Yeah. 368 00:23:45,678 --> 00:23:46,478 It was 5,000. 369 00:23:46,478 --> 00:23:50,018 So instead of, we only had 1200 actual juror responses. 370 00:23:50,018 --> 00:23:59,698 So we scaled it up to 5,000 real human responses where we would say, you know, go read this, first take on this, you know, this demographic or take on a random demographic from 371 00:23:59,698 --> 00:24:07,158 the range present in the real jurors, then read the thing, then respond to the thing as if you are from, you know, said demographic. 372 00:24:07,247 --> 00:24:08,747 And then we just scaled it. 373 00:24:08,747 --> 00:24:14,655 it's just like, just ran my computer for like an evening, basically, just let it grind through 5,000 iterations. 374 00:24:14,655 --> 00:24:18,178 And then we took that and measured it against the real human responses. 375 00:24:18,395 --> 00:24:20,192 and how did they compare? 376 00:24:20,192 --> 00:24:27,512 Much better, much better, which it was one of those things where I wasn't sure if just fine tuning was going to be enough. 377 00:24:27,512 --> 00:24:29,672 It felt like something that was more. 378 00:24:29,672 --> 00:24:36,372 You might have to have like some kind of a genetic workflow and you know, frankly, we can't afford to pre train a model of our own. 379 00:24:36,372 --> 00:24:39,732 So I, I wasn't sure if just fine tuning would be enough. 380 00:24:39,732 --> 00:24:48,959 And then the other problem is when you're fine tuning a model, even a small model, like 24 billion parameters, we had 1200 Jason. 381 00:24:48,959 --> 00:24:52,639 data points, know, JSON rows or data points to use. 382 00:24:52,719 --> 00:24:58,099 Most people do like 50,000 and we'll see benefits up to 100,000 fine tuning. 383 00:24:58,099 --> 00:25:04,259 So we just thought it was possibly just way too small of a training set to have any, you know, measurable impact on it. 384 00:25:04,259 --> 00:25:10,999 And the fact that at 1200 it did was like, Whoa, Whoa, you know, this is something that again, it was fairly cheap. 385 00:25:10,999 --> 00:25:11,899 It could do it for free. 386 00:25:11,899 --> 00:25:15,959 If you had the right hardware, like the Mac studio that we're running all this stuff on, we can do it on that. 387 00:25:16,093 --> 00:25:17,925 So it was really interesting to go. 388 00:25:17,925 --> 00:25:22,947 was like one of those breaks, know, one of the million aha moments I seem to have every month with this technology. 389 00:25:22,947 --> 00:25:24,993 It was like, whoa, we can actually do this. 390 00:25:24,993 --> 00:25:26,646 You know, it's like what it felt like. 391 00:25:26,646 --> 00:25:31,206 And when you say much better, like give me a sense of magnitude. 392 00:25:31,206 --> 00:25:32,926 Like how, how did it do? 393 00:25:32,926 --> 00:25:46,466 I don't know if this was a percentage that you use to measure alignment, but like give me a sense of, it doesn't have to be the exact number, but was it two, two X three X better 394 00:25:46,466 --> 00:25:48,206 50 % better. 395 00:25:48,655 --> 00:25:55,106 It was, geez, I don't want to state it, because I had it right here, but my computer locked. 396 00:25:55,106 --> 00:25:59,408 It was, I wrote it down just so I could specifically say this. 397 00:25:59,408 --> 00:26:02,394 I remember reading it was significant. 398 00:26:02,394 --> 00:26:12,013 95 % accuracy and geez, like essentially the margin of error was half, half of what it was with the model, the frontier models. 399 00:26:12,013 --> 00:26:17,839 So it was, you know, twice as good effectively as chat GPT trying to do this on its own. 400 00:26:17,839 --> 00:26:24,385 And so, and again, like, you know, there was a lot of technical things that I didn't do that I would do differently. 401 00:26:24,385 --> 00:26:27,188 Now I would use a different style of fine tuning if I was going to do it. 402 00:26:27,188 --> 00:26:28,330 We would have, you know, 403 00:26:28,330 --> 00:26:29,950 even scale with synthetic data. 404 00:26:29,950 --> 00:26:36,870 So that was something that I thought about, but she didn't want to, uh, cause a lot of people had just have a reflexive bad reaction to synthetic data. 405 00:26:36,870 --> 00:26:44,830 But had we taken that and then scaled it up to, know, 50,000, you know, take 12, take 1200 and then use that to scale up to 50,000. 406 00:26:44,830 --> 00:26:46,530 And then they use that to train a model. 407 00:26:46,530 --> 00:26:49,210 It probably would have been substantially better even then. 408 00:26:49,210 --> 00:26:56,710 So, uh, you know, just for a small investment and a small amount of training data, we got pretty big wins. 409 00:26:56,710 --> 00:26:58,077 And that was like the real like, 410 00:26:58,077 --> 00:26:59,601 breakthrough, I guess, of the paper. 411 00:26:59,601 --> 00:27:01,435 Especially as compared to the platform. 412 00:27:01,435 --> 00:27:07,962 Yeah, and definitely substantially better if you just use like the API to Gemini or Claude or Google or OpenAI. 413 00:27:07,962 --> 00:27:08,313 Yeah. 414 00:27:08,313 --> 00:27:10,835 And what conclusions can we draw from this? 415 00:27:10,835 --> 00:27:24,621 So back to my earlier comment about there's a, it is a constant point of discussion in the legal tech community about are these legal specific solutions going to get steamrolled by 416 00:27:24,621 --> 00:27:26,322 the frontier models? 417 00:27:26,362 --> 00:27:33,916 And your study makes me think maybe, maybe not, because if the frontier models are 418 00:27:33,916 --> 00:27:41,270 have controls, we'll call them, in place that make them a Swiss army knife where they're good for a lot of things. 419 00:27:41,270 --> 00:27:46,953 So they're a mile wide and an inch deep, as opposed to being an inch wide and a mile deep. 420 00:27:47,134 --> 00:27:57,720 It seems like, I don't know if it's a fair conclusion or not, but your study for niche legal use cases, the frontier models might not be the best solution. 421 00:27:57,720 --> 00:28:00,341 I don't know, Haley, what's your thought on that? 422 00:28:00,562 --> 00:28:03,143 think that's exactly right, at least as of now. 423 00:28:03,143 --> 00:28:14,755 I think that our project really proves the need to have these models specifically trained on human data, human judgment in the realm of evaluating evidence in these legal 424 00:28:14,755 --> 00:28:15,415 scenarios. 425 00:28:15,415 --> 00:28:23,829 I think it could be huge in terms of, as you mentioned earlier, trial prep, things that lots of trial boutiques would really latch onto and utilize. 426 00:28:23,829 --> 00:28:25,699 But right now they're just not there. 427 00:28:25,699 --> 00:28:29,310 And I think it is because of that Swiss army knife nature. 428 00:28:29,478 --> 00:28:33,250 But I do think our fine tuning model shows that it is possible. 429 00:28:33,250 --> 00:28:44,038 So with lots more data and lots more training, there could be an essentially mock jury bot at some point in the future that could be reliable for trial prep and strategy and a lot 430 00:28:44,038 --> 00:28:44,789 of other things. 431 00:28:44,789 --> 00:28:48,792 Yeah, I know, right now the big the frontier models obviously have the data problem. 432 00:28:48,792 --> 00:28:55,457 There's only four good places you can get, you know, US legal data, Vlex, now Clio, guess, Vlex still, Westlaw, Lexus, Bloomberg. 433 00:28:55,457 --> 00:28:57,096 Now those are the only data sets. 434 00:28:57,096 --> 00:29:04,192 and they haven't purchased one of those yet, which probably indicates, which, you when I hear people talk about it, it's that they just don't think it's a big enough market right 435 00:29:04,192 --> 00:29:04,401 now. 436 00:29:04,401 --> 00:29:06,254 They're going after literally everyone. 437 00:29:06,254 --> 00:29:14,400 But if at some point they do purchase one of those data sets and have that on the backend, then that's really gonna blow up the business model of like a Lexus or Westlaw or Vlex, 438 00:29:14,400 --> 00:29:22,352 unless they're offering something else in addition to it, some sort of agentic workflow, some sort of tool that plugs into your firm, which I think is. 439 00:29:22,352 --> 00:29:27,216 why probably the Clio Vlex acquisition is really interesting to me right now. 440 00:29:27,216 --> 00:29:31,170 But right now, the frontier models have a data problem and they're just not focused on legal. 441 00:29:31,170 --> 00:29:38,045 They're too big for legal, which is crazy to think about with the 1.5 billion or whatever that's going into legal investment this year. 442 00:29:38,045 --> 00:29:41,659 It's like, no, we're still considered small dollars to them, I think. 443 00:29:41,659 --> 00:29:46,170 Yeah, well, and honestly, and the industry is extremely fragmented. 444 00:29:46,170 --> 00:29:51,452 And if you add up the entire Amlaw 100 revenue, it's $140 billion. 445 00:29:51,452 --> 00:29:55,694 Like that's, that would be like a fortune 100 company. 446 00:29:55,694 --> 00:29:58,689 So that's the entire, that's the top 100. 447 00:29:58,835 --> 00:30:05,598 So it is, it is a relatively small industry, when compared to others, but you know, 448 00:30:05,778 --> 00:30:09,040 There was an AI company that did try to buy one of the information providers. 449 00:30:09,040 --> 00:30:12,450 Harvey made a play for Vlex, I think as part of their series. 450 00:30:12,450 --> 00:30:17,865 They were trying to raise, yeah, five, 600 million to buy Vlex and that didn't happen. 451 00:30:17,865 --> 00:30:18,590 And then... 452 00:30:18,590 --> 00:30:26,445 I know some people who work at like, you know, the open legal data sets that are out there, some of the more free ones, and they've been approached by the likes of open AI for 453 00:30:26,445 --> 00:30:26,995 purchasing. 454 00:30:26,995 --> 00:30:31,006 So they don't have a complete set, but they have a really good set that could be really useful. 455 00:30:31,006 --> 00:30:38,058 You know, the benefit of having the Lexus and Westlaw is you've got all the secondary materials, which are beautiful for training, uh even on, you know, very niche kind of 456 00:30:38,058 --> 00:30:39,378 specific areas of the law. 457 00:30:39,378 --> 00:30:44,140 You've got this really good rich data that's curated by attorneys that have worked in the field for a billion years and whatever. 458 00:30:44,140 --> 00:30:44,831 So. 459 00:30:44,831 --> 00:30:45,942 That is the benefit of those ones. 460 00:30:45,942 --> 00:30:50,213 You Felix went the really interesting opposite route, which is we don't have the secondary materials. 461 00:30:50,213 --> 00:30:53,889 We're just going to make it all a Gentic and we're going to craft you an on-demand treatise. 462 00:30:53,889 --> 00:30:59,763 So I think, you know, unless they can get ahold of Westlar or Lexis, I can't imagine Westlar wants to sell ever. 463 00:30:59,763 --> 00:31:01,624 You know, their data is there. 464 00:31:01,624 --> 00:31:05,387 That is their gold mine that they're sitting on, you know, Thompson writers. 465 00:31:05,387 --> 00:31:08,509 But if they can get ahold of Lexis, that would be some pretty interesting content. 466 00:31:08,509 --> 00:31:11,961 And, know, Harvey now has the API to Lexis. 467 00:31:11,961 --> 00:31:13,252 So I don't know. 468 00:31:13,252 --> 00:31:13,582 see. 469 00:31:13,582 --> 00:31:14,305 We'll see. 470 00:31:14,305 --> 00:31:18,674 Yeah, I've heard that integration is very surface level. 471 00:31:18,674 --> 00:31:19,961 um 472 00:31:19,961 --> 00:31:20,932 struck me as strange. 473 00:31:20,932 --> 00:31:26,681 think a lot of techie people, I think it struck us as strange because it's like, know, what you want is good structured data. 474 00:31:26,681 --> 00:31:33,520 You know, it's like saying, here's this big gulp and here's one those little coffee straws to drink out of it, you know, with the API. 475 00:31:33,520 --> 00:31:37,154 So it's like, ah maybe they'll fix that in the future. 476 00:31:37,154 --> 00:31:37,667 I don't know. 477 00:31:37,667 --> 00:31:42,469 Yeah, it's and you know, this brings up another interesting question and something I talk about a lot. 478 00:31:42,469 --> 00:31:57,285 So I'm of the opinion that in the future, let's say five years from now, once we have a enter the tech enabled legal service delivery era, um, I'm a big believer that law firms 479 00:31:57,285 --> 00:32:02,138 are going to have to leverage their data and in order to differentiate, right? 480 00:32:02,138 --> 00:32:04,329 Buying an off the shelf tool is not going to 481 00:32:04,329 --> 00:32:08,160 to differentiate you because your competitor can go buy the same tool. 482 00:32:08,160 --> 00:32:10,421 But you know what is unique to you? 483 00:32:10,421 --> 00:32:15,812 All of those documents that provided winning outcomes for your clients, right? 484 00:32:15,812 --> 00:32:30,569 All the precedent libraries of model documents that have been battle tested and the way legal works today, it's a very bespoke industry and everybody kind of has their own. 485 00:32:30,705 --> 00:32:39,770 You can ask 10 different lawyers at the same law firm in the same practice for the same document and get different iterations of it. 486 00:32:39,811 --> 00:32:44,453 So, you know, I, it makes me wonder something's got to give there, right? 487 00:32:44,453 --> 00:32:55,391 Like if we're going to tech enable legal service delivery, how, how we can't train models to be reflect one lawyers. 488 00:32:55,897 --> 00:33:00,299 perspective, even though that is what Crosby's doing, by the way, I don't know if you know about Crosby AI. 489 00:33:00,299 --> 00:33:11,494 Um, they, they're, I think it was Sequoia is one of the VCs in their cap table and they, they have a podcast called training data and I listened to that and that's, that is what 490 00:33:11,494 --> 00:33:11,983 they're doing. 491 00:33:11,983 --> 00:33:14,295 They just, but they just have a handful of attorneys. 492 00:33:14,295 --> 00:33:20,737 So they are literally training the attorneys, if I understood it correctly, um, to think like that lawyer. 493 00:33:21,028 --> 00:33:24,611 and that's, that's great and makes a lot of sense as 494 00:33:24,611 --> 00:33:26,332 this transitional phase, right? 495 00:33:26,332 --> 00:33:29,574 Like, I don't think that scales, right? 496 00:33:29,574 --> 00:33:34,256 It's expensive and it's one-off and what happens if they lateral? 497 00:33:34,256 --> 00:33:46,294 they take, you know, it's just, so I don't know, what is your thought on kind of the future and how much law firm data is going to come into play and how firms differentiate 498 00:33:46,294 --> 00:33:47,284 themselves? 499 00:33:47,491 --> 00:33:49,131 I think it's gonna be super interesting. 500 00:33:49,131 --> 00:33:55,411 you know, I think, I went to, once I was, I had a consulting appointment at a law firm. 501 00:33:55,411 --> 00:34:02,871 They wanted to implement something very simple, which is just a rag pipeline for all their pleadings, you know, so that they could generate stuff. 502 00:34:03,031 --> 00:34:05,111 And I went over and I said, okay, where's your data? 503 00:34:05,111 --> 00:34:07,191 And they said, well, some of it's in OneDrive. 504 00:34:07,191 --> 00:34:08,451 And I said, okay. 505 00:34:08,511 --> 00:34:11,771 And some of it's, you know, some of it's in NetDocs. 506 00:34:11,771 --> 00:34:12,771 I said, okay. 507 00:34:12,771 --> 00:34:19,071 And they said, some of it's on this old server that we had, you know, like in a box is that some of it's just paper on a shelf. 508 00:34:19,071 --> 00:34:24,491 But, then when you got to the, got to the planes themselves, all different formats, no good metadata, blah, blah. 509 00:34:24,491 --> 00:34:32,091 And I'm like, it's just going to take a lot to get your, you know, your information into a place where you could actually ingest it into any of these AI systems. 510 00:34:32,171 --> 00:34:35,211 So I think that's obviously step one is just getting 511 00:34:35,211 --> 00:34:43,219 a lot of law firms which have kind of lagged behind just modern data practices, get it all into a data lake or whatever so you can actually make use of this thing is going to be a 512 00:34:43,219 --> 00:34:43,959 big part of it. 513 00:34:43,959 --> 00:34:52,506 But I, you know, I think you're right that there is going to definitely be a period of time where there is going to be a custom tool. 514 00:34:52,506 --> 00:34:56,929 You Joe only do immigration law here is your custom tool that uses your internal things. 515 00:34:56,929 --> 00:35:00,092 and by the way, fine tuning now costs $14 or whatever, you know. 516 00:35:00,092 --> 00:35:09,732 So we can fine tune your little tiny model that just sits on prem on your stuff and really understands your workflow and who Ted is in the context of your firm and all of those 517 00:35:09,732 --> 00:35:10,392 things. 518 00:35:10,392 --> 00:35:12,072 I think there will be a while for that. 519 00:35:12,072 --> 00:35:21,232 think at some point it's probably gonna get eaten up by the big guys where they're gonna, know, VLex is already launching the ability to create your own workflows for your firm. 520 00:35:21,232 --> 00:35:24,892 Harvey has some sort of version of that from what I understand. 521 00:35:24,892 --> 00:35:29,692 So I think they're gonna roll out tools that are probably good for 70 % of the market. 522 00:35:30,054 --> 00:35:40,079 But there may be this place on the margins where you can outperform them substantially by using your own data and leveraging it within an internal ecosystem that is good for AI. 523 00:35:40,079 --> 00:35:43,503 So we were going to talk about a couple of things here. 524 00:35:43,503 --> 00:35:50,960 I want to skip down to legal education because I had a conversation this morning with Ed Walters from Felix. 525 00:35:51,602 --> 00:35:53,143 Yeah, he's a great guy. 526 00:35:53,143 --> 00:35:55,646 He's actually going to be on our 100th episode. 527 00:35:55,646 --> 00:35:57,103 um 528 00:35:57,103 --> 00:35:58,084 He's always a good get. 529 00:35:58,084 --> 00:35:59,785 Always has something interesting to say. 530 00:35:59,805 --> 00:36:01,107 And he's really nice. 531 00:36:01,107 --> 00:36:02,489 So I see him at conferences. 532 00:36:02,489 --> 00:36:05,331 He's just a really, really nice guy too. 533 00:36:06,553 --> 00:36:07,101 Yeah. 534 00:36:07,101 --> 00:36:12,405 we were talking a little bit about lawyer training and you guys are at the forefront of that. 535 00:36:12,566 --> 00:36:27,159 So I'm curious, Haley, like how do you think about how in an academic setting, lawyer training is going to need to evolve from where it is today to where we may be heading 536 00:36:27,159 --> 00:36:29,587 three, four years ago when the 537 00:36:29,587 --> 00:36:37,629 blocking and tackling work is really automated and document generation and we need to elevate lawyers to be more consultants. 538 00:36:37,629 --> 00:36:42,185 Like, how are you guys thinking about it at the law school or are you? 539 00:36:42,185 --> 00:36:45,089 Is it kind of a wait and see situation? 540 00:36:45,089 --> 00:36:45,877 What's that look like? 541 00:36:45,877 --> 00:36:47,618 think it depends on who you talk to. 542 00:36:47,618 --> 00:36:49,820 There's definitely resistance. 543 00:36:49,820 --> 00:36:57,926 I'm a relatively new professor and obviously I like AI and new things and so I think I'm a little bit more forward thinking about all of these things. 544 00:36:57,926 --> 00:37:10,600 But there is a very serious discussion that I think everyone is like engaging in about our students and their future because what first, second, third year attorneys usually do, 545 00:37:10,600 --> 00:37:15,505 Doc review, things like that are all the things that AI is going to be able to do in two seconds, right? 546 00:37:15,505 --> 00:37:22,882 And so what I think is really important for us as educators is to train students to know how to use these tools. 547 00:37:22,882 --> 00:37:27,996 They can come in as the people that can train all the other people in the law firms, right? 548 00:37:27,996 --> 00:37:30,180 So they can bring value that way. 549 00:37:30,180 --> 00:37:36,737 So like in my classes, I don't mind if students use AI to help them with practice problems, generate potential. 550 00:37:36,737 --> 00:37:40,449 multiple choice questions, essay answers, tell them what they did right or wrong. 551 00:37:40,449 --> 00:37:42,101 I encourage them to do that. 552 00:37:42,101 --> 00:37:51,466 And I think, you know, it's pretty similar, you know, even if AI is helping you generate your response to something, like that's what's essentially going on in law firms right now 553 00:37:51,466 --> 00:37:52,487 anyway, right? 554 00:37:52,487 --> 00:37:55,218 There's shell pleadings, motions. 555 00:37:55,218 --> 00:38:04,143 No one ever starts from scratch if you're in a law firm like that, although their data might be in boxes or on old servers, em as Sean recently found out, right? 556 00:38:04,201 --> 00:38:09,193 But I think it's coming no matter what we want to do about it or how we think about it. 557 00:38:09,193 --> 00:38:18,988 And so my general approach is just to encourage students to learn as much as they can, go to the digital initiatives with Sean that he puts on with Kitten all the time, which is 558 00:38:18,988 --> 00:38:20,559 really great for students. 559 00:38:20,559 --> 00:38:27,153 And I'm trying to learn as well so that I can be a resource for them to put them in the best position possible to succeed. 560 00:38:27,153 --> 00:38:29,266 know that Deans are thinking about it for sure. 561 00:38:29,266 --> 00:38:34,671 Because a huge portion of their your US World News ranking is based on your employment data of your students. 562 00:38:34,671 --> 00:38:43,129 And so we're finally, you know, people have been predicting these like catastrophic job layoffs since Chad GVT came around, we're finally now getting studies that are showing 563 00:38:43,129 --> 00:38:44,110 that this is really happening. 564 00:38:44,110 --> 00:38:54,499 I think Accenture just laid off 11,000 people there was that big study that came out of I think it was Stanford that had 62 million back end data points from the payroll system. 565 00:38:54,499 --> 00:38:59,899 that showed a 13.2 % decline in hiring among younger people. 566 00:39:00,159 --> 00:39:10,319 I, know, with the kind of traditional law school or law firm pyramid where this bottom layer of people, this document review discovery, simple, you know, kind of simple drafting 567 00:39:10,319 --> 00:39:14,639 research, like AI does all of those things very well and collapses it down. 568 00:39:14,639 --> 00:39:19,331 you know, I think when we think of law firms going forward, it's either going to be kind of a cylinder. 569 00:39:19,331 --> 00:39:25,931 or even a diamond shape is what other people have predicted, which I think is pretty interesting, but it raises a lot of problems for law schools. 570 00:39:26,191 --> 00:39:35,351 the way that I've thought about filling in what are those people gonna go do if they're not going to work at Kirkland, Junior's, Jackson Walker, McAfee Taft, one of the big firms 571 00:39:35,351 --> 00:39:37,731 down here, go and don't leave me where your mom works. 572 00:39:38,031 --> 00:39:47,171 I think non-traditional careers, you I kind of think we're headed for probably a 2008, 2009 type scenario where they're just not hiring first year associates as much. 573 00:39:47,202 --> 00:39:54,178 And so my impulse is to say to my students, hey, especially the ones my AI in the practice of law class, you know, it'd be really cool career legal ops. 574 00:39:54,619 --> 00:40:02,766 And that's not something that law schools traditionally use legal leverage their career services and send people to clock and talk to people you don't like you who are really 575 00:40:02,766 --> 00:40:03,537 into these areas. 576 00:40:03,537 --> 00:40:09,313 And I'm like, that's a great lifestyle, you know, you don't have to go work at a firm, you can still make great money. 577 00:40:09,313 --> 00:40:12,260 And if you're a techie person, like you know me or 578 00:40:12,260 --> 00:40:15,374 you know, like when I was graduating from law school, wouldn't want to go do document review at big firm. 579 00:40:15,374 --> 00:40:20,931 I would have loved to have done something like that instead, but I didn't know that was an option and it wasn't the traditional path for law students. 580 00:40:20,931 --> 00:40:28,592 So I think we need to be a little more creative and hopefully we're setting up the students that come through here so that they could go right into those careers and be 581 00:40:28,592 --> 00:40:29,713 really effective. 582 00:40:29,779 --> 00:40:30,119 Yeah. 583 00:40:30,119 --> 00:40:37,133 So, you I've been an entrepreneur for 32 years and during that time I have consumed a lot of legal services. 584 00:40:37,133 --> 00:40:47,450 And one of the big challenges I've talked about it once or twice on the show before is Lawyers are really good at identifying risk. 585 00:40:47,450 --> 00:40:50,613 as a business person, that's only half the equation. 586 00:40:50,613 --> 00:40:56,322 is a, almost every decision is a risk reward or a cost benefit. 587 00:40:56,322 --> 00:40:57,733 decision, right? 588 00:40:57,733 --> 00:40:59,114 You're balancing those scales. 589 00:40:59,114 --> 00:41:08,250 And lawyers historically have not taken the time to get to know my business because I'd have to pay them to do that, right? 590 00:41:08,250 --> 00:41:09,701 And that would be really expensive. 591 00:41:09,701 --> 00:41:19,128 I understand why they haven't, but you know, like that's always been a challenge from where I sit because I've had lawyers, you know, basically advise me, don't sign this 592 00:41:19,128 --> 00:41:19,888 agreement. 593 00:41:19,888 --> 00:41:21,959 And I'm like, you can't say that. 594 00:41:22,016 --> 00:41:25,406 What you can tell me is you can inform me on the risk side of the equation. 595 00:41:25,406 --> 00:41:31,008 I, as the business person, need to balance the business opportunity, i.e. 596 00:41:31,008 --> 00:41:33,709 the reward side of the equation. 597 00:41:33,709 --> 00:41:35,769 What you're bringing to me is very important. 598 00:41:35,769 --> 00:41:37,770 It helps me balance the scales. 599 00:41:37,810 --> 00:41:46,905 So what I would love to see as a consumer of legal services is lawyers get more in an advisory capacity and bring the human element. 600 00:41:46,905 --> 00:41:48,247 I'll give you an example. 601 00:41:48,247 --> 00:41:59,690 So we hired someone away from a, they're not really a competitor, but a company who had a non-compete and we had to craft his job description in a way where we wouldn't run a foul 602 00:41:59,690 --> 00:42:00,931 of that non-compete. 603 00:42:00,931 --> 00:42:06,056 And we hired a L &E attorney and he gave great advice and it wasn't really legal. 604 00:42:06,056 --> 00:42:09,662 He's like, look, you know, the reality is, you know, for the first, 605 00:42:09,662 --> 00:42:13,462 X number of months, there's probably going to be a lot of scrutiny around this. 606 00:42:13,462 --> 00:42:22,162 And then over time that will probably, you know, there, there are a bill, a judge wouldn't look kindly if six months down the road, they, they threw a flag, right? 607 00:42:22,162 --> 00:42:23,842 So he was giving me human advice. 608 00:42:23,842 --> 00:42:27,002 Like, you know, this is how the company is going to look at this. 609 00:42:27,002 --> 00:42:31,882 This is how the judge would look at an action that came six months later. 610 00:42:31,882 --> 00:42:37,082 Like that sort of stuff is going to be really hard to get out of chat GPT. 611 00:42:37,082 --> 00:42:38,382 It requires 612 00:42:38,458 --> 00:42:39,730 Experience is it possible? 613 00:42:39,730 --> 00:42:42,724 I guess I right 614 00:42:42,836 --> 00:42:48,489 you have to fill up the current 1 million token context window in order to have all that context if then, you know, even. 615 00:42:48,489 --> 00:43:01,557 Yeah, and even then, this guy was a seasoned vet and had spent 30 years on plaintiff side cases. 616 00:43:01,557 --> 00:43:10,102 And again, how you load that into an AI model and get solid judgment back, I think, presents challenges. 617 00:43:10,102 --> 00:43:11,825 uh 618 00:43:11,825 --> 00:43:16,207 when they survey, they survey, you know, our customers from law firms, they all hate us. 619 00:43:16,207 --> 00:43:17,807 They hate to see us. 620 00:43:17,807 --> 00:43:18,888 They hate to talk to us. 621 00:43:18,888 --> 00:43:26,162 They hate to pick up the phone in advance, which is, know, in many instances would remediate a lot of risk if they could just, hey, Sean, what about this? 622 00:43:26,162 --> 00:43:27,332 They hate doing all that stuff. 623 00:43:27,332 --> 00:43:35,445 So, I mean, that's why when you hear it, like a lot of the access to justice, people talk about this technology, the more optimistic ones, they go look, 78 % of the market doesn't 624 00:43:35,445 --> 00:43:38,956 have access to a civil attorney, an attorney in a civil context. 625 00:43:38,956 --> 00:43:40,633 If this drops the cost. 626 00:43:40,633 --> 00:43:50,178 cost and it allows a whole nother layer of people to have better legal outcomes or God forbid uh makes us have just better interactions with our clients so that they want to 627 00:43:50,178 --> 00:43:50,969 come talk to us. 628 00:43:50,969 --> 00:43:55,507 They enjoy picking up the phone because they feel more reassured about their business or whatever. 629 00:43:55,507 --> 00:43:57,192 You know, there is a lot of promise with this stuff. 630 00:43:57,192 --> 00:44:02,525 I try not to be completely starry eyed and rose colored glasses about it, but I'm just an optimistic person. 631 00:44:02,525 --> 00:44:08,749 I think there are a lot of wins here or just our interactions with our clients and better legal outcomes for regular people. 632 00:44:08,749 --> 00:44:09,749 think too. 633 00:44:10,078 --> 00:44:10,258 Yeah. 634 00:44:10,258 --> 00:44:15,169 And you have to ask yourself, why did, why do they hate answering the phone or calling the lawyers? 635 00:44:15,169 --> 00:44:17,631 You know, and, I actually don't hate it. 636 00:44:17,631 --> 00:44:19,552 Uh, what I do hate. 637 00:44:19,552 --> 00:44:27,475 Yeah, no, I mean, you know, we've got, we've got, you know, probably the number one startup law firm in the country. 638 00:44:27,475 --> 00:44:29,216 You can figure out who that is. 639 00:44:29,216 --> 00:44:35,179 And, it is, it is an education process every time I get to, talk to my counterpart over there. 640 00:44:35,179 --> 00:44:36,683 And I learned so much. 641 00:44:36,683 --> 00:44:37,393 And he's great. 642 00:44:37,393 --> 00:44:40,966 And he does think about things in risk reward. 643 00:44:40,966 --> 00:44:43,947 And that's why he's my attorney. 644 00:44:43,947 --> 00:44:46,209 And I enjoy picking up the phone. 645 00:44:46,209 --> 00:44:48,229 I don't enjoy getting the bills. 646 00:44:49,370 --> 00:44:54,174 That part's not fun, but I do feel like I get a lot of value from it. 647 00:44:54,174 --> 00:44:56,415 I've not always felt that way. 648 00:44:56,436 --> 00:45:00,378 And I don't know if this is a function, cause I've been in a lot of different businesses. 649 00:45:00,378 --> 00:45:02,779 My wife and I own five gyms here in St. 650 00:45:02,779 --> 00:45:03,528 Louis. 651 00:45:03,528 --> 00:45:06,519 fair amount of legal work that went into building that those out. 652 00:45:06,519 --> 00:45:07,859 I owned a collection agency. 653 00:45:07,859 --> 00:45:09,898 used to sue people all the time. 654 00:45:09,898 --> 00:45:11,461 A lot of legal work there. 655 00:45:11,461 --> 00:45:18,624 I've not always enjoyed it and I'm not sure if it's because these guys are super niche and they focus on venture backed startups like we are. 656 00:45:18,624 --> 00:45:25,206 But they really know our business and I learned something every time I talked to them and I enjoy it. 657 00:45:25,692 --> 00:45:28,436 Yeah, and there's a whole, you know, so I think there's two parts to that. 658 00:45:28,436 --> 00:45:36,043 Number one, some lawyers when you talk to them are sort of like when I talk to my IT department, excuse me, and I go, hey, I need to download this really interesting open 659 00:45:36,043 --> 00:45:43,788 source, completely unvetted software, because I need to keep my skills, you know, hot and, you know, up to the bleeding edge. 660 00:45:43,788 --> 00:45:45,229 And I want to use a cool new thing. 661 00:45:45,229 --> 00:45:46,460 And they just go, absolutely not. 662 00:45:46,460 --> 00:45:48,641 have no risk, no risk tolerance whatsoever. 663 00:45:48,641 --> 00:45:53,366 Because what's the incentive to them if I can use a new AI thing? 664 00:45:53,366 --> 00:45:54,546 Zero. 665 00:45:54,746 --> 00:45:58,988 What's the risk if I huge, you know, and so they just dial up risk. 666 00:45:58,988 --> 00:46:07,350 And I'm like, well, I have other interests, like remaining relevant or, you know, with library resources that happens very frequently with a thing we will not put behind 10 667 00:46:07,350 --> 00:46:09,431 different walls before the user can get to it. 668 00:46:09,431 --> 00:46:13,732 And a librarian goes, nobody's going to use it if it's a hassle to get in and use the thing. 669 00:46:13,732 --> 00:46:19,623 So it's just like, like you said, I think just the incentives are just pointed in the wrong direction for some lawyers and law firms. 670 00:46:19,623 --> 00:46:20,907 Yeah. 671 00:46:20,907 --> 00:46:23,667 Do you have anything to add to AI stuff I was talking about? 672 00:46:23,667 --> 00:46:24,987 Sorry, I got a pun intended. 673 00:46:25,107 --> 00:46:33,007 it just kind of reminded me when you're talking about the experience that the labor and employment attorney had that you found really valuable. 674 00:46:33,007 --> 00:46:43,227 I was a law clerk for a few judges and tons of firms and clients found it very valuable because I worked in judges chambers for years at a time. 675 00:46:43,227 --> 00:46:49,379 I understood how the judges I worked for thought, how their chambers operated, all this information that 676 00:46:49,385 --> 00:46:50,936 AI can't get its hands on, right? 677 00:46:50,936 --> 00:47:01,722 And so that was value I could bring to clients that is aside from document review and, you know, it's advice that can actually be beneficial that you're not going to get from AI or 678 00:47:01,722 --> 00:47:02,933 even every attorney. 679 00:47:02,933 --> 00:47:08,887 And so there are ways I think that attorneys just in general can utilize things that AI just can't do. 680 00:47:08,887 --> 00:47:11,148 Now, to me, though, it brings up this issue. 681 00:47:11,148 --> 00:47:13,799 You know, you're talking about the 30 years of experience. 682 00:47:13,799 --> 00:47:16,681 Well, if attorneys can't even get in the door. 683 00:47:16,681 --> 00:47:25,926 you know, after law school, because AI has taken over all of their tasks, we're going to end up with a huge shortage of attorneys at some point that have have any experience and 684 00:47:25,926 --> 00:47:35,261 then that you know, value might be lost if there's not a way to kind of have attorneys recalibrate like what they bring to the table and then get that experience as years go on. 685 00:47:35,261 --> 00:47:42,585 Yeah, I some people have, you know, when I go to other conferences, like Dan Lena was from Northwestern was talking about this where he was saying, you know, essentially, what we're 686 00:47:42,585 --> 00:47:45,256 going to see is just attorneys just highly specialized. 687 00:47:45,256 --> 00:47:53,221 So I'm an expert in e-discovery, in transactional, in California, for these real estate, you know, whatever, just really, really narrowly specialized. 688 00:47:53,221 --> 00:47:55,912 And so maybe that would be a situation like yours. 689 00:47:55,912 --> 00:47:59,514 You now got this attorney who he has dialed into your exact sector. 690 00:47:59,514 --> 00:48:03,617 You know that he's competent, you know, he has the right level of risk tolerance for you. 691 00:48:03,617 --> 00:48:07,279 And he gives you this white glove service that's very customized to Ted. 692 00:48:07,279 --> 00:48:13,054 And so maybe it'll just be that kind of times everything in law as what he, you know, they've hypothesized. 693 00:48:13,054 --> 00:48:14,950 They say the riches are in the niches. 694 00:48:14,950 --> 00:48:16,273 So, um, 695 00:48:16,273 --> 00:48:16,864 what I my students. 696 00:48:16,864 --> 00:48:18,296 I said, you know, this is a great time. 697 00:48:18,296 --> 00:48:24,403 Like you want lost services Chipotle burrito, you know, you want extra guac and so Fritas instead of carnitas. 698 00:48:24,403 --> 00:48:26,115 Let me get you exactly what you want, sir. 699 00:48:26,115 --> 00:48:28,068 That expensive burrito. 700 00:48:28,068 --> 00:48:29,930 Deliver it to your table for you. 701 00:48:29,930 --> 00:48:31,571 And here's the drink that you enjoy. 702 00:48:31,571 --> 00:48:33,665 Well, this has been a great conversation. 703 00:48:33,665 --> 00:48:38,031 really appreciate you guys coming and spending a little bit of time with me on the show. 704 00:48:38,031 --> 00:48:39,147 um 705 00:48:39,147 --> 00:48:42,850 got to kick the tires on our new podcast studio that Jim put together for us. 706 00:48:42,850 --> 00:48:44,211 Yeah, it's been great. 707 00:48:44,211 --> 00:48:44,611 Thank you. 708 00:48:44,611 --> 00:48:46,323 Awesome. 709 00:48:46,323 --> 00:48:47,183 I'm really impressed with it. 710 00:48:47,183 --> 00:48:49,543 I'm gonna do all my big presentations in here from now on. 711 00:48:49,543 --> 00:48:50,796 Yeah, you should. 712 00:48:50,796 --> 00:48:57,030 How do folks find out more about the work that you're doing there at the law school and you individually? 713 00:48:57,030 --> 00:48:58,863 What's the best way for them to do that? 714 00:48:59,111 --> 00:49:03,126 Probably for me, LinkedIn is kind of where I consolidate all my professional stuff. 715 00:49:03,126 --> 00:49:06,059 And it's, I'll have links to like my SSRN articles. 716 00:49:06,059 --> 00:49:09,102 And when I presented a thing, if I'm going to Ulta and presenting or whatever. 717 00:49:09,102 --> 00:49:09,793 Yeah. 718 00:49:09,793 --> 00:49:12,716 And then we're both, we both have pages on OU Laws website. 719 00:49:12,716 --> 00:49:17,470 So you can just search for our name, all of our papers and work will kind of be linked there as well. 720 00:49:17,470 --> 00:49:18,230 Okay. 721 00:49:18,230 --> 00:49:18,820 Well, awesome. 722 00:49:18,820 --> 00:49:22,332 We'll include those links in the show notes so folks can access them. 723 00:49:22,332 --> 00:49:28,893 And thank you so much for being on the show and hopefully we get to meet in person at a future conference. 724 00:49:30,974 --> 00:49:32,085 You as well. 00:00:02,083 Sean Haley, how are you today? 2 00:00:02,083 --> 00:00:03,453 Good to see you. 3 00:00:03,453 --> 00:00:05,655 I really appreciate you guys joining me, man. 4 00:00:05,655 --> 00:00:12,260 I enjoyed your Michael Scott paper quite a bit and we're going to talk about it today. 5 00:00:12,260 --> 00:00:14,602 But before we do, let's get you guys introduced. 6 00:00:14,602 --> 00:00:20,985 So you're both academics at University of, is it Oklahoma? 7 00:00:21,766 --> 00:00:22,786 University of Oklahoma. 8 00:00:22,786 --> 00:00:24,027 In my notes, have Arkansas. 9 00:00:24,027 --> 00:00:25,988 That's wrong. 10 00:00:26,068 --> 00:00:29,450 Why don't, Hailey, we start with you, introduce yourself and then we'll... 11 00:00:29,450 --> 00:00:30,811 Hand it over to Sean. 12 00:00:31,017 --> 00:00:31,387 Sure. 13 00:00:31,387 --> 00:00:33,128 So I'm Haley Stilwell. 14 00:00:33,128 --> 00:00:37,950 I am an associate professor of law at the University of Oklahoma College of Law. 15 00:00:37,950 --> 00:00:40,211 I teach evidence and criminal law. 16 00:00:40,211 --> 00:00:45,813 I write in those fields as well, but I also have a particular interest in exploring AI in my scholarship. 17 00:00:45,813 --> 00:00:49,015 And that's kind of where Sean and I's relationship began. 18 00:00:49,015 --> 00:00:49,255 Yeah. 19 00:00:49,255 --> 00:00:50,025 And I'm Sean Harrington. 20 00:00:50,025 --> 00:00:57,178 I'm the director of technology innovation here at OU Law, which is kind of a nebulous title, you know, at different institutions that means different things. 21 00:00:57,178 --> 00:00:59,711 Here what it means is I teach AI in the practice of law. 22 00:00:59,711 --> 00:01:06,231 And I also run what's called their digital initiative, which gets tech training to all the students, essentially. 23 00:01:06,326 --> 00:01:09,869 And I met someone from your law school. 24 00:01:09,869 --> 00:01:10,440 Yeah. 25 00:01:10,440 --> 00:01:11,930 Was it Kenton? 26 00:01:12,792 --> 00:01:13,572 Yeah. 27 00:01:13,572 --> 00:01:19,107 Um, we were, he happened to be attending a session that I was helping, helping facilitate and sat at my table. 28 00:01:19,107 --> 00:01:23,601 So it was pretty cool when I saw your paper and was like, yeah. 29 00:01:23,601 --> 00:01:26,404 And connected the dots and, here we are today. 30 00:01:27,626 --> 00:01:28,459 So 31 00:01:28,459 --> 00:01:29,190 to your podcast. 32 00:01:29,190 --> 00:01:31,432 And so he was like, my gosh, you got to go on that. 33 00:01:31,432 --> 00:01:33,373 I was like, Okay, yeah, I'm in. 34 00:01:33,434 --> 00:01:34,014 Good. 35 00:01:34,014 --> 00:01:34,374 Good. 36 00:01:34,374 --> 00:01:34,594 Yeah. 37 00:01:34,594 --> 00:01:44,574 We started off, you know, we started off primarily in the knowledge management and innovation realm at large law firms as our audience. 38 00:01:44,574 --> 00:01:46,174 And it's really evolved from that. 39 00:01:46,174 --> 00:01:49,954 I mean, we have, I've got people from the marketing function. 40 00:01:49,954 --> 00:01:52,985 I've got attorneys, I've got managing partners, I have academics. 41 00:01:52,985 --> 00:01:59,721 So, you know, the innovation topic is very timely in the legal world. 42 00:01:59,721 --> 00:02:00,477 right. 43 00:02:00,477 --> 00:02:03,718 Like the timing was, the timing was very fortunate for us. 44 00:02:03,718 --> 00:02:07,940 So I started right around the time chat GPT was released. 45 00:02:07,940 --> 00:02:08,740 Yeah. 46 00:02:08,740 --> 00:02:10,260 So, I mean, just pure luck. 47 00:02:10,260 --> 00:02:13,181 Like I didn't see that and go, we need a podcast. 48 00:02:13,181 --> 00:02:18,754 was just, we had been selling into not the world of knowledge management for many, many years. 49 00:02:18,754 --> 00:02:26,164 And then we saw this interesting dynamic where all saw these innovation roles start to pop up and 50 00:02:26,164 --> 00:02:26,958 Right. 51 00:02:27,244 --> 00:02:31,688 started having conversations with, what's the difference between knowledge management and innovation? 52 00:02:31,688 --> 00:02:35,612 And those conversations were very enlightening. 53 00:02:35,612 --> 00:02:37,372 And I thought, you know what? 54 00:02:37,473 --> 00:02:40,115 I should record these and make a podcast. 55 00:02:40,115 --> 00:02:54,757 So I recorded three episodes, didn't tell anyone about it, posted them, you know, just on the usual platforms, Spotify, Apple podcasts, like three weeks before, Ilta Ilta con where 56 00:02:54,757 --> 00:02:55,338 we met. 57 00:02:55,338 --> 00:03:00,578 Kenton and I had like 10 people come up to me in the booth and go, Hey, I your podcast. 58 00:03:00,578 --> 00:03:01,018 It's great. 59 00:03:01,018 --> 00:03:03,398 I was like, how the hell did you find it? 60 00:03:03,398 --> 00:03:12,958 And I guess people just, I don't know if they keep Google alerts on certain keywords, but, um, that was almost a hundred episodes. 61 00:03:12,958 --> 00:03:15,458 We're going to have our hundredth episode here in a few weeks. 62 00:03:15,458 --> 00:03:18,115 And, uh, yeah, it's been, it's been a lot of fun. 63 00:03:18,115 --> 00:03:18,495 awesome. 64 00:03:18,495 --> 00:03:27,835 mean, the thing is you go to something like Ilta and you just have, you know, 100 of those types of conversations and you run into Damien real or whatever, you know, and you have a 65 00:03:27,835 --> 00:03:32,075 conversation like, man, I wish I had somebody to save this, you know, other people could repurpose this. 66 00:03:32,075 --> 00:03:37,175 I want to have so I have this little, uh, have you seen the limitless AI pendant? 67 00:03:37,334 --> 00:03:42,692 No, but I've used, I've actually used the, think it's called plod, which I think it's similar. 68 00:03:43,437 --> 00:03:43,849 Yeah. 69 00:03:43,849 --> 00:03:46,802 And so he just sticks right here and then you have that knowledge base for forever. 70 00:03:46,802 --> 00:03:50,816 And I'm like, man, anytime I want to go to one of those conferences, I'm to be wearing that thing. 71 00:03:50,816 --> 00:03:59,666 Yeah, you know, the whole privacy thing around that, like, do you have to tell everybody like, hey, I'm recording that, you know, I was, yeah. 72 00:03:59,666 --> 00:04:05,546 of don't have like the social etiquette for it because obviously, you know, people knew that you were recording them, they would speak to you very differently. 73 00:04:05,546 --> 00:04:13,146 But you want the other conversation because the utility of retaining that information and having it in a data set that you could manipulate is like so high. 74 00:04:13,146 --> 00:04:20,886 It's just like, we're gonna have to figure out something like a little green, you know, light on my lapel that says Sean's always recording or something. 75 00:04:20,886 --> 00:04:22,406 think I turn it off. 76 00:04:22,406 --> 00:04:25,146 Nobody's want Nobody's gonna want to talk to me anyways. 77 00:04:25,146 --> 00:04:26,582 That's only gonna make it worse. 78 00:04:26,582 --> 00:04:29,211 in the legal crowd, right? 79 00:04:29,211 --> 00:04:31,918 I mean, they're very privacy sensitive. 80 00:04:31,918 --> 00:04:37,265 I had a faculty member here that I was just explaining what the thing did. 81 00:04:37,265 --> 00:04:38,096 It was in its dock. 82 00:04:38,096 --> 00:04:39,149 I was like, it's not recording. 83 00:04:39,149 --> 00:04:40,551 I was explaining what it was. 84 00:04:40,551 --> 00:04:44,115 And she made up an excuse to get up and leave my office almost immediately. 85 00:04:44,383 --> 00:04:45,003 That's funny. 86 00:04:45,003 --> 00:04:45,423 Yeah. 87 00:04:45,423 --> 00:04:54,763 I mean, we, um, we use a platform called fireflies for our meetings and 100 % of our, of our client base is law, our law firms, large law firms. 88 00:04:55,183 --> 00:04:59,983 And, um, surprisingly we don't get a lot of pushback on that. 89 00:05:00,063 --> 00:05:09,803 And, you know, but I do think that people manage their candor when they see and, and rightly so I get it. 90 00:05:09,803 --> 00:05:11,319 Um, but 91 00:05:11,355 --> 00:05:14,071 Anyway, let's talk about your paper. 92 00:05:14,071 --> 00:05:15,706 So it really caught my eye. 93 00:05:15,706 --> 00:05:19,112 It was, and the reason is because I'm a big office fan. 94 00:05:19,484 --> 00:05:20,464 Love it. 95 00:05:20,464 --> 00:05:22,324 So the title was all Haley. 96 00:05:22,324 --> 00:05:24,264 She came up with that. 97 00:05:25,924 --> 00:05:27,264 Oh, yeah. 98 00:05:27,264 --> 00:05:27,535 Yeah. 99 00:05:27,535 --> 00:05:29,077 they're great. 100 00:05:29,077 --> 00:05:35,343 So if I remember correctly, the title was Michael Scott is not a juror. 101 00:05:36,244 --> 00:05:36,766 Okay. 102 00:05:36,766 --> 00:05:40,654 And, um, give us a little bit of background on the paper. 103 00:05:40,654 --> 00:05:45,203 Uh, I don't know whoever wants to start and then we can jump into that a little bit. 104 00:05:45,321 --> 00:05:46,491 Yeah, you started, it your idea. 105 00:05:46,491 --> 00:05:51,744 Yeah, so the origin story kind of comes from my just academic research in general. 106 00:05:51,744 --> 00:06:01,828 There's this standard that's often utilized em in courts called the reasonable juror standard where courts are basically trying to determine what a perfectly behaving 107 00:06:01,828 --> 00:06:07,210 impartial juror that follows all the court's directions, how they would think and understand evidence. 108 00:06:07,210 --> 00:06:14,387 And it comes into play a lot when em courts are assessing whether the admission or exclusion of evidence was 109 00:06:14,387 --> 00:06:16,939 in violation of a rule or the constitution. 110 00:06:16,939 --> 00:06:26,874 And so I was bothered by some of the court's reasonable jury determinations because it just didn't track at all with how I thought real people would think and understand about 111 00:06:26,874 --> 00:06:27,985 certain evidence. 112 00:06:27,985 --> 00:06:38,390 So that prompted me to want to create a mock jury study to compare what the court has determined reasonable jurors would think and compare it to what real people would think 113 00:06:38,390 --> 00:06:43,301 and kind of highlight where the court has got it wrong and in turn results in 114 00:06:43,301 --> 00:06:45,922 probable constitutional violations. 115 00:06:45,922 --> 00:06:57,508 And so with that project, what I wanted to then do is find a way to empirically ground the reasonable juror where possible to bring the reasonable juror and real jurors closer 116 00:06:57,508 --> 00:07:00,389 together when courts were making those determinations. 117 00:07:00,389 --> 00:07:06,682 And the easiest way to do that is with mock jury data like we did in this study, but it takes way too much time. 118 00:07:06,682 --> 00:07:08,867 It's way too expensive to do. 119 00:07:08,867 --> 00:07:20,567 And so when uh Sean and I were kind of talking about it, we got the idea, what if we could create a mock jury bot that could output mock jury information instantaneously to help 120 00:07:20,567 --> 00:07:27,852 inform courts when they're making these determinations to protect criminal defendants and all people that are appealing their cases. 121 00:07:27,852 --> 00:07:38,704 And so we thought it would be pretty easy, just use existing platforms to feed them, you know, what this, the demographic profile of a juror and spit out information and 122 00:07:38,704 --> 00:07:41,156 Hopefully it would track what real jurors would do. 123 00:07:41,156 --> 00:07:43,587 And that is just not even close to what happened. 124 00:07:43,587 --> 00:07:44,958 And that's where this paper came. 125 00:07:44,958 --> 00:07:45,308 Yeah. 126 00:07:45,308 --> 00:07:50,642 And like the other thing is, you know, we looked up a bunch of studies that were trying to do this and it's like sample size, 112 people. 127 00:07:50,642 --> 00:07:56,566 So there just wasn't a lot of research and uh actual surveys of humans done on this stuff. 128 00:07:56,566 --> 00:08:00,669 So, you know, you quickly realize that it's really expensive to survey people. 129 00:08:00,669 --> 00:08:05,773 And so there just wasn't like a good academic research in this area. 130 00:08:05,773 --> 00:08:06,649 So we're like, well, 131 00:08:06,649 --> 00:08:10,635 That's even another reason then to integrate AI into this thing. 132 00:08:10,801 --> 00:08:11,111 Yeah. 133 00:08:11,111 --> 00:08:20,087 And you know, I, we talked to, we get a lot of attorneys on the show, some practicing, some not the ones who practice. 134 00:08:20,087 --> 00:08:28,072 I always like to ask, you know, what are the valuable use cases with AI trial prep is near the top of the list. 135 00:08:28,072 --> 00:08:31,034 And the data doesn't show this. 136 00:08:31,034 --> 00:08:33,396 really it's kind of drafting due diligence. 137 00:08:33,396 --> 00:08:36,923 you see as kind of the most common, document review. 138 00:08:36,923 --> 00:08:38,559 Yeah, document review would seem like fun. 139 00:08:38,559 --> 00:08:41,859 Yeah, it seemed to be the most common use cases cited. 140 00:08:41,879 --> 00:08:49,650 But when I hear about in terms of value, we had a guy on the podcast, name's Stephen Embry. 141 00:08:49,650 --> 00:08:50,750 You familiar with? 142 00:08:50,750 --> 00:08:51,650 Yeah. 143 00:08:51,810 --> 00:08:58,650 So Steve is a former, he was a 30 year attorney at I believe Finnegan, which is a, they're a client of ours. 144 00:08:58,650 --> 00:09:00,930 They're an intellectual property firm. 145 00:09:01,270 --> 00:09:04,914 And he was telling me how 146 00:09:05,111 --> 00:09:19,716 either him or his colleagues would use it for trial preparation in basically having AI challenge the evidence in some way, like ask all of the questions that opposing counsel 147 00:09:19,716 --> 00:09:22,977 might present in a courtroom. 148 00:09:22,977 --> 00:09:30,249 he said, you know, he talked about how, how useful that was, which I found very intriguing and makes perfect sense. 149 00:09:30,249 --> 00:09:32,760 Like I'm a daily user of AI. 150 00:09:32,860 --> 00:09:33,877 super familiar with. 151 00:09:33,877 --> 00:09:34,647 drive here for you. 152 00:09:34,647 --> 00:09:36,118 said, pretend. 153 00:09:36,118 --> 00:09:37,298 I said, here's our paper. 154 00:09:37,298 --> 00:09:40,469 Pretend that you're Ted and ask me some stuff. 155 00:09:40,469 --> 00:09:45,541 I mean, it really lowers the temperature coming into the environment because you're like, okay, I'm feeling pretty confident. 156 00:09:45,541 --> 00:09:49,172 You know, I've kind of been challenged on things that where I wasn't super confident. 157 00:09:49,172 --> 00:09:50,132 I looked a few things up. 158 00:09:50,132 --> 00:09:51,983 So it's like, you feel ready to roll. 159 00:09:51,983 --> 00:09:55,230 So I could definitely see that having a lot of value for attorneys. 160 00:09:55,230 --> 00:09:58,441 Yeah, I mean, it's almost like a practice run that you get. 161 00:09:58,441 --> 00:10:03,943 So when I heard the study, the name was catchy, so that drew me in initially. 162 00:10:03,943 --> 00:10:11,946 But then when I read what you guys were actually doing, I was like, wow, this seems like a perfect use case for AI. 163 00:10:11,946 --> 00:10:18,468 And it turns out it wasn't that perfect using the frontier models. 164 00:10:18,468 --> 00:10:21,699 Tell us about what your findings were initially. 165 00:10:22,083 --> 00:10:28,305 Well, the first thing we tried to do was just let the frontier models try to create a jury. 166 00:10:28,305 --> 00:10:34,016 So we said, create for us a jury pool that is similar to what a federal jury pool would be. 167 00:10:34,016 --> 00:10:36,438 And that's where Michael Scott emerged. 168 00:10:36,438 --> 00:10:39,100 It was really hilarious. 169 00:10:39,100 --> 00:10:41,411 They would output the demographics of the jurors. 170 00:10:41,411 --> 00:10:48,433 So it was a white man in his mid-40s who is the manager of a mid-sized paper firm in Scranton, Pennsylvania. 171 00:10:48,433 --> 00:10:52,145 which you and I would obviously know is Michael Scott. 172 00:10:52,145 --> 00:10:57,867 Michael Scott is not a real person, let alone a real juror in the federal jury pool, right? 173 00:10:57,867 --> 00:11:06,069 We also had a lot of other interesting combinations of, there was a 90 year old woman who was a part-time botanist, part-time DJ. 174 00:11:06,069 --> 00:11:07,437 I love that one. 175 00:11:07,437 --> 00:11:10,311 We had an abolitionist podcaster. 176 00:11:10,311 --> 00:11:15,824 So it seemed like when these platforms were left to their own devices, they were generating 177 00:11:15,824 --> 00:11:25,106 jurors that were more for show kind of eye catching types of backgrounds that really didn't reflect what we needed for our purposes. 178 00:11:25,106 --> 00:11:29,717 What real people on a jury would actually look like demographically. 179 00:11:29,717 --> 00:11:30,478 Yeah. 180 00:11:30,478 --> 00:11:39,560 And you can tell that, know, there a kid in, you know, Washington is using them right now to study who's 12 years old and maybe using it for creative writing. 181 00:11:39,560 --> 00:11:44,079 So, you know, there's a big range of why people are using these tools and they have the dial. 182 00:11:44,079 --> 00:11:51,150 on certain types of representation, which could be very useful, obviously, in a creative writing context, but in ours, that was, you know, catastrophic, because it was wasn't 183 00:11:51,150 --> 00:11:52,451 representing reality. 184 00:11:52,896 --> 00:11:55,888 Well, and look, I love Michael Scott. 185 00:11:55,888 --> 00:11:59,200 He'd be the last person I'd ever choose to be on a jury. 186 00:11:59,200 --> 00:12:01,301 Not, not known for good judgment. 187 00:12:01,301 --> 00:12:04,543 Um, well, and that's interesting. 188 00:12:04,543 --> 00:12:19,501 So I have had a lot of debate myself and I've seen a lot of debate in cyberspace, LinkedIn and Reddit about the frontier models coming and essentially crushing the legal specific 189 00:12:19,501 --> 00:12:27,750 tools just based on just how quickly the trajectory that they're on in terms of capabilities and 190 00:12:27,750 --> 00:12:38,941 I think your findings will first tell us what you found as potential root cause as to why this experiment was not successful with the Frontier models. 191 00:12:38,941 --> 00:12:39,461 Yeah. 192 00:12:39,461 --> 00:12:45,143 So first of all, you know, they have lengthy system prompts, which people, you know, leak on GitHub all the time. 193 00:12:45,143 --> 00:12:50,608 And you can see kind of why they're training these models to do things like be a good study tool for a student. 194 00:12:50,608 --> 00:12:57,852 And then we also just saw that just rampant bias on certain topics that just was completely at odds with reality. 195 00:12:57,852 --> 00:13:07,117 And so when we were looking, you when we actually finally ran our actual mock juror surveys, what we discovered is that people are like fairly moderate on average. 196 00:13:07,117 --> 00:13:10,320 And these would be kind of extreme versions of people. 197 00:13:10,320 --> 00:13:17,727 And again, you know, I think it's just because they're, you know, kind of meant to do everything instead of replicate juror juror scenarios. 198 00:13:17,727 --> 00:13:18,787 What else? 199 00:13:19,108 --> 00:13:24,132 Well, I think to covering kind of the second part of our experiment is important to explain that as well. 200 00:13:24,132 --> 00:13:33,380 After we left the platforms to their own devices, that's where the mock jury data came in to where, OK, we were going to tell the platforms now. 201 00:13:33,518 --> 00:13:35,409 here are the demographics of the jurors. 202 00:13:35,409 --> 00:13:37,099 You're not gonna pick them anymore. 203 00:13:37,099 --> 00:13:38,349 No more Michael Scott's. 204 00:13:38,349 --> 00:13:38,699 So. 205 00:13:38,699 --> 00:13:44,451 part coded in all of the demographics with a lengthy system prompt ah using Lang chain if anyone cares. 206 00:13:44,451 --> 00:13:54,354 Yeah, so we use basically the demographics from the real humans and told the platforms generate those exact humans and predict how incriminating this piece of evidence was. 207 00:13:54,354 --> 00:13:58,395 So we had an apples to apples comparison, how accurate were these platforms? 208 00:13:58,395 --> 00:13:59,956 And they weren't very accurate. 209 00:13:59,956 --> 00:14:01,368 Not only were they not accurate, 210 00:14:01,368 --> 00:14:03,340 but they were biased in different ways. 211 00:14:03,340 --> 00:14:16,733 ChatGPT was actually not very biased, then Jim and I thought the evidence was a little more incriminating on average, and then Claude was a little underrating in terms of 212 00:14:16,733 --> 00:14:17,364 incrimination. 213 00:14:17,364 --> 00:14:23,951 That was consistent basically across every type of demographic, so we saw those consistent skews in every single one. 214 00:14:23,951 --> 00:14:26,012 you know, X type of person, was deterministic. 215 00:14:26,012 --> 00:14:32,196 So if you're X type of person, then you necessarily would believe why you would, you know, if you're a certain type person, maybe you don't like the police. 216 00:14:32,196 --> 00:14:37,419 And so you would always find them, you know, guilty, you know, those types of things just happened over and over and over again. 217 00:14:37,419 --> 00:14:41,162 And we knew from our real world human data that that's not how real people behave at all. 218 00:14:41,162 --> 00:14:43,964 You could be a type of person and think something completely different. 219 00:14:43,964 --> 00:14:48,727 And so that, you know, just the replication of stereotypes was rampant through them. 220 00:14:48,727 --> 00:14:51,515 And this was something that as we were doing it, 221 00:14:51,515 --> 00:14:56,108 as somebody who sits in these things all day effectively, I was like, yeah, they're definitely gonna do that. 222 00:14:56,169 --> 00:14:59,801 But that was the most interesting thing to Haley, understandably, I think. 223 00:14:59,801 --> 00:15:04,505 And then when I presented this topic at a couple of different conferences, that was the thing that everybody latched onto. 224 00:15:04,505 --> 00:15:09,979 Like really, these are just wildly over-representing stereotypes and replicating stereotypes. 225 00:15:09,979 --> 00:15:12,360 That's catastrophic for the law in many ways, obviously. 226 00:15:12,360 --> 00:15:14,973 So that was something where I was like, people are interested in this. 227 00:15:14,973 --> 00:15:15,677 And so we... 228 00:15:15,677 --> 00:15:22,277 Spent a long time in the paper talking about that because it was something that maybe doesn't occur to a lot of people like the system prompts. 229 00:15:22,337 --> 00:15:25,597 If you're not, maybe Ted, you know what a system prompt is. 230 00:15:25,597 --> 00:15:34,377 But I think the average chat GPT user doesn't realize before you enter anything into the system, there's a five and 10 and 20 page potentially system prompt that has all of these 231 00:15:34,377 --> 00:15:37,377 instructions, do this, don't do that type stuff. 232 00:15:37,517 --> 00:15:41,737 so those types of things, I think maybe don't occur to the average lawyer. 233 00:15:41,737 --> 00:15:45,417 And they're definitely happening on the legal tools as well. 234 00:15:45,841 --> 00:15:47,739 You know, Westlaw has a lengthy system prompt. 235 00:15:47,739 --> 00:15:51,655 Lexus says they all have lengthy system prompts that tell it what it can and cannot do. 236 00:15:51,655 --> 00:15:52,195 Interesting. 237 00:15:52,195 --> 00:16:03,144 So is this kind of the Swiss army knife scenario where, know, they've got a lot of tools, including a little, little scissors there, but if you ever had to cut a thick piece of 238 00:16:03,144 --> 00:16:05,325 canvas with it, it's just not a fit. 239 00:16:05,325 --> 00:16:07,997 So it's good at doing a lot of things. 240 00:16:07,997 --> 00:16:08,588 Okay. 241 00:16:08,588 --> 00:16:11,410 But not really anything specialized. 242 00:16:11,410 --> 00:16:13,671 Is that really the dynamic at play? 243 00:16:13,870 --> 00:16:17,612 Yeah, and that's how tell the students, know, lawyers are very worried about hallucinations. 244 00:16:17,612 --> 00:16:18,972 In large language models. 245 00:16:18,972 --> 00:16:25,715 I saw students hallucinations at this point in time are basically solved if you're using Vlex's Vincent AI for using Lexus AI fusing Westlaw. 246 00:16:25,715 --> 00:16:27,476 It has a hyperlink to the case. 247 00:16:27,476 --> 00:16:30,477 It's your duty as an attorney to click that and go read the case. 248 00:16:30,477 --> 00:16:31,698 That's why they pay you the big bucks. 249 00:16:31,698 --> 00:16:33,899 But if you're using chat GBT, that's not the case. 250 00:16:33,899 --> 00:16:38,059 And so I think, you know, it is, think the Swiss Army Knife is a really good call. 251 00:16:38,059 --> 00:16:41,099 You're using a Swiss Army Knife when you should be using a skillsaw or something. 252 00:16:41,099 --> 00:16:46,159 And that skillsaw is Lexus, know, Vlex, whatever it is, which is a law specific tool. 253 00:16:46,159 --> 00:16:54,859 And here, what our results really showed is that these platforms were not good at replicating the noise in the real human responses. 254 00:16:54,859 --> 00:17:00,419 You know, it really seemed like demographics ended up being deterministic of what 255 00:17:00,455 --> 00:17:08,084 how incriminating they thought each evidence was, but we saw in our actual real human data that there was a lot more variation among demographics. 256 00:17:08,084 --> 00:17:18,587 so really at a high level, these platforms are just not equipped to replicate human judgment the way it actually comes out with real humans and their varying backgrounds. 257 00:17:18,587 --> 00:17:18,867 Yeah. 258 00:17:18,867 --> 00:17:22,250 And you know, I mean, I've actually noticed some interesting dynamics. 259 00:17:22,250 --> 00:17:26,773 The safety and alignment protocols are implemented very differently. 260 00:17:26,773 --> 00:17:35,568 Um, you know, there's four models that I use primarily, which are in chat, GBT, Claude, Gemini and Grok. 261 00:17:35,568 --> 00:17:40,912 And really I use them in that order in terms of frequency. 262 00:17:40,912 --> 00:17:45,685 but I ran into a really interesting little exercise where 263 00:17:45,823 --> 00:17:50,395 I saw there was a graphic on LinkedIn and it was a caricature of two people that I know. 264 00:17:50,395 --> 00:17:52,436 Um, in fact, they're both podcasters. 265 00:17:52,436 --> 00:17:55,217 was Richard Trowman's and sac of Bramowitz. 266 00:17:55,217 --> 00:18:02,380 And they had this picture and I was like, the guy who's supposed to be Richard Trowman's looks just like this actor, but I can't think of who it is. 267 00:18:02,380 --> 00:18:12,164 So I took a quick screenshot and put it into, I think I did Claude first and was just like, what famous actor does this per does the person given the thumbs up look like? 268 00:18:12,204 --> 00:18:13,458 And it said, 269 00:18:13,458 --> 00:18:14,228 I'm sorry. 270 00:18:14,228 --> 00:18:17,593 I actually have the, let's see if I can find it real quick. 271 00:18:17,593 --> 00:18:19,854 what it said, cause it was really interesting. 272 00:18:19,854 --> 00:18:28,569 it basically refused to do it and it did so on the basis of it was trying to be, here we go. 273 00:18:28,569 --> 00:18:29,189 Let's see. 274 00:18:29,189 --> 00:18:35,262 It says, I see the image shows two people in what appears to be a promotional blah, blah. 275 00:18:35,262 --> 00:18:35,965 However, 276 00:18:35,965 --> 00:18:48,165 I can't identify who specific people in images look like or compare them to famous individuals as this could involve making assumptions about people's identity based on 277 00:18:48,165 --> 00:18:49,385 their appearance. 278 00:18:49,405 --> 00:18:52,545 And I was like, all right, that's interesting. 279 00:18:52,545 --> 00:18:54,125 Let me go to ChatGPT. 280 00:18:54,125 --> 00:18:56,525 It's a little goofy, but okay, whatever. 281 00:18:56,525 --> 00:19:03,085 They're probably trying to solve some sort of a bias. 282 00:19:03,085 --> 00:19:05,361 or privacy thing or like you said, bias. 283 00:19:05,361 --> 00:19:08,939 It doesn't want to say that I look like somebody stereotypically a thing, whatever. 284 00:19:08,939 --> 00:19:09,510 Yeah. 285 00:19:09,510 --> 00:19:10,852 So I went to ChatGBT. 286 00:19:10,852 --> 00:19:12,905 ChatGBT just flat out refused. 287 00:19:12,905 --> 00:19:15,188 It just said, I cannot help you with this request. 288 00:19:15,188 --> 00:19:16,891 Gemini did it. 289 00:19:16,891 --> 00:19:20,207 I still don't know who the actor was, um unfortunately. 290 00:19:20,207 --> 00:19:22,027 So that's surprising when we were running. 291 00:19:22,027 --> 00:19:23,529 we're talking about refusal rates. 292 00:19:23,529 --> 00:19:26,822 When we were running ours, Jim and I had the highest refusal rate, right? 293 00:19:26,822 --> 00:19:27,703 Jim and I had it. 294 00:19:27,703 --> 00:19:32,717 And even for the nerds in the audience, even through the API, it would say, sorry, I can't do that. 295 00:19:32,717 --> 00:19:34,479 I can't, I can't engage in legal advice. 296 00:19:34,479 --> 00:19:40,724 Now subsequently, you know, and this is all obviously moving, moving terrain under our feet at all times. 297 00:19:40,724 --> 00:19:42,766 So you can do it one day and it'll refuse to do it the next day. 298 00:19:42,766 --> 00:19:45,869 We'll do it because they've dialed something back or down. 299 00:19:45,869 --> 00:19:47,864 But that was um for us, it was 300 00:19:47,864 --> 00:19:50,426 Gemini and Claude were the two highest refusal rates. 301 00:19:50,426 --> 00:19:52,097 ChatGPT was a little bit better. 302 00:19:52,097 --> 00:19:57,561 And even through the API, I just like, I'm not going to engage in, you know, essentially legal advice. 303 00:19:57,561 --> 00:20:00,273 And you know, what you realize is there was a time there. 304 00:20:00,273 --> 00:20:09,449 And I think that that time has probably passed where, know, if the state of California can sue legal zoom for forms and that's considered unauthorized practice of law, these are 305 00:20:09,449 --> 00:20:12,802 pretty uncomfortable with, you know, anything kind of approximating legal advice. 306 00:20:12,802 --> 00:20:14,783 Now, now they give legal advice, no problem. 307 00:20:14,783 --> 00:20:16,290 They'll issue a bunch of caveats. 308 00:20:16,290 --> 00:20:18,775 but they will give you lots of legal advice. 309 00:20:18,775 --> 00:20:24,327 I don't know what movement behind the scenes has happened, but they all kind of collectively said, we can give some legal advice. 310 00:20:24,327 --> 00:20:25,100 That's what it seems like. 311 00:20:25,100 --> 00:20:26,220 You should try your picture out again. 312 00:20:26,220 --> 00:20:26,821 Yeah. 313 00:20:26,821 --> 00:20:27,633 now it'll do it. 314 00:20:27,633 --> 00:20:28,079 Yeah. 315 00:20:28,079 --> 00:20:28,990 that's a good point. 316 00:20:28,990 --> 00:20:32,063 This was, this was probably a couple of months ago. 317 00:20:32,063 --> 00:20:36,308 So, yeah, there's not a date on this, but Grok did it, of course, which you would expect. 318 00:20:36,308 --> 00:20:39,282 Grok, not a lot of guard rails with Grok. 319 00:20:39,282 --> 00:20:45,067 And, you know, did, did you guys try Grok at all in this experiment? 320 00:20:45,192 --> 00:20:47,442 No, so we just stuck with those three. 321 00:20:47,442 --> 00:20:52,057 I just for a dumb reason, I just didn't have a GROCK subscription at that point so I can get to the API. 322 00:20:52,057 --> 00:20:53,188 So I didn't, I do now. 323 00:20:53,188 --> 00:20:55,370 And there's some stuff where GROCK is really useful. 324 00:20:55,370 --> 00:20:56,640 know, first of all, it's very fast. 325 00:20:56,640 --> 00:20:57,831 just in the chat interface. 326 00:20:57,831 --> 00:21:00,394 And then number two, it performs really, really well. 327 00:21:00,394 --> 00:21:09,761 think, you know, Elon Musk has kind of been in the political news so much that they don't report on his model the same way, but that it's a really interesting and powerful model. 328 00:21:09,761 --> 00:21:11,822 And the other thing is like, if you need 329 00:21:11,860 --> 00:21:19,204 You know, if you're on your phone and you need kind of a survey of kind of what is going on in social media, it's really useful because as Twitter data. 330 00:21:19,947 --> 00:21:22,087 I've done stock stuff for my wife with it. 331 00:21:22,087 --> 00:21:27,113 yeah, I was curious if there were any different results with Grok. 332 00:21:27,113 --> 00:21:32,297 Well, what about, um okay, so you tried the Frontier models, weren't successful. 333 00:21:32,297 --> 00:21:34,789 What was the next step? 334 00:21:34,789 --> 00:21:37,561 It sounds like you might've tried some open source models. 335 00:21:37,703 --> 00:21:39,445 Yeah, so I went on and eat. 336 00:21:39,445 --> 00:21:47,115 So I went on the leaderboard and kind of tried to find something that we could run locally so we could save token costs, of course, and was trying to look at, you know, less 337 00:21:47,115 --> 00:21:49,097 censored models was the idea. 338 00:21:49,097 --> 00:21:59,036 And actually, a handful of the models, even one of the Mistral models that was a French company, their Frontier Lab, and even one of those wouldn't do kind of law, advice, 339 00:21:59,036 --> 00:22:02,495 practice of law, replicate human experimentation type things. 340 00:22:02,495 --> 00:22:08,587 I was able to find a less censored version of Mistral that would do kind of whatever I told it to do, which was very useful. 341 00:22:08,587 --> 00:22:10,228 It was very useful for what we're doing. 342 00:22:10,228 --> 00:22:21,501 And so then we decided to try to take all of our real human responses, turn that into a JSON training set, and then fine tune that less censored model as just kind of like, will 343 00:22:21,501 --> 00:22:22,881 this do anything? 344 00:22:22,881 --> 00:22:26,332 And it ended up being much better than any of the other models that we did. 345 00:22:26,332 --> 00:22:28,621 So that was pretty interesting. 346 00:22:28,621 --> 00:22:29,981 And it was cheap too. 347 00:22:29,981 --> 00:22:34,661 So I remember I used to fine tune models kind of early in the chat, GPC revolution. 348 00:22:34,661 --> 00:22:36,001 And it was expensive. 349 00:22:36,001 --> 00:22:41,261 I mean, it would be too even for a small model, like a 24 billion parameter model would be hundreds of dollars. 350 00:22:41,261 --> 00:22:46,121 When I did this one, it was, I think it was $14 or something. 351 00:22:47,941 --> 00:22:49,321 I use a service. 352 00:22:49,321 --> 00:22:50,741 Well, I just use tokens for mine. 353 00:22:50,741 --> 00:22:52,121 So I did it with two different packages. 354 00:22:52,121 --> 00:22:54,061 There's an open source package called oxalotl. 355 00:22:54,061 --> 00:22:58,436 And then there's a service called open pipe, which is no, I guess no code. 356 00:22:58,452 --> 00:23:04,338 As long as you have the JSON training set, you can just throw it in there and you know, do whatever model and it'll also host it for inference, which is pretty cool. 357 00:23:04,338 --> 00:23:08,421 But initially I just did it at home on my home computer with axolotl. 358 00:23:08,883 --> 00:23:10,964 Open source, fine tuning package. 359 00:23:11,387 --> 00:23:12,969 did you try llama? 360 00:23:13,102 --> 00:23:14,822 No, we so we didn't do the llama. 361 00:23:14,822 --> 00:23:18,602 Oh, so we tested the llama model, but it ended up being even worse. 362 00:23:18,602 --> 00:23:24,382 And so I couldn't even I couldn't even fine tune it enough to get it to be anywhere near human level. 363 00:23:24,382 --> 00:23:29,353 And so I was like, all right, we have to find something less censored than llama, all the meta models. 364 00:23:29,353 --> 00:23:37,413 And then after the fine tuning, were, what, was the, how was the testing after, after you fine tuned? 365 00:23:37,838 --> 00:23:42,518 So we essentially just did the same thing that we did to the real humans and scaled it. 366 00:23:42,518 --> 00:23:45,358 I think we did it up to, was it 5,000? 367 00:23:45,358 --> 00:23:45,678 Yeah. 368 00:23:45,678 --> 00:23:46,478 It was 5,000. 369 00:23:46,478 --> 00:23:50,018 So instead of, we only had 1200 actual juror responses. 370 00:23:50,018 --> 00:23:59,698 So we scaled it up to 5,000 real human responses where we would say, you know, go read this, first take on this, you know, this demographic or take on a random demographic from 371 00:23:59,698 --> 00:24:07,158 the range present in the real jurors, then read the thing, then respond to the thing as if you are from, you know, said demographic. 372 00:24:07,247 --> 00:24:08,747 And then we just scaled it. 373 00:24:08,747 --> 00:24:14,655 it's just like, just ran my computer for like an evening, basically, just let it grind through 5,000 iterations. 374 00:24:14,655 --> 00:24:18,178 And then we took that and measured it against the real human responses. 375 00:24:18,395 --> 00:24:20,192 and how did they compare? 376 00:24:20,192 --> 00:24:27,512 Much better, much better, which it was one of those things where I wasn't sure if just fine tuning was going to be enough. 377 00:24:27,512 --> 00:24:29,672 It felt like something that was more. 378 00:24:29,672 --> 00:24:36,372 You might have to have like some kind of a genetic workflow and you know, frankly, we can't afford to pre train a model of our own. 379 00:24:36,372 --> 00:24:39,732 So I, I wasn't sure if just fine tuning would be enough. 380 00:24:39,732 --> 00:24:48,959 And then the other problem is when you're fine tuning a model, even a small model, like 24 billion parameters, we had 1200 Jason. 381 00:24:48,959 --> 00:24:52,639 data points, know, JSON rows or data points to use. 382 00:24:52,719 --> 00:24:58,099 Most people do like 50,000 and we'll see benefits up to 100,000 fine tuning. 383 00:24:58,099 --> 00:25:04,259 So we just thought it was possibly just way too small of a training set to have any, you know, measurable impact on it. 384 00:25:04,259 --> 00:25:10,999 And the fact that at 1200 it did was like, Whoa, Whoa, you know, this is something that again, it was fairly cheap. 385 00:25:10,999 --> 00:25:11,899 It could do it for free. 386 00:25:11,899 --> 00:25:15,959 If you had the right hardware, like the Mac studio that we're running all this stuff on, we can do it on that. 387 00:25:16,093 --> 00:25:17,925 So it was really interesting to go. 388 00:25:17,925 --> 00:25:22,947 was like one of those breaks, know, one of the million aha moments I seem to have every month with this technology. 389 00:25:22,947 --> 00:25:24,993 It was like, whoa, we can actually do this. 390 00:25:24,993 --> 00:25:26,646 You know, it's like what it felt like. 391 00:25:26,646 --> 00:25:31,206 And when you say much better, like give me a sense of magnitude. 392 00:25:31,206 --> 00:25:32,926 Like how, how did it do? 393 00:25:32,926 --> 00:25:46,466 I don't know if this was a percentage that you use to measure alignment, but like give me a sense of, it doesn't have to be the exact number, but was it two, two X three X better 394 00:25:46,466 --> 00:25:48,206 50 % better. 395 00:25:48,655 --> 00:25:55,106 It was, geez, I don't want to state it, because I had it right here, but my computer locked. 396 00:25:55,106 --> 00:25:59,408 It was, I wrote it down just so I could specifically say this. 397 00:25:59,408 --> 00:26:02,394 I remember reading it was significant. 398 00:26:02,394 --> 00:26:12,013 95 % accuracy and geez, like essentially the margin of error was half, half of what it was with the model, the frontier models. 399 00:26:12,013 --> 00:26:17,839 So it was, you know, twice as good effectively as chat GPT trying to do this on its own. 400 00:26:17,839 --> 00:26:24,385 And so, and again, like, you know, there was a lot of technical things that I didn't do that I would do differently. 401 00:26:24,385 --> 00:26:27,188 Now I would use a different style of fine tuning if I was going to do it. 402 00:26:27,188 --> 00:26:28,330 We would have, you know, 403 00:26:28,330 --> 00:26:29,950 even scale with synthetic data. 404 00:26:29,950 --> 00:26:36,870 So that was something that I thought about, but she didn't want to, uh, cause a lot of people had just have a reflexive bad reaction to synthetic data. 405 00:26:36,870 --> 00:26:44,830 But had we taken that and then scaled it up to, know, 50,000, you know, take 12, take 1200 and then use that to scale up to 50,000. 406 00:26:44,830 --> 00:26:46,530 And then they use that to train a model. 407 00:26:46,530 --> 00:26:49,210 It probably would have been substantially better even then. 408 00:26:49,210 --> 00:26:56,710 So, uh, you know, just for a small investment and a small amount of training data, we got pretty big wins. 409 00:26:56,710 --> 00:26:58,077 And that was like the real like, 410 00:26:58,077 --> 00:26:59,601 breakthrough, I guess, of the paper. 411 00:26:59,601 --> 00:27:01,435 Especially as compared to the platform. 412 00:27:01,435 --> 00:27:07,962 Yeah, and definitely substantially better if you just use like the API to Gemini or Claude or Google or OpenAI. 413 00:27:07,962 --> 00:27:08,313 Yeah. 414 00:27:08,313 --> 00:27:10,835 And what conclusions can we draw from this? 415 00:27:10,835 --> 00:27:24,621 So back to my earlier comment about there's a, it is a constant point of discussion in the legal tech community about are these legal specific solutions going to get steamrolled by 416 00:27:24,621 --> 00:27:26,322 the frontier models? 417 00:27:26,362 --> 00:27:33,916 And your study makes me think maybe, maybe not, because if the frontier models are 418 00:27:33,916 --> 00:27:41,270 have controls, we'll call them, in place that make them a Swiss army knife where they're good for a lot of things. 419 00:27:41,270 --> 00:27:46,953 So they're a mile wide and an inch deep, as opposed to being an inch wide and a mile deep. 420 00:27:47,134 --> 00:27:57,720 It seems like, I don't know if it's a fair conclusion or not, but your study for niche legal use cases, the frontier models might not be the best solution. 421 00:27:57,720 --> 00:28:00,341 I don't know, Haley, what's your thought on that? 422 00:28:00,562 --> 00:28:03,143 think that's exactly right, at least as of now. 423 00:28:03,143 --> 00:28:14,755 I think that our project really proves the need to have these models specifically trained on human data, human judgment in the realm of evaluating evidence in these legal 424 00:28:14,755 --> 00:28:15,415 scenarios. 425 00:28:15,415 --> 00:28:23,829 I think it could be huge in terms of, as you mentioned earlier, trial prep, things that lots of trial boutiques would really latch onto and utilize. 426 00:28:23,829 --> 00:28:25,699 But right now they're just not there. 427 00:28:25,699 --> 00:28:29,310 And I think it is because of that Swiss army knife nature. 428 00:28:29,478 --> 00:28:33,250 But I do think our fine tuning model shows that it is possible. 429 00:28:33,250 --> 00:28:44,038 So with lots more data and lots more training, there could be an essentially mock jury bot at some point in the future that could be reliable for trial prep and strategy and a lot 430 00:28:44,038 --> 00:28:44,789 of other things. 431 00:28:44,789 --> 00:28:48,792 Yeah, I know, right now the big the frontier models obviously have the data problem. 432 00:28:48,792 --> 00:28:55,457 There's only four good places you can get, you know, US legal data, Vlex, now Clio, guess, Vlex still, Westlaw, Lexus, Bloomberg. 433 00:28:55,457 --> 00:28:57,096 Now those are the only data sets. 434 00:28:57,096 --> 00:29:04,192 and they haven't purchased one of those yet, which probably indicates, which, you when I hear people talk about it, it's that they just don't think it's a big enough market right 435 00:29:04,192 --> 00:29:04,401 now. 436 00:29:04,401 --> 00:29:06,254 They're going after literally everyone. 437 00:29:06,254 --> 00:29:14,400 But if at some point they do purchase one of those data sets and have that on the backend, then that's really gonna blow up the business model of like a Lexus or Westlaw or Vlex, 438 00:29:14,400 --> 00:29:22,352 unless they're offering something else in addition to it, some sort of agentic workflow, some sort of tool that plugs into your firm, which I think is. 439 00:29:22,352 --> 00:29:27,216 why probably the Clio Vlex acquisition is really interesting to me right now. 440 00:29:27,216 --> 00:29:31,170 But right now, the frontier models have a data problem and they're just not focused on legal. 441 00:29:31,170 --> 00:29:38,045 They're too big for legal, which is crazy to think about with the 1.5 billion or whatever that's going into legal investment this year. 442 00:29:38,045 --> 00:29:41,659 It's like, no, we're still considered small dollars to them, I think. 443 00:29:41,659 --> 00:29:46,170 Yeah, well, and honestly, and the industry is extremely fragmented. 444 00:29:46,170 --> 00:29:51,452 And if you add up the entire Amlaw 100 revenue, it's $140 billion. 445 00:29:51,452 --> 00:29:55,694 Like that's, that would be like a fortune 100 company. 446 00:29:55,694 --> 00:29:58,689 So that's the entire, that's the top 100. 447 00:29:58,835 --> 00:30:05,598 So it is, it is a relatively small industry, when compared to others, but you know, 448 00:30:05,778 --> 00:30:09,040 There was an AI company that did try to buy one of the information providers. 449 00:30:09,040 --> 00:30:12,450 Harvey made a play for Vlex, I think as part of their series. 450 00:30:12,450 --> 00:30:17,865 They were trying to raise, yeah, five, 600 million to buy Vlex and that didn't happen. 451 00:30:17,865 --> 00:30:18,590 And then... 452 00:30:18,590 --> 00:30:26,445 I know some people who work at like, you know, the open legal data sets that are out there, some of the more free ones, and they've been approached by the likes of open AI for 453 00:30:26,445 --> 00:30:26,995 purchasing. 454 00:30:26,995 --> 00:30:31,006 So they don't have a complete set, but they have a really good set that could be really useful. 455 00:30:31,006 --> 00:30:38,058 You know, the benefit of having the Lexus and Westlaw is you've got all the secondary materials, which are beautiful for training, uh even on, you know, very niche kind of 456 00:30:38,058 --> 00:30:39,378 specific areas of the law. 457 00:30:39,378 --> 00:30:44,140 You've got this really good rich data that's curated by attorneys that have worked in the field for a billion years and whatever. 458 00:30:44,140 --> 00:30:44,831 So. 459 00:30:44,831 --> 00:30:45,942 That is the benefit of those ones. 460 00:30:45,942 --> 00:30:50,213 You Felix went the really interesting opposite route, which is we don't have the secondary materials. 461 00:30:50,213 --> 00:30:53,889 We're just going to make it all a Gentic and we're going to craft you an on-demand treatise. 462 00:30:53,889 --> 00:30:59,763 So I think, you know, unless they can get ahold of Westlar or Lexis, I can't imagine Westlar wants to sell ever. 463 00:30:59,763 --> 00:31:01,624 You know, their data is there. 464 00:31:01,624 --> 00:31:05,387 That is their gold mine that they're sitting on, you know, Thompson writers. 465 00:31:05,387 --> 00:31:08,509 But if they can get ahold of Lexis, that would be some pretty interesting content. 466 00:31:08,509 --> 00:31:11,961 And, know, Harvey now has the API to Lexis. 467 00:31:11,961 --> 00:31:13,252 So I don't know. 468 00:31:13,252 --> 00:31:13,582 see. 469 00:31:13,582 --> 00:31:14,305 We'll see. 470 00:31:14,305 --> 00:31:18,674 Yeah, I've heard that integration is very surface level. 471 00:31:18,674 --> 00:31:19,961 um 472 00:31:19,961 --> 00:31:20,932 struck me as strange. 473 00:31:20,932 --> 00:31:26,681 think a lot of techie people, I think it struck us as strange because it's like, know, what you want is good structured data. 474 00:31:26,681 --> 00:31:33,520 You know, it's like saying, here's this big gulp and here's one those little coffee straws to drink out of it, you know, with the API. 475 00:31:33,520 --> 00:31:37,154 So it's like, ah maybe they'll fix that in the future. 476 00:31:37,154 --> 00:31:37,667 I don't know. 477 00:31:37,667 --> 00:31:42,469 Yeah, it's and you know, this brings up another interesting question and something I talk about a lot. 478 00:31:42,469 --> 00:31:57,285 So I'm of the opinion that in the future, let's say five years from now, once we have a enter the tech enabled legal service delivery era, um, I'm a big believer that law firms 479 00:31:57,285 --> 00:32:02,138 are going to have to leverage their data and in order to differentiate, right? 480 00:32:02,138 --> 00:32:04,329 Buying an off the shelf tool is not going to 481 00:32:04,329 --> 00:32:08,160 to differentiate you because your competitor can go buy the same tool. 482 00:32:08,160 --> 00:32:10,421 But you know what is unique to you? 483 00:32:10,421 --> 00:32:15,812 All of those documents that provided winning outcomes for your clients, right? 484 00:32:15,812 --> 00:32:30,569 All the precedent libraries of model documents that have been battle tested and the way legal works today, it's a very bespoke industry and everybody kind of has their own. 485 00:32:30,705 --> 00:32:39,770 You can ask 10 different lawyers at the same law firm in the same practice for the same document and get different iterations of it. 486 00:32:39,811 --> 00:32:44,453 So, you know, I, it makes me wonder something's got to give there, right? 487 00:32:44,453 --> 00:32:55,391 Like if we're going to tech enable legal service delivery, how, how we can't train models to be reflect one lawyers. 488 00:32:55,897 --> 00:33:00,299 perspective, even though that is what Crosby's doing, by the way, I don't know if you know about Crosby AI. 489 00:33:00,299 --> 00:33:11,494 Um, they, they're, I think it was Sequoia is one of the VCs in their cap table and they, they have a podcast called training data and I listened to that and that's, that is what 490 00:33:11,494 --> 00:33:11,983 they're doing. 491 00:33:11,983 --> 00:33:14,295 They just, but they just have a handful of attorneys. 492 00:33:14,295 --> 00:33:20,737 So they are literally training the attorneys, if I understood it correctly, um, to think like that lawyer. 493 00:33:21,028 --> 00:33:24,611 and that's, that's great and makes a lot of sense as 494 00:33:24,611 --> 00:33:26,332 this transitional phase, right? 495 00:33:26,332 --> 00:33:29,574 Like, I don't think that scales, right? 496 00:33:29,574 --> 00:33:34,256 It's expensive and it's one-off and what happens if they lateral? 497 00:33:34,256 --> 00:33:46,294 they take, you know, it's just, so I don't know, what is your thought on kind of the future and how much law firm data is going to come into play and how firms differentiate 498 00:33:46,294 --> 00:33:47,284 themselves? 499 00:33:47,491 --> 00:33:49,131 I think it's gonna be super interesting. 500 00:33:49,131 --> 00:33:55,411 you know, I think, I went to, once I was, I had a consulting appointment at a law firm. 501 00:33:55,411 --> 00:34:02,871 They wanted to implement something very simple, which is just a rag pipeline for all their pleadings, you know, so that they could generate stuff. 502 00:34:03,031 --> 00:34:05,111 And I went over and I said, okay, where's your data? 503 00:34:05,111 --> 00:34:07,191 And they said, well, some of it's in OneDrive. 504 00:34:07,191 --> 00:34:08,451 And I said, okay. 505 00:34:08,511 --> 00:34:11,771 And some of it's, you know, some of it's in NetDocs. 506 00:34:11,771 --> 00:34:12,771 I said, okay. 507 00:34:12,771 --> 00:34:19,071 And they said, some of it's on this old server that we had, you know, like in a box is that some of it's just paper on a shelf. 508 00:34:19,071 --> 00:34:24,491 But, then when you got to the, got to the planes themselves, all different formats, no good metadata, blah, blah. 509 00:34:24,491 --> 00:34:32,091 And I'm like, it's just going to take a lot to get your, you know, your information into a place where you could actually ingest it into any of these AI systems. 510 00:34:32,171 --> 00:34:35,211 So I think that's obviously step one is just getting 511 00:34:35,211 --> 00:34:43,219 a lot of law firms which have kind of lagged behind just modern data practices, get it all into a data lake or whatever so you can actually make use of this thing is going to be a 512 00:34:43,219 --> 00:34:43,959 big part of it. 513 00:34:43,959 --> 00:34:52,506 But I, you know, I think you're right that there is going to definitely be a period of time where there is going to be a custom tool. 514 00:34:52,506 --> 00:34:56,929 You Joe only do immigration law here is your custom tool that uses your internal things. 515 00:34:56,929 --> 00:35:00,092 and by the way, fine tuning now costs $14 or whatever, you know. 516 00:35:00,092 --> 00:35:09,732 So we can fine tune your little tiny model that just sits on prem on your stuff and really understands your workflow and who Ted is in the context of your firm and all of those 517 00:35:09,732 --> 00:35:10,392 things. 518 00:35:10,392 --> 00:35:12,072 I think there will be a while for that. 519 00:35:12,072 --> 00:35:21,232 think at some point it's probably gonna get eaten up by the big guys where they're gonna, know, VLex is already launching the ability to create your own workflows for your firm. 520 00:35:21,232 --> 00:35:24,892 Harvey has some sort of version of that from what I understand. 521 00:35:24,892 --> 00:35:29,692 So I think they're gonna roll out tools that are probably good for 70 % of the market. 522 00:35:30,054 --> 00:35:40,079 But there may be this place on the margins where you can outperform them substantially by using your own data and leveraging it within an internal ecosystem that is good for AI. 523 00:35:40,079 --> 00:35:43,503 So we were going to talk about a couple of things here. 524 00:35:43,503 --> 00:35:50,960 I want to skip down to legal education because I had a conversation this morning with Ed Walters from Felix. 525 00:35:51,602 --> 00:35:53,143 Yeah, he's a great guy. 526 00:35:53,143 --> 00:35:55,646 He's actually going to be on our 100th episode. 527 00:35:55,646 --> 00:35:57,103 um 528 00:35:57,103 --> 00:35:58,084 He's always a good get. 529 00:35:58,084 --> 00:35:59,785 Always has something interesting to say. 530 00:35:59,805 --> 00:36:01,107 And he's really nice. 531 00:36:01,107 --> 00:36:02,489 So I see him at conferences. 532 00:36:02,489 --> 00:36:05,331 He's just a really, really nice guy too. 533 00:36:06,553 --> 00:36:07,101 Yeah. 534 00:36:07,101 --> 00:36:12,405 we were talking a little bit about lawyer training and you guys are at the forefront of that. 535 00:36:12,566 --> 00:36:27,159 So I'm curious, Haley, like how do you think about how in an academic setting, lawyer training is going to need to evolve from where it is today to where we may be heading 536 00:36:27,159 --> 00:36:29,587 three, four years ago when the 537 00:36:29,587 --> 00:36:37,629 blocking and tackling work is really automated and document generation and we need to elevate lawyers to be more consultants. 538 00:36:37,629 --> 00:36:42,185 Like, how are you guys thinking about it at the law school or are you? 539 00:36:42,185 --> 00:36:45,089 Is it kind of a wait and see situation? 540 00:36:45,089 --> 00:36:45,877 What's that look like? 541 00:36:45,877 --> 00:36:47,618 think it depends on who you talk to. 542 00:36:47,618 --> 00:36:49,820 There's definitely resistance. 543 00:36:49,820 --> 00:36:57,926 I'm a relatively new professor and obviously I like AI and new things and so I think I'm a little bit more forward thinking about all of these things. 544 00:36:57,926 --> 00:37:10,600 But there is a very serious discussion that I think everyone is like engaging in about our students and their future because what first, second, third year attorneys usually do, 545 00:37:10,600 --> 00:37:15,505 Doc review, things like that are all the things that AI is going to be able to do in two seconds, right? 546 00:37:15,505 --> 00:37:22,882 And so what I think is really important for us as educators is to train students to know how to use these tools. 547 00:37:22,882 --> 00:37:27,996 They can come in as the people that can train all the other people in the law firms, right? 548 00:37:27,996 --> 00:37:30,180 So they can bring value that way. 549 00:37:30,180 --> 00:37:36,737 So like in my classes, I don't mind if students use AI to help them with practice problems, generate potential. 550 00:37:36,737 --> 00:37:40,449 multiple choice questions, essay answers, tell them what they did right or wrong. 551 00:37:40,449 --> 00:37:42,101 I encourage them to do that. 552 00:37:42,101 --> 00:37:51,466 And I think, you know, it's pretty similar, you know, even if AI is helping you generate your response to something, like that's what's essentially going on in law firms right now 553 00:37:51,466 --> 00:37:52,487 anyway, right? 554 00:37:52,487 --> 00:37:55,218 There's shell pleadings, motions. 555 00:37:55,218 --> 00:38:04,143 No one ever starts from scratch if you're in a law firm like that, although their data might be in boxes or on old servers, em as Sean recently found out, right? 556 00:38:04,201 --> 00:38:09,193 But I think it's coming no matter what we want to do about it or how we think about it. 557 00:38:09,193 --> 00:38:18,988 And so my general approach is just to encourage students to learn as much as they can, go to the digital initiatives with Sean that he puts on with Kitten all the time, which is 558 00:38:18,988 --> 00:38:20,559 really great for students. 559 00:38:20,559 --> 00:38:27,153 And I'm trying to learn as well so that I can be a resource for them to put them in the best position possible to succeed. 560 00:38:27,153 --> 00:38:29,266 know that Deans are thinking about it for sure. 561 00:38:29,266 --> 00:38:34,671 Because a huge portion of their your US World News ranking is based on your employment data of your students. 562 00:38:34,671 --> 00:38:43,129 And so we're finally, you know, people have been predicting these like catastrophic job layoffs since Chad GVT came around, we're finally now getting studies that are showing 563 00:38:43,129 --> 00:38:44,110 that this is really happening. 564 00:38:44,110 --> 00:38:54,499 I think Accenture just laid off 11,000 people there was that big study that came out of I think it was Stanford that had 62 million back end data points from the payroll system. 565 00:38:54,499 --> 00:38:59,899 that showed a 13.2 % decline in hiring among younger people. 566 00:39:00,159 --> 00:39:10,319 I, know, with the kind of traditional law school or law firm pyramid where this bottom layer of people, this document review discovery, simple, you know, kind of simple drafting 567 00:39:10,319 --> 00:39:14,639 research, like AI does all of those things very well and collapses it down. 568 00:39:14,639 --> 00:39:19,331 you know, I think when we think of law firms going forward, it's either going to be kind of a cylinder. 569 00:39:19,331 --> 00:39:25,931 or even a diamond shape is what other people have predicted, which I think is pretty interesting, but it raises a lot of problems for law schools. 570 00:39:26,191 --> 00:39:35,351 the way that I've thought about filling in what are those people gonna go do if they're not going to work at Kirkland, Junior's, Jackson Walker, McAfee Taft, one of the big firms 571 00:39:35,351 --> 00:39:37,731 down here, go and don't leave me where your mom works. 572 00:39:38,031 --> 00:39:47,171 I think non-traditional careers, you I kind of think we're headed for probably a 2008, 2009 type scenario where they're just not hiring first year associates as much. 573 00:39:47,202 --> 00:39:54,178 And so my impulse is to say to my students, hey, especially the ones my AI in the practice of law class, you know, it'd be really cool career legal ops. 574 00:39:54,619 --> 00:40:02,766 And that's not something that law schools traditionally use legal leverage their career services and send people to clock and talk to people you don't like you who are really 575 00:40:02,766 --> 00:40:03,537 into these areas. 576 00:40:03,537 --> 00:40:09,313 And I'm like, that's a great lifestyle, you know, you don't have to go work at a firm, you can still make great money. 577 00:40:09,313 --> 00:40:12,260 And if you're a techie person, like you know me or 578 00:40:12,260 --> 00:40:15,374 you know, like when I was graduating from law school, wouldn't want to go do document review at big firm. 579 00:40:15,374 --> 00:40:20,931 I would have loved to have done something like that instead, but I didn't know that was an option and it wasn't the traditional path for law students. 580 00:40:20,931 --> 00:40:28,592 So I think we need to be a little more creative and hopefully we're setting up the students that come through here so that they could go right into those careers and be 581 00:40:28,592 --> 00:40:29,713 really effective. 582 00:40:29,779 --> 00:40:30,119 Yeah. 583 00:40:30,119 --> 00:40:37,133 So, you I've been an entrepreneur for 32 years and during that time I have consumed a lot of legal services. 584 00:40:37,133 --> 00:40:47,450 And one of the big challenges I've talked about it once or twice on the show before is Lawyers are really good at identifying risk. 585 00:40:47,450 --> 00:40:50,613 as a business person, that's only half the equation. 586 00:40:50,613 --> 00:40:56,322 is a, almost every decision is a risk reward or a cost benefit. 587 00:40:56,322 --> 00:40:57,733 decision, right? 588 00:40:57,733 --> 00:40:59,114 You're balancing those scales. 589 00:40:59,114 --> 00:41:08,250 And lawyers historically have not taken the time to get to know my business because I'd have to pay them to do that, right? 590 00:41:08,250 --> 00:41:09,701 And that would be really expensive. 591 00:41:09,701 --> 00:41:19,128 I understand why they haven't, but you know, like that's always been a challenge from where I sit because I've had lawyers, you know, basically advise me, don't sign this 592 00:41:19,128 --> 00:41:19,888 agreement. 593 00:41:19,888 --> 00:41:21,959 And I'm like, you can't say that. 594 00:41:22,016 --> 00:41:25,406 What you can tell me is you can inform me on the risk side of the equation. 595 00:41:25,406 --> 00:41:31,008 I, as the business person, need to balance the business opportunity, i.e. 596 00:41:31,008 --> 00:41:33,709 the reward side of the equation. 597 00:41:33,709 --> 00:41:35,769 What you're bringing to me is very important. 598 00:41:35,769 --> 00:41:37,770 It helps me balance the scales. 599 00:41:37,810 --> 00:41:46,905 So what I would love to see as a consumer of legal services is lawyers get more in an advisory capacity and bring the human element. 600 00:41:46,905 --> 00:41:48,247 I'll give you an example. 601 00:41:48,247 --> 00:41:59,690 So we hired someone away from a, they're not really a competitor, but a company who had a non-compete and we had to craft his job description in a way where we wouldn't run a foul 602 00:41:59,690 --> 00:42:00,931 of that non-compete. 603 00:42:00,931 --> 00:42:06,056 And we hired a L &E attorney and he gave great advice and it wasn't really legal. 604 00:42:06,056 --> 00:42:09,662 He's like, look, you know, the reality is, you know, for the first, 605 00:42:09,662 --> 00:42:13,462 X number of months, there's probably going to be a lot of scrutiny around this. 606 00:42:13,462 --> 00:42:22,162 And then over time that will probably, you know, there, there are a bill, a judge wouldn't look kindly if six months down the road, they, they threw a flag, right? 607 00:42:22,162 --> 00:42:23,842 So he was giving me human advice. 608 00:42:23,842 --> 00:42:27,002 Like, you know, this is how the company is going to look at this. 609 00:42:27,002 --> 00:42:31,882 This is how the judge would look at an action that came six months later. 610 00:42:31,882 --> 00:42:37,082 Like that sort of stuff is going to be really hard to get out of chat GPT. 611 00:42:37,082 --> 00:42:38,382 It requires 612 00:42:38,458 --> 00:42:39,730 Experience is it possible? 613 00:42:39,730 --> 00:42:42,724 I guess I right 614 00:42:42,836 --> 00:42:48,489 you have to fill up the current 1 million token context window in order to have all that context if then, you know, even. 615 00:42:48,489 --> 00:43:01,557 Yeah, and even then, this guy was a seasoned vet and had spent 30 years on plaintiff side cases. 616 00:43:01,557 --> 00:43:10,102 And again, how you load that into an AI model and get solid judgment back, I think, presents challenges. 617 00:43:10,102 --> 00:43:11,825 uh 618 00:43:11,825 --> 00:43:16,207 when they survey, they survey, you know, our customers from law firms, they all hate us. 619 00:43:16,207 --> 00:43:17,807 They hate to see us. 620 00:43:17,807 --> 00:43:18,888 They hate to talk to us. 621 00:43:18,888 --> 00:43:26,162 They hate to pick up the phone in advance, which is, know, in many instances would remediate a lot of risk if they could just, hey, Sean, what about this? 622 00:43:26,162 --> 00:43:27,332 They hate doing all that stuff. 623 00:43:27,332 --> 00:43:35,445 So, I mean, that's why when you hear it, like a lot of the access to justice, people talk about this technology, the more optimistic ones, they go look, 78 % of the market doesn't 624 00:43:35,445 --> 00:43:38,956 have access to a civil attorney, an attorney in a civil context. 625 00:43:38,956 --> 00:43:40,633 If this drops the cost. 626 00:43:40,633 --> 00:43:50,178 cost and it allows a whole nother layer of people to have better legal outcomes or God forbid uh makes us have just better interactions with our clients so that they want to 627 00:43:50,178 --> 00:43:50,969 come talk to us. 628 00:43:50,969 --> 00:43:55,507 They enjoy picking up the phone because they feel more reassured about their business or whatever. 629 00:43:55,507 --> 00:43:57,192 You know, there is a lot of promise with this stuff. 630 00:43:57,192 --> 00:44:02,525 I try not to be completely starry eyed and rose colored glasses about it, but I'm just an optimistic person. 631 00:44:02,525 --> 00:44:08,749 I think there are a lot of wins here or just our interactions with our clients and better legal outcomes for regular people. 632 00:44:08,749 --> 00:44:09,749 think too. 633 00:44:10,078 --> 00:44:10,258 Yeah. 634 00:44:10,258 --> 00:44:15,169 And you have to ask yourself, why did, why do they hate answering the phone or calling the lawyers? 635 00:44:15,169 --> 00:44:17,631 You know, and, I actually don't hate it. 636 00:44:17,631 --> 00:44:19,552 Uh, what I do hate. 637 00:44:19,552 --> 00:44:27,475 Yeah, no, I mean, you know, we've got, we've got, you know, probably the number one startup law firm in the country. 638 00:44:27,475 --> 00:44:29,216 You can figure out who that is. 639 00:44:29,216 --> 00:44:35,179 And, it is, it is an education process every time I get to, talk to my counterpart over there. 640 00:44:35,179 --> 00:44:36,683 And I learned so much. 641 00:44:36,683 --> 00:44:37,393 And he's great. 642 00:44:37,393 --> 00:44:40,966 And he does think about things in risk reward. 643 00:44:40,966 --> 00:44:43,947 And that's why he's my attorney. 644 00:44:43,947 --> 00:44:46,209 And I enjoy picking up the phone. 645 00:44:46,209 --> 00:44:48,229 I don't enjoy getting the bills. 646 00:44:49,370 --> 00:44:54,174 That part's not fun, but I do feel like I get a lot of value from it. 647 00:44:54,174 --> 00:44:56,415 I've not always felt that way. 648 00:44:56,436 --> 00:45:00,378 And I don't know if this is a function, cause I've been in a lot of different businesses. 649 00:45:00,378 --> 00:45:02,779 My wife and I own five gyms here in St. 650 00:45:02,779 --> 00:45:03,528 Louis. 651 00:45:03,528 --> 00:45:06,519 fair amount of legal work that went into building that those out. 652 00:45:06,519 --> 00:45:07,859 I owned a collection agency. 653 00:45:07,859 --> 00:45:09,898 used to sue people all the time. 654 00:45:09,898 --> 00:45:11,461 A lot of legal work there. 655 00:45:11,461 --> 00:45:18,624 I've not always enjoyed it and I'm not sure if it's because these guys are super niche and they focus on venture backed startups like we are. 656 00:45:18,624 --> 00:45:25,206 But they really know our business and I learned something every time I talked to them and I enjoy it. 657 00:45:25,692 --> 00:45:28,436 Yeah, and there's a whole, you know, so I think there's two parts to that. 658 00:45:28,436 --> 00:45:36,043 Number one, some lawyers when you talk to them are sort of like when I talk to my IT department, excuse me, and I go, hey, I need to download this really interesting open 659 00:45:36,043 --> 00:45:43,788 source, completely unvetted software, because I need to keep my skills, you know, hot and, you know, up to the bleeding edge. 660 00:45:43,788 --> 00:45:45,229 And I want to use a cool new thing. 661 00:45:45,229 --> 00:45:46,460 And they just go, absolutely not. 662 00:45:46,460 --> 00:45:48,641 have no risk, no risk tolerance whatsoever. 663 00:45:48,641 --> 00:45:53,366 Because what's the incentive to them if I can use a new AI thing? 664 00:45:53,366 --> 00:45:54,546 Zero. 665 00:45:54,746 --> 00:45:58,988 What's the risk if I huge, you know, and so they just dial up risk. 666 00:45:58,988 --> 00:46:07,350 And I'm like, well, I have other interests, like remaining relevant or, you know, with library resources that happens very frequently with a thing we will not put behind 10 667 00:46:07,350 --> 00:46:09,431 different walls before the user can get to it. 668 00:46:09,431 --> 00:46:13,732 And a librarian goes, nobody's going to use it if it's a hassle to get in and use the thing. 669 00:46:13,732 --> 00:46:19,623 So it's just like, like you said, I think just the incentives are just pointed in the wrong direction for some lawyers and law firms. 670 00:46:19,623 --> 00:46:20,907 Yeah. 671 00:46:20,907 --> 00:46:23,667 Do you have anything to add to AI stuff I was talking about? 672 00:46:23,667 --> 00:46:24,987 Sorry, I got a pun intended. 673 00:46:25,107 --> 00:46:33,007 it just kind of reminded me when you're talking about the experience that the labor and employment attorney had that you found really valuable. 674 00:46:33,007 --> 00:46:43,227 I was a law clerk for a few judges and tons of firms and clients found it very valuable because I worked in judges chambers for years at a time. 675 00:46:43,227 --> 00:46:49,379 I understood how the judges I worked for thought, how their chambers operated, all this information that 676 00:46:49,385 --> 00:46:50,936 AI can't get its hands on, right? 677 00:46:50,936 --> 00:47:01,722 And so that was value I could bring to clients that is aside from document review and, you know, it's advice that can actually be beneficial that you're not going to get from AI or 678 00:47:01,722 --> 00:47:02,933 even every attorney. 679 00:47:02,933 --> 00:47:08,887 And so there are ways I think that attorneys just in general can utilize things that AI just can't do. 680 00:47:08,887 --> 00:47:11,148 Now, to me, though, it brings up this issue. 681 00:47:11,148 --> 00:47:13,799 You know, you're talking about the 30 years of experience. 682 00:47:13,799 --> 00:47:16,681 Well, if attorneys can't even get in the door. 683 00:47:16,681 --> 00:47:25,926 you know, after law school, because AI has taken over all of their tasks, we're going to end up with a huge shortage of attorneys at some point that have have any experience and 684 00:47:25,926 --> 00:47:35,261 then that you know, value might be lost if there's not a way to kind of have attorneys recalibrate like what they bring to the table and then get that experience as years go on. 685 00:47:35,261 --> 00:47:42,585 Yeah, I some people have, you know, when I go to other conferences, like Dan Lena was from Northwestern was talking about this where he was saying, you know, essentially, what we're 686 00:47:42,585 --> 00:47:45,256 going to see is just attorneys just highly specialized. 687 00:47:45,256 --> 00:47:53,221 So I'm an expert in e-discovery, in transactional, in California, for these real estate, you know, whatever, just really, really narrowly specialized. 688 00:47:53,221 --> 00:47:55,912 And so maybe that would be a situation like yours. 689 00:47:55,912 --> 00:47:59,514 You now got this attorney who he has dialed into your exact sector. 690 00:47:59,514 --> 00:48:03,617 You know that he's competent, you know, he has the right level of risk tolerance for you. 691 00:48:03,617 --> 00:48:07,279 And he gives you this white glove service that's very customized to Ted. 692 00:48:07,279 --> 00:48:13,054 And so maybe it'll just be that kind of times everything in law as what he, you know, they've hypothesized. 693 00:48:13,054 --> 00:48:14,950 They say the riches are in the niches. 694 00:48:14,950 --> 00:48:16,273 So, um, 695 00:48:16,273 --> 00:48:16,864 what I my students. 696 00:48:16,864 --> 00:48:18,296 I said, you know, this is a great time. 697 00:48:18,296 --> 00:48:24,403 Like you want lost services Chipotle burrito, you know, you want extra guac and so Fritas instead of carnitas. 698 00:48:24,403 --> 00:48:26,115 Let me get you exactly what you want, sir. 699 00:48:26,115 --> 00:48:28,068 That expensive burrito. 700 00:48:28,068 --> 00:48:29,930 Deliver it to your table for you. 701 00:48:29,930 --> 00:48:31,571 And here's the drink that you enjoy. 702 00:48:31,571 --> 00:48:33,665 Well, this has been a great conversation. 703 00:48:33,665 --> 00:48:38,031 really appreciate you guys coming and spending a little bit of time with me on the show. 704 00:48:38,031 --> 00:48:39,147 um 705 00:48:39,147 --> 00:48:42,850 got to kick the tires on our new podcast studio that Jim put together for us. 706 00:48:42,850 --> 00:48:44,211 Yeah, it's been great. 707 00:48:44,211 --> 00:48:44,611 Thank you. 708 00:48:44,611 --> 00:48:46,323 Awesome. 709 00:48:46,323 --> 00:48:47,183 I'm really impressed with it. 710 00:48:47,183 --> 00:48:49,543 I'm gonna do all my big presentations in here from now on. 711 00:48:49,543 --> 00:48:50,796 Yeah, you should. 712 00:48:50,796 --> 00:48:57,030 How do folks find out more about the work that you're doing there at the law school and you individually? 713 00:48:57,030 --> 00:48:58,863 What's the best way for them to do that? 714 00:48:59,111 --> 00:49:03,126 Probably for me, LinkedIn is kind of where I consolidate all my professional stuff. 715 00:49:03,126 --> 00:49:06,059 And it's, I'll have links to like my SSRN articles. 716 00:49:06,059 --> 00:49:09,102 And when I presented a thing, if I'm going to Ulta and presenting or whatever. 717 00:49:09,102 --> 00:49:09,793 Yeah. 718 00:49:09,793 --> 00:49:12,716 And then we're both, we both have pages on OU Laws website. 719 00:49:12,716 --> 00:49:17,470 So you can just search for our name, all of our papers and work will kind of be linked there as well. 720 00:49:17,470 --> 00:49:18,230 Okay. 721 00:49:18,230 --> 00:49:18,820 Well, awesome. 722 00:49:18,820 --> 00:49:22,332 We'll include those links in the show notes so folks can access them. 723 00:49:22,332 --> 00:49:28,893 And thank you so much for being on the show and hopefully we get to meet in person at a future conference. 724 00:49:30,974 --> 00:49:32,085 You as well. -->

Subscribe

Stay up on the latest innovations in legal technology and knowledge management.