This company is either run by someone who doesn't understand the tech or is willfully fraudulent. ChatGPT and company are far from good enough to be entrusted with law. Having interacted extensively with modern LLMs, I absolutely know something like this would happen:
> Defendant (as dictated by AI): The Supreme Court ruled in Johnson v. Smith in 1978...
> Judge: There was no case Johnson v. Smith in 1978.
LLMs hallucinate, and there is absolutely no space for hallucination in a court of law. The legal profession is perhaps the closest one to computer programming, and absolute precision is required, not a just-barely-good-enough statistical machine.
Pretty sure the whole reason why DoNotPay actually exists is because defending against parking tickets didn't actually require a strong defense. The tickets were flawed automation, and their formulaic nature justified and equally formulaic response, or something to that effect. Whether the LLM was actually going to output answers directly, or just be used to drive a behavior tree or something like that, is a question I don't see answered anywhere.
That said, if it's such a catastrophically stupid idea, I'm not really sure why it had to be shot down so harshly: seems like that problem would elegantly solve itself. I assume the real reason it was shot down was out of fear that it would work well. Does anyone else have a better explanation for why there was such a visceral response?
> Does anyone else have a better explanation for why there was such a visceral response?
I can't speak for lawyers in general or what everyone's motivations would be, but my initial reaction was that it seemed like a somewhat unethical experiment. I assume the client would have agreed or represented themselves, but even there -- legal advice is tricky because it's advice -- it feels unethical to tell a person to rely on something that is very likely going to give them sub-par legal representation.
Sneaking it into a courtroom without the judge's knowledge feels a lot like a PR stunt, and one that might encourage further legal malpractice in the future.
I assume there are other factors at play, I assume many lawyers felt insulted or threatened, but ignoring that, it's not an experiment I personally would have lauded even as a non-lawyer who wishes the legal industry was, well... less of an industry. The goal of automating parts of the legal industry and improving access to representation is a good goal that I agree with. And maybe there are ways where AI can help with that, sure. I'm optimistic, I guess. But this feels to me like a startup company taking advantage of someone who's in legal trouble for a publicity stunt, not like an ethically run experiment with controls and with efforts made to mitigate harm.
Details have been scarce, so maybe there were other safety measures put in place; I could be wrong. But my understanding was that this was planned to be secret representation where the judge didn't know. And I can't think of any faster way to get into trouble with a judge then pulling something like that. Even if the AI was brilliant, it apparently wasn't brilliant enough to counsel its own developers that running experiments on judges is a bad legal strategy.
From what I've read recently, the legal profession is the one most at risk of adverse financial effects from AI. Not the court appearances nor the specialized work. But the run-of-the-mill boilerplate legal writing that is the bread and butter profit center of most first. You bet they are threatened and will push back.
Now the question is this. If an AI is doing something illegal like practicing law, how does one sanction an AI?
> Now the question is this. If an AI is doing something illegal like practicing law, how does one sanction an AI?
As far as I'm aware, no LLM has reached sentience and started taking on projects of its own volition. So it's easy - you sanction whoever ran the software for an illegal purpose or whoever marketed and sold the software for an illegal purpose.
> An AI is not a person, and therefore can't be sanctioned for practicing law - my take anyway.
"Personhood" in a legal sense doesn't necessarily mean a natural person. In this case, the company behind it is a person and is practicing law (so no pro se litigant using the company to generate legal arguments). In addition, if you want something entered into court, you need a (natural person) lawyer to do it, who has a binding ethical duty to supervise the work of his or her subordinates. Blindly dumping AI-generated work product into open court is about as clear-cut an ethical violation as you can find.
To your larger point, law firms would love to automate a bunch of paralegal and associate-level work; I've been involved in some earlier efforts to do things like automated deposition analysis, and there's plenty of precedent in the way the legal profession jumped on shepardizing tools to rapidly cite cases. Increased productivity isn't going to be reflected by partners earning any less, after all.
The legal profession is at the least risk of adverse financial effects from anything, because the people who make the laws are largely lawyers, and will shape the law to their advantage.
Automating boilerplate seems like a great use for AI if you can then have someone go over the writing and check that it's accurate.
I'd prefer that the boilerplate actually be reduced instead, but... I don't have any issue with someone using AI to target tasks that are essentially copy-paste operations anyway. I think this was kind of different.
> If an AI is doing something illegal like practicing law, how does one sanction an AI?
IANAL, but AIs don't have legal personhood, so it would be kind of like trying to sanction a hammer. I don't think that the AI was being threatened with legal action over this stunt, DoNotPay was being threatened.
In an instance where an AI just exists and is Open Source and there is no party at fault beyond the person who decides to download and use it, then as long as that person isn't violating court procedure there's probably no one to sanction? It's likely a bad move, but :shrug:.
But this comes into play with stuff like self-driving as well. The law doesn't think of AI as something that's special. If your AI drives you into the side of the wall, it's the same situation as if your back-up camera didn't beep and you backed into another car. Either the manufacturer is at fault because the tool failed, or you're at fault and you didn't have a reasonable expectation that the tool wouldn't fail or you used it improperly. Or maybe nobody's at fault because everyone (both you and the manufacturer) acted reasonably. In all of those cases, the AI doesn't have any more legal rights or masking of liability than your break pads do, it's not treated as a unique entity -- and using an AI doesn't change a manufacturer's liability around advertising.
That gets slightly more complicated with copyright law surrounding AIs, but even there, it's not that AIs are special entities that have their own legal status that can't own copyright, it's that (currently, we'll see if that precedent holds in the future) US courts rule that using an AI is not a sufficiently creative act to generate copyright protections.
> Law is different because the bar has a legally enforced monopoly on doing legal work.
I don't see how this would decrease DoNotPay's liability.
Regardless of how you feel about the bar, I don't think that changes anything about who they would sanction or why. Having a legal monopoly means they're even less likely to go along with a "the AI did it, not me" explanation than a normal market would be.
I mean, no matter what, they're not sanctioning the AI. They don't recognize the AI as a person, they recognize it as a tool that a person/organization is using to perform an action.
> Now the question is this. If an AI is doing something illegal like practicing law, how does one sanction an AI?
Its not and you don’t.
When a legal person (either a natural person or corporation) is doing something illegal like unauthorized practice of law, you sanction that person. The fact that they use an AI as a key tool in their unauthorized law practice is not particularly significant, legally.
I'm going to sit on that particular hill and see what happens. Even if DoNotPay's AI is not ready to do the job, the idea that AI could one day argue the law by focusing on logic and precedent instead of circumstance and interpretation is exceedingly threatening to a lawyer's career. No offense intended to the lawyers out there, of course. Were I in your shoes, I'd feel a bit fidgity over this, too.
i feel like lawyers will be able to legally keep AI out of their field for a while yet. they have the tools at their disposal to do so and a huge incentive.
> i feel like lawyers will be able to legally keep AI out of their
field for a while yet. they have the tools at their disposal to do
so and a huge incentive, other fields like journalism not so much.
That was my initial response too.
Artists, programmers, musicians, teachers are threatened... but shrug
and say "that's the future, what can you do". If lawyers feel
"threatened" by AI, they get it shot down.
I suddenly have a newfound respect for lawyers :)
Yet if we think about it, we all have exactly the same tools at our
disposal - which is just not playing that game. Difference is, while
most professions have got used to rolling with whatever "progressive
technology" is foisted on us, lawyers have a long tradition of caution
and moderating external pressure to "modernise". I'm not sure
Microsoft have much influence in the legal field.
When you're poor you have the choice between an AI that may work or you'll be defending yourself.
Access to legal assistance is almost as unobtainable as a dentist these days.
> When you're poor you have the choice between an AI that may work or you'll be defending yourself.
This is a thing that lots of people say about unethical businesses, and I'm a little skeptical about it at this point. A couple of objections I have:
- You have a constitutional right to legal representation when accused of a crime by the US government, and while we don't to abandon people who are suffering now because of some theoretical future fix, we also don't want to normalize the idea that constitutional rights only exist when a private market accommodates them. That's explicitly a bad direction for the country to go.
- Saying "well, this works here and now, and people don't have access to anything better" is in my mind only a really effective argument when we know that the thing here and now actually works. But we don't know that this works, which changes a lot about the equation.
- Is sneaking an AI into a courtroom through an earpiece really a cost-effective accessible strategy for poor people? Nothing about this screams "accessibility" to me.
I think summing up the last two points, if the AI was proven to actually work in a court of law, and was an accessible option, then sure, at that point I think the argument would have a lot more weight. It wouldn't be ideal, it would be a bad state for us to be in because your constitutional rights should not depend on an AI. But I could see a strong argument for using the AI in the meantime.
But that doesn't mean that DoNotPay should do unethical things right now to get to that point. The way that your choice is being phrased is begging the question: it assumes that the AI is the only choice other than no representation, that it does work, and that it will produce better outcomes.
But we don't actually know if the AI does work in a court of law, and DoNotPay's decision was to "move fast and break things"; it was to start releasing it into the wild without knowing what would happen. We don't know if asking people to represent themselves with a secret earpiece is a good legal strategy or if it's accessible. We don't know what happens when something goes wrong. We don't know that this actually is a working solution. But they were putting someone's legal outcome on the line anyway.
I think there's a big difference between making an imperfect solution available to poor people because we don't have anything better to offer, and using poor people as experimental fodder to build an imperfect solution that might not work at all. There's a lot of assumption here that using their AI would be better than representing yourself, and I don't know that's true. A judge is not going to pleased with being used as an experiment. And I've been hearing people say that the AI subpoenaed the officer involved in the ticket? That's not a good legal strategy.
The proper way to build a solution like this is to make sure it works before you start using it on people, and I think it's unethical to give someone bad legal advice and to try and justify it because giving that person bad legal advice might allow the company to help other people down the line. A lot of our laws around legal representation are predicated on the idea that legal advice should be solely focused on the good of the client, and not focused on the lawyer's career, or on someone else the lawyer wants to help, or on what the lawyer will be able to do in the future. Based on what we know about the state of the AI today, it doesn't seem like DoNotPay was thinking solely about the good of the person they were advising. We really don't want the legal industry to be an industry that embraces "the ends justify the means."
Yeah I feel like you're right on the money on re: the ethics of using someone who is in legal trouble who will have to live with the results. It's not as sexy but they should just build a fake case (or just use an already settled one if possible) and play out the scenario. No reason it wouldn't be just as effective as a "real" case.
I'd have no objections at all to them setting up a fake test case with a real judge or real prosecutors and doing controlled experiments where there's no actual legal risk and where everyone knows it's not a real court case. You're right that it wouldn't be as attention-grabbing, but I suspect it would be a lot more useful for actually determining the AI's capabilities, with basically zero of the ethical downsides. I'd be fully in support of an experiment like that.
Run it multiple times with multiple defendants, set up a control group that's receiving remote advice from actual lawyers, mask which group is which to the judges, then ask the judge(s) at the end to rank the cases and see which defendants did best.
That would be a lot more work, but it would also be much higher quality data than what they were trying to do.
And in some ways it’s less work! The risks of using a real court case are massive if you ask me. We are a wildly litigious country. No amount of waivers will stop an angry American.
> Run it multiple times with multiple defendants, set up a control group
And also
> That would be a lot more work, but it would also be much higher quality data
I don’t know much about the field of law, but anecdotally it doesn’t strike me as particularly data driven. So I think, even before introducing any kind of AI, the above would be met with a healthy dose of gatekeeping.
Like the whole sport of referencing prior rulings, based on opinions at a point in time doesn’t seem much different than anecdotes to me.
It's about volume. A fake case would be expensive to run and running dozens of them a day would be hard.
That said. The consequence of most traffic tickets is increased insurance and a fine. Yes these do have an impact on the accused, but they are the least impactful legal cases, so it would make sense to focus on them as test cases.
Is this not what moot court is? Seems like a great place to test and refine this kind of technology. The same place lawyers in training are tested and refined.
> Pretty sure the whole reason why DoNotPay actually exists is because defending against parking tickets didn't actually require a strong defense. The tickets were flawed automation...
I have some past experience working in the courts in my state, and I know there are many judges who are perfectly fine with dismissing minor traffic infractions for no reason other than that they feel like it. If you've got an otherwise clean traffic abstract and sent in a reasonable sounding letter contesting the infraction, these judges probably aren't going to thoroughly read through every word of it and contrast it with what was alleged in the citation. They don't really care about the city making an extra $173 off your parking ticket -- they just want to get through their citation reviews before lunch. Case dismissed.
So I am not surprised at all by the success of DoNotPay for minor traffic infractions. Most traffic courts are heavily strained by heavy case loads. If you give them a reason to throw your case out so they can go home on time, by all means, they will take it.
And I don't think anyone here has an issue with DoNotPay providing pre-trial advice and tips for someone defending themselves. It's bringing that into the courtroom that crosses a line from defending yourself to hiring an AI lawyer, and that line is where I'm very uncomfortable.
Thinking about how the problem would "elegantly solve itself" seems to illustrate the issue.
Someone using it in an actual courtroom would make a boneheadedly dumb argument or refer to a nonexistent precedent or something. Then maybe the judge gets upset and gives them the harshest punishment or contempt of court or they just lose the case. They may or may not ever get a chance to fix it.
A failure mode of jail time and/or massive fines for your customers doesn't sound all that elegant to me. This isn't a thing to show people cat pictures, I don't think move fast and break things is a good strategy.
Not to say that there aren't some entrenched possibly corrupt and self-serving interests here. But that doesn't mean they don't have a point.
It's probably better than the existing alternative. Which is roughly plead guilty because you don't have money to pay a lawyer. Or don't sue someone because you don't have money to pay a lawyer.
Judges would be absolutely right to punish lawyers or defendants that are bullshitting the court. They are wasting time and resources that would otherwise go towards cases where people are actually representing themselves in good faith.
The specific scenario doesn't matter. It's illegal to represent someone else in court if you're not a lawyer. There are a lot of things that you can't get a second chance at if your lawyer messes up that suing them can't fix. Lawyers and judges also negotiate, which a machine can't do because nobody feels an obligation to cut them some slack. Also now you're tainting case law with machine-generated garbage. Everything about the justice system assumes humans in the loop. You can't bolt on this one thing without denying people justice.
Not tricky at all. If someone is receiving counsel, then someone is giving counsel. Hiding behind a machine adds a pretty minor extra step to identifying the culprits, but does not create ambiguity over whether they are culpable.
On the other hand, here's a lawyer who thinks it would not count as legal representation, but hasn't seen the arguments made against it yet. Food for thought.
If you can sell a book that helps teach someone how to represent themselves, why can't you sell a person access to a robot that helps teach them how to represent themselves?
You're still illegally providing legal counsel if you're not a lawyer, or commiting malpractice if you are. Using a machine to commit the same crime doesn't change anything.
"Speech" would be like publishing a book about self-representation. "Counsel" would be providing advice to a defendant about their specific case. The machine would be in the courtroom advising the defendant on their trial, so that's counsel.
If the book was written about a particular case, that seems like specific legal advice.
If the book was a generalized "choose your own adventure" where you compose a sensible legal argument from selecting a particular template and filling it in with relevant data - use of the book essentially lets the user find the pre-existing legal advice that is relevant to their situation.
Chatbots as a system are arguably a lot more like the latter than the former - its a tool that someone can use to 'legal advise' themselves.
Are you still referring to the scenario from the article, or a different one where it's a resource you use outside of court?
> Here's how it was supposed to work: The person challenging a speeding ticket would wear smart glasses that both record court proceedings and dictate responses into the defendant's ear from a small speaker.
Also, probably wouldn't matter. The interactive human-ish-like nature might cross the line to being considered as counsel, even if you said it wasn't. See my response to your other comment.
Right, this strikes me as exactly the kind of "I'm not touching you!" argument that basically never works in a court of law. The law's not like code. "Well it's not any different than publishing a book, so this is just free speech and not legal representation"; "OK, cool, well, we both know that's sophist bullshit, judgement against you, next case."
By providing the words to say and arguments to make to the court, in response to a specific case or circumstance, DoNotPay was giving protected "legal advice" as opposed "legal information". There is ambiguity to find between legal advice and legal information, but that isn't.
A book gives legal information, not specific to a certain circumstance or case. If your chatbot is considering the specifics of a case before advising on a course of action, it's probably giving legal advice.
> Does anyone else have a better explanation for why there was such a visceral response?
It doesn't really matter if it'd work well or poorly. Lawyers don't want to be replaced, and being a lawyer entails a great ability to be annoying to delay/prevent things you don't want to happen.
It had to be shot down harshly because there are some premises to a courtroom proceeding that aren't met by an AI as we currently have.
One of those is that the lawyer arguing a case is properly credentialed and has been admitted to the bar, and is a professional subject to malpractice standards, who can be held responsible for their performance. An AI spitting out statistically likely responses can't be considered an actual party to the proceedings in that sense.
If a lawyer cites a non-existent precedent, they can make their apologies to the court or be sanctioned. If the AI cites a non-existent precedent, there's literally no way to incorporate that error back into the AI because there's no factual underlying model against which to check the AI's output--unless you had an actual lawyer checking it, in which case, what's the point of the AI?
Someone standing in court, repeating what they hear through an earpiece, is literally committing a fraud on the court by presenting themselves as a credentialled attorney. The stunt of "haha, it was really just chatGPT!" would have had severe legal consequences for everyone involved. The harsh response saved DoNotPay from itself.
> If the AI cites a non-existent precedent, there's literally no way to incorporate that error back into the AI because there's no factual underlying model against which to check the AI's output--unless you had an actual lawyer checking it, in which case, what's the point of the AI?
IANAL, but I would bet the level of effort to fact check an AI's output would be orders of magnitude lower than researching and building all your own facts.
I used it to generate some ffmpeg commands. I had to verify all the flags myself, but it was like 5 minutes of work compared to probably hours it would have taken me to figure them all out on my own.
Fact-checking nonsensical output would take a lot longer than researching a single body of law, which you can generally do by just looking up a recent case on the matter. You don't need to check every cite; that will have been done for you by the lawyers and judges involved in that case.
But checking every cite in an AI's output: many of those citations won't exist, and for the ones that do, you'll need to closely read all of them to confirm that they say what the AI claims they say, or are even within the ballpark of what the AI claims they say.
Fact checking an AI is still massively easier than finding and reading all the precedent yourself. Real lawyers of course already know the important precedent in the areas they deal in, and they still have teams behind the scene to search out more that might apply, and then only read the ones the team says look important.
Of course there could be a difference between an reading all the cases an AI says are important and actually finding the important cases including ones the AI didn't point you at. However this is not what the bet was about.
> Fact checking an AI is still massively easier than finding and reading all the precedent yourself.
Actually fact-checking an AI requires finding and reading all the precedent yourself to verify that the AI has both cited accurately and not missed contradictory precedent that is more relevant (whether newer, from a higher court, or more specifically on-point.)
If it has got an established track record, just as with a human assistant, you can make an informed decision about what corners you can afford to cut on that, but then you aren't really fact-checking it.
OTOH, an AI properly trained on one of the existing human-curated and annotated databases linking case law to issues and tracking which cases apply, overrule, or modify holdings from others might be extremely impressive—but those are likely to be expensive products tied to existing offerings from Westlaw, LexisNexis, etc.
What do you mean "finding"? The AI would just return links or raw text of the cases. Reading the findings would be the same as reading any precedence. But the AI could weight the results, and you'd only have to read the high scoring results. If the AI got it wrong, you'd just refine the search and the AI would be trained.
To the cost. If it removed the need for one legal assistant or associate then anything less than the cost of employing said person would be profit. So if it cost < 50k a year you'd be saving. (cost of employing is more than just salary)
You can't validate that it is making the right citations by only checking the cases it is citing, and the rankings it provides of those and other cases. You have to validate the non-existence of other, particularly contrary, cases it should be citing either additionally or instead, which it may or may not have ranked as relevant.
> You don't need to check every cite; that will have been done for you by the lawyers and judges involved in that case.
Why would this be different with an AI assistant to help you? It's not a binary "do or do not". Just because you have an assistant doesn't mean you don't do anything. Kind of like driver assist can handle some of the load vs full self-driving.
> But checking every cite in an AI's output: many of those citations won't exist, and for the ones that do, you'll need to closely read all of them to confirm that they say what the AI claims they say, or are even within the ballpark of what the AI claims they say.
But you'd have to do this anyway if you did all the research yourself. At least the AI assistant can help give you some good leads so you don't have to start from scratch. A lazy lawyer could skip some verifying, but a good lawyer would still benefit from an AI assistant as was my original bet, just like they would benefit from interns or paralegals, etc. And all those interns and paralegals could still be there, helping verify facts.
But you'd have to do this anyway if you did all the research yourself. At least the AI assistant can help give you some good leads so you don't have to start from scratch.
No, that's exactly the opposite of what I'm saying. If you did the research yourself, you wouldn't need to verify every cite once you find a relevant source/cite, because previous lawyers would have already validated the citations contained within that source. (A good lawyer should validate at least some of those cites, but frequently that's not necessary unless you're dealing with big stakes.)
And the AI assistant, at least this one and the ones based on ChatGPT, don't provide good leads. They provide crap leads that not only don't exist, but increase the amount of work. And any "AI" based on LLM will never be capable of providing good cites, because they'll never understand what they're reading and/or citing, and they'll miss relevant citations that are not statistically likely (i.e., new case law, or cases with similar facts, or similar law, or otherwise similar contexts that can be applied to the case at hand) that a context-aware AI or living breathing human would find easily.
At best, LLM-based AI might be able to help people with very simple legal situations. But you don't need AI for that. A single decision tree is easier to implement, and it's even easier to verify the domain-specific process and outcomes to make sure you don't get something silly like happened with this "AI".
But when appearing in court you're in real-time: you can't take 5 minutes to validate the AI output before passing it on. You can do that for your opening statements but once faced with the judge's rulings or cross-examination you'll be in the weeds.
Yeah that's fair, although if it was AI-assisted lawyer then presumably you'd have done the research ahead of time. But, for spontaneous stuff, you're totally right. My original statement was thinking about it as a "prep time" exercise, but spontaneous stuff would appear in court. Although, the human lawyer (who should still be simiarly prepared for court) would be there to handle those, possibly with some quick assistance.
If it was AI-assisted lawyer, it would be a whole different discussion. Aside from requiring a live feed of interactions to a remote system and other technical details, “lawyers using supportive tools while exercising their own judgement on behalf of their client” isn’t controversial the way marketing an automated system as, or as a substitute for, legal counsel and representation is.
I don't understand the "cites a non-existent precedent" bit. Presumably the AI would have a database of a pile of precedent. It wouldn't make up cites. It would have "knowledge" of so much precedence, it could likely find something to win either side of the argument.
I think you're misunderstanding how the model works. It predicts next tokens based on past tokens and the LLM trained on large bodies of text. It doesn't have an underlying database of "factual" elements it incorporates or searches, and its output doesn't have an underlying semantic structure that can be verified or reasoned about. The entirety of the quality of its output can only be judged by whether it "sounds" like the rest of the text on which it was trained.
I think making the connection between the predictive output and an underlying representation of reality is the next great step, but until that happens, chatGPT's output is just amazing mimicry of human language.
>That said, if it's such a catastrophically stupid idea, I'm not really sure why it had to be shot down so harshly
The title of the article seems misleading.
A techbro who doesn't appear to be a lawyer or has any understanding of the law wants to use AI so people can defend themselves. It doesn't seem like any of this was done with input from any bar associations. Without seeing the emails and "threats", and ignoring the emotional language it sounds like these people were helping him out:
>"In particular, Browder said one state bar official noted that the unauthorized practice of law is a misdemeanor in some states punishable up to six months in county jail."
Were these emails "angry" or just stating very plainly and with forceful language, that if you do this without the AI having the appropriate qualifications, you are most probably going to jail?
It even sounds like Browder didn't really widely publicise the fact that a case defended by an AI was about to happen.
>As word got out, an uneasy buzz began to swirl among various state bar officials, according to Browder. He says angry letters began to pour in.
Really sounds like these letter writers did him a favour.
> That said, if it's such a catastrophically stupid idea, I'm not really sure why it had to be shot down so harshly
To avoid the catastrophy that makes it a catastrophically bad idea.
> I assume the real reason it was shot down was out of fear that it would work well. Does anyone else have a better explanation for why there was such a visceral response?
It had already worked badly (subpoenaeing the key adverse witness, who would provide a basically automatic defense win, and one of the most common wins for this kind of case, if they failed to show up.)
DoNotPay exists because AI vaporware is the new crypto vaporware, which was the new IoT vaporware, which was the new Web2 vaporware, and so on. Build a "product" on AI and you get (in this case) $28 million in funding. Pull stunts like this to generate a little buzz for the next round of funding. Then bail out with your golden parachute. Now you have experience founding a startup - do it again for $50 million.
This is the obvious point: they fear it would work well and they will have to slowly say good bye to their extremely well paid profession.
We are so close from a new disruptive revolution where a lot of jobs (not just lawyers) will be made obsolete. Possibly similar to inventions like assembly lines, or cars. Such an exciting time to be alive!
1) The legal profession tries to instill a sense of ethics into the lawyers they train, and what DoNotPay was proposing violated that ethics. I don't want to overyhype the legal profession, but many (maybe even most) lawyers really do want to do the right thing, which by their lights means clients get the best representation possible. Which an LLM currently is not, so you get a visceral reaction from people at what they see as discussion of/advocacy of unethical conduct.
2) As a practical matter, it was likely to yield very bad outcomes for all concerned, including professional and legal consequences. Bluntly, DoNotPay was proposing to do something illegal, it really could have resulted in jail time, and possible disbarrment for any lawyers who were involved. For good or ill, judges have immense power to control what goes on in their courtrooms, and the risk of a judge taking offense to this is high. And the higher profile the case (and DoNotPay was offering $1m for a lawyer who would repeat whatever the chatbot said in a Supreme Court case) the higher the stakes. That really could be a career ending mistake for a junior lawyer (plus, almost by definition, anything that reaches the Supreme Court is important; bad representation could throw the result, with potentially terrible consequences for the country).
3) This is not, by any means, the first attempt to try and automate or streamline the provision of basic legal services. And it's a field ripe for such things; simple stuff like many rental agreements, employment contracts, sueing in small claims court, divorce agreements, etc., etc. all seem like they should be able to be generated by filling out a form and pressing a button, not hiring a high priced professional to craft a bespoke document. But over and over, such attempts have failed badly, and lawyers are getting reflexively defensive to anything that looks like that, not because they threaten their jobs, but because they keep being so terrible.
4) And DoNotPay looks a lot like these previous failures. See, eg, https://www.techdirt.com/2023/01/24/the-worlds-first-robot-l... which makes them look less like some cutting edge AI lawyer that will put an army of paralegals out of work, and more like yet another shitty site trying to resell Mechanical Turk at an enormous markup. Now, maybe what they're showing to the public is entirely different than what they want to use in courtrooms, but...it doesn't build confidence.
In many ways I think this is analagous to self driving cars. There is scope for enormous gains here, it's probably inevitable eventually, and it will probably put a lot of delivery drivers, taxi drivers, and truck drivers out of a job when it finally arrives. But right now deciding to hook your new self driving AI up to an unmarked truck and do a secret test in an unnamed US city without warning anyone or getting any licenses or permits is unethical and illegal, and you'd expect a very negative reaction if you announce your plans on Twitter. Especially if your website offers some software that they claim can control an RC car, but when you try it, it keeps steering into walls.
> I assume the real reason it was shot down was out of fear that it would work well
The next lawyer (or AI expert!) I see who thinks it would work well will be the first. There's just an enormous mismatch between the requirements, what chatbots in general seem to be capable of right now, and what DoNotPay in specific seems to be capable of. (Again, see the the Tech Dirt link.)
And again, note that while there's a pretty strong argument that it should be faster, easier, and cheaper to create a rental agreement (or whatever), DoNotPay was making a big deal out of how they wanted to argue in front of the Supreme Court. DoNotPay doesn't seem like they're going to nail the "draft a rental agreement for cheap" any time soon, but maybe they (or someone else) can. But handing a Supreme Court case? Obviously not. Now you might argue (and one certainly hopes!) the whole Supreme Court thing was just an unserious PR stunt, but when a company is making undifferentiated claims that their tech can do something semi-plausible and something entirely impossible, it makes the entire thing look like a scam created by or aimed at people who just don't know any better. Not a good look!
In short: I think the negative reaction is very, very understandable.
Browder (founder) appeared to also acknowledge that it was not fit for purpose as well [0].
If something that's providing input to a formal legal process (which, let's not forget, means false or inaccurate statements have real and potentially prejudicial repercussions), "makes facts up and exaggerates", then there seems to be no reason they should be talking about taking this anywhere near a courthouse.
This feels a lot like "move fast and break things" being applied - where the people silly enough to use this tool and say whatever it came up with would end up with more serious legal issues. It seems like that only stopped when the founder himself was the one facing the serious legal issues - 'good enough for thee, but not for me'...
I think what many are overlooking is that bad inputs to the legal system can jeopardise someone's position in future irretrievably, with little or no recourse (due to his class action/arbitration waiver). Once someone starts down the road of legal action, there's real consequences if you get it wrong - not only through exposure, but also through prejudicing your own position and making it impossible to take a different route, having previously argued something.
I think you’re right here and it’s the same reason I see AI as a tool in the software profession. You can use it to speed up your work, but you have to have someone fully trained who can tell the difference between looks good but is wrong, versus is actually usable.
I’ve been using copilot for half a year now and it’s helpful, but often wrong. I carefully verify anything it gives me before using it. I’ve had one bug that made it to production where I think copilot inserted code that I didn’t notice and which slipped past code review. I’ve had countless garbage suggestions I ignored, and a surprising amount of code that seemed reasonable but was subtly broken.
This will still require a human lawyer (and/or intern, depending on the stakes) to check its output carefully. I am not now, nor have I ever been afraid that AI is coming after my job. When it does, we’re dangerously close to general AI and a paradigm shift of such magnitude and consequence that it’s called The Singularity. At which point we may have bigger worries than jobs.
> I’ve had one bug that made it to production where I think copilot inserted code that I didn’t notice and which slipped past code review
I'm not saying this is good but come on. Humans do that all the time, why aren't we so harsh on humans?
> I am not now, nor have I ever been afraid that AI is coming after my job
I am. This thing amazed me, and even if it won't be able to 100% replace humans (which I doubt), it can make juniors orders of magnitude more productive for example. This will be a complete disruption of the industry and doesn't bode well for salaries.
I'm in the top x% of my profession, after 20 years of grinding. I'm unafraid. I don't see my salary taking a cut anytime soon. When I was searching for a job back in March, I had a 50% offer to response rate (if the company responded to my application.)
People who are just skating by may have cause for concern. But those are the people with the most to gain from it, so maybe not even them.
Demand is so high in the business, I have trouble imagining that any tool could make a meaningful impact on that.
It would need to multiply productivity by a huge number, and nothing has that impact. Copilot is barely, optimistically, above 10%. I don't really think I get that much, but just for arguments sake.
Broken window fallaccy. Software is so bad, that AI making is more productive won't put us out of work, it will make software suck a bit less. It eill5 make bugfixes and tech debt payback more affordable.
> I’ve been using copilot for half a year now and it’s helpful, but often wrong.
I wonder if that is because of the training set, us humans are often wrong or different. If given a room of programmers and asked to implement a Fibonacci algorithm would they all get it right, would they all do it via iteration, recursion, dynamic programming. Co-pilot might not replace you, but it just needs to replace some of those programmers. Then add tools like automated AI reviews or integration tests for example and now you removed another population of tech workers.
I am not sure if that is cause for alarm or the fact that such improvements could be rather beneficial. Some tools will replace people, some will be assistive, and as they improve and other layers are added they will reduce the need for people in areas, improving efficiency and productivity. Robots in manufacturing for example, improve productivity and reduce human labor.
And this leads on to this, these tools are narrow to general and have achieved this in a very short period of time. The cost factors have also massively reduced, if you could pay OpenAI $10 a month to make 1000 mistakes and still deliver code, or a human $120k a year to do the same then which one would you target. AI might not be coming for your job soon but it will be taking away your options to get a job. This is not unique to AI it is the basis for any technological improvement vs labor, yes new ideas and opportunities may come out of this but I don't think they will be equal in volume.
Yes! I think the legal system would and should look differently at a tool like this in the ear of a licensed lawyer, and AI tools will be invaluable for legal research. I just don't think the output of an AI should be fed directly into a non-lawyer's ear, any more than I think a non-programmer should try to build a startup with ChatGPT as technical co-founder.
What's interesting is that sometimes it does a great job at something like telling you the holdings of a case, but then other times it gives you a completely incorrect response. If you ask it for things like "the X factor test from Johnson v. Smith" sometimes it will dutifully report the correct test in bullets, but other times will say the completely wrong thing.
The issue I think is that it's pulling from too many sources. There are plenty of sources that are pretty machine readable that will give it good answers. There's a lot of training that can be eked out from the legal databases that already exist that could make it a lot better. If it takes in too much information from too many sources, it tends to get garbled.
There are also a lot of areas where it will confuse concepts from different areas of law, like mixing up criminal battery with civil battery, but that's not the worst of the problems.
> The issue I think is that it's pulling from too many sources. There are plenty of sources that are pretty machine readable that will give it good answers. There's a lot of training that can be eked out from the legal databases that already exist that could make it a lot better. If it takes in too much information from too many sources, it tends to get garbled.
No, this is a common misunderstanding about the way these things work. A LLM is not really pulling from any sources specifically. It has no concept of a source. It has a bunch of weights that were trained to predict the next likely word, and those weights were tuned by feeding in a large amount of text from the internet.
Improving the quality of the sources used to train the weights would likely help, but would not solve the fundamental problem that this isn't actually a lossless knowledge compression algorithm. It's a statistical machine designed to guess the next word. That makes it fundamentally non-deterministic and unsuitable for any task where factual correctness matters (and there's no knowledgeable human in the loop to issue corrections).
One useful way to think of language models is that they are statistical completion engines. It attempts to create a completion to the prompt, and then evaluates the likelihood, in a statistical sense, that the completion would follow the prompt, based on the patterns in the training data.
A citation in legalese is very common. A citation that is similar or identical to actual citations, in similar contexts, is therefore an excellent candidate for the completion. A fake citation that looks like a real citation is also a rather good candidate, and will sometimes squeak past the "is this real or fake?" metric used to evaluate generated potential responses.
This may seem like "pulling from a source" but there is no token, semantic information, or even any information in the model about where and when the citation was encountered. There is no identifiable structure or object (so far as anyone can tell anyway) in the model that is a token related to and containing the citation. It just learns to create fake citations so convincingly, that most of the time they're actually real citations.
This explains some of the particular errors that I've seen when poking and prodding it on complex legal questions and in trying to get it to brief cases.
ChatGPT can provide correct citations because somewhere deep in its weights it does lossily encode real texts and citations to real texts. That makes real citations in some cases be its most confident guess for what is supposed to come next in the sentence. But when there isn't a real text it is confident about giving a citation about, but it still feels like a citation should be next in the output, it will happily invent realistic looking citations to texts that have never existed and it has never seen in any sources. On the outside, as readers, it's hard to tell when this occurs without getting an outside confirmation. I'm assuming though that to some degree it is itself aware that a linked citation doesn't refer to anything
A few days ago on HN there was a short story of five paragraphs that started badly & finished OK, and I wondered if some operations research tricks could be applied. One is forward-backward-forward planning, this produces better schedules, and if applied would create a better opening, and perhaps ending (i.e. model is run three times).
In the case of citations, you really need a language model & a fact model. The language model then passes over to the fact model, then back to the language model. This means double(+) training.
I suppose the fact model could include things like Wolfram (also discussed on HN).
Asking "Can you cite some legal precedence for lemon law cases?" gives an answer containing
"In California, for example, the California Supreme Court in the case of Lemon v. Kurtzman (1941) held that a vehicle which did not meet the manufacturer's express warranty was a "lemon" and the manufacturer was liable for damages."
I dont think that case exist, there is a first amendment case Lemon v. Kurtzman, 403 U.S. 602 (1971) though.
I can't find any reference to Kurtzman or 1941 in any of the references. I think the answer is that the AI generating the text, and the code supplying the references are distinct and do not interact.
The example you give isn't necessarily a valid one. You're asking for a specific piece of knowable, measurable data -- one that has a single right answer and many wrong answers. Legal questions may have conflicting answers, they may have answers that are correct in one venue but not in another, etc., I have't yet seen any examples of an AI drawing the distinctions necessary for those situations.
ChatGPT has a lot of trouble understanding jurisdiction and what constitutes controlling precedent over what in which jurisdictions. As in it has no conception of it at all and gets it really badly mixed up. It doesn't understand the hierarchy of any court system so there are some questions that it will just always get wrong.
> If it is never pulling from a source, then why is it able to provide citations?
If you have the training set, and models that summarize text and/or assess similarity, and some basic search engine style tools to reduce or prioritize the problem space, it seems intuitively possible to synthesize probably-credible citations from a draft response without the response being drawn from citations the way a human author would.
Kind of a variant of how plagiarism detection works.
You.com is hugged to death right now, but from what I can see it's a different kind of chatbot. It looks closer to Google's featured snippets than it is to ChatGPT.
That kind of chatbot has different limitations that would make it unsuitable to be an unsupervised legal advice generator.
What if we prompt the LLM to generate a response with citations, and then we have program which looks up the citations in a citation database to validate their correctness? Responses with hallucinated citations are thrown away, and the LLM is asked to try again. Then, we could retrieve the text of the cited article, and get another LLM to opine on whether the article text supports the point it is being cited in favour of. I think a lot of these problems with LLMs could be worked around with a few more moving parts.
Definitely, no one is arguing that an AI lawyer will be the near future, but I can totally see it being good enough for the vast majority of small scale lawsuits within 10-20 years.
DoNotPay seems to know very well what they’re doing.
It really doesn’t strike me as true that law requires absolute precision. There are many adjacent (both near and far) arguments that can work in law for any given case, since the interpreter is a human. You just need no silly mistakes that shatter credibility, but that’s very different from “get one thing wrong and the system doesn’t work at all or works in wildly unexpected ways.”
Low end law will be one of the first areas to go due to this tech. DoNotPay actually has already been doing this stuff successfully for a while (not in court proceedings themselves though).
There are also many adjacent algorithms that could solve the same problem, but you still need to execute the algorithm correctly. LLMs are not ready for unsupervised use in any domain outside of curiosity, and what DoNotPay is proposing would be to let one roam free in a courtroom.
I'm not at all opposed to using LLMs in the research and discovery phase. But having a naive defendant with no legal experience parroting an LLM in court is deeply problematic, even if the stakes are low in this particular category of law.
That’s nowhere near analogous because between every working algorithm are massive gulfs of syntactic and semantic failure zones. This is not the case with human language, which is the whole power of both language production and language interpretation.
Is it more problematic than this person 1) not being represented, 2) having to pay exorbitant fees to be represented, or 3) having an extremely overworked and disinterested public defender?
I’m not convinced.
The idea that we need to wait to do this stuff until the people whose profession is under threat give “permission” is dismissible on its face and is exactly why we should be doing this is as quickly as possible. For what it’s worth, I mostly agree with you: I’m doubtful the technology is there yet. But that’s a call for each defendant to make in each new case and so long as they’re of sound mind, they should be free to pick whatever legal counsel they want.
> 1) not being represented, 2) having to pay exorbitant fees to be represented, or 3) having an extremely overworked and disinterested public defender?
You’re leaving off being put in jail for contempt of court, perjuring oneself, making procedural errors that result in an adverse default ruling, racking up fines, et cetera. Bad legal representation is ruinous.
Gee good thing everyone in court gets good representation at reasonable prices eh?
I get that lawyers think their profession is important (it is) and that by and large they’re positive for their clients (they are), but there are a lot of people who simply do not have access to any worthwhile representation. I saw Spanish-speaking kids sent to juvenile detention for very minor charges conducted in English with no interpreter and a completely useless public defender. So in my view that is the alternative for many people, not Pretty Decent Representation.
There are people who can stomach the downside risks to push this tech forward for people who cannot stomach the downside risks of the current reality.
Do you know that? How do you know that’s what the model would do? How do you know that the defendant doesn’t have an 80 IQ and an elementary school grasp on English? Do you think this doesn’t happen today and that these people don’t get absolutely dragged by the system?
We know that subpoenaing [possibly] no-show cops is what the model will do because that is what the CEO says the model did in the run up to this particular case.
Someone with an 80 IQ and an elementary school grasp of English is going to get absolutely dragged by the system with or without a "robot lawyer" if they insist on fighting without competent representation, but they'd probably still stand a better chance of getting off a fine on a technicality if they weren't paying a VC-backed startup to write a letter make sure the cops turned up to testify against them.
They'd also be more likely to not get absolutely dragged if they listened to a human that told them not to bother than a signup flow that encouraged them to purchase further legal representation to pursue the case.
Last time I checked with a lawyer about a traffic ticket he told me that it wasn't worth his time to go to court (this was the school provided free legal service and the case was just a burnt headlight that I didn't have verified fixed within a week, you decide if he should have gone to court with me or if he would have for a more complex case), but I was instructed how to present my case. I got my fine reduced at least, which was important as a student paying my own way (I'm one of the last to pay for college just by working jobs between class, so this was important)
Yeah, that's because he's an attorney that gets paid a lot. go to the traffic court and you'll find the ones that don't get paid a lot. That's why they are hanging out in traffic court representing litigants there.
I mean, not contesting the ticket is likely to be a better option than delegating your chances of not being convicted of contempt of court or perjury to the truthfulness and legal understanding of an LLM...
Sure if your objective is to minimize your own personal exposure. If your goal is to push toward a world where poor folks aren’t coerced into not contesting their tickets because they can’t afford to go to court or get representation, then maybe it is a good option.
I prefer a world in which people pay a small fine or make their own excuses rather than pay the fine money to a VC-backed startup for access to an LLM to mount a legally unsound defence that ends up getting them into a lot more trouble, yes.
If your goal is to ensure poor folks are pushed towards paying an utterly unaccountable service further fees to escalate legal cases they don't have the knowledge to win so the LPs of Andreesen Horowitz have that bit more growth in their portfolio, I can see how you would think differently.
> goal is to push toward a world where poor folks aren’t coerced into not contesting their tickets
Invest in a company doing this properly and not pushing half-baked frauds into the world. Supervised learning, mock trials. You’re proposing turning those poor folk into Guinea pigs, gambling with their freedom from afar.
This company has been doing this stuff for years. Yes this is a big step forward but it’s not from zero, as you’re suggesting. What makes you think they haven’t been doing mock trials and tons of supervised learning?
And no, I’m not. I don’t think Defendant Zero (or one, or two, or 100) should be people whose lives would be seriously affected by errors. I’m pretty sure DNP doesn’t either.
> What makes you think they haven’t been doing mock trials and tons of supervised learning?
The CEO tweeting they fucked up a default judgement [1]. That not only communicates model failure, but also lack of domain expertise and controls at the organisational level.
This is a false dichotomy that you are making up in order to politically justify your narrative that is otherwise completely made up nonsense best described as legal malpractice.
Alternative spin on the "know very well what they're doing": they know very well that it's unlicensed practice of law and they'd have to withdraw from the case.
But doing so generates lots of publicity for their online wizards that send out boilerplate legal letters.
The CEO tweeted about the system subpoenaing the traffic cop. If they actually built a system which is so advanced it can handle a court case in real time and yet so ignorant of the basics of fighting a traffic ticket it subpoenas the traffic cop it's... a very odd approach to product management for the flagship product of a legal specialist, and a bit scary to think anyone would use it. Easier to make the mistake of claiming your flagship system does stuff it shouldn't be doing if it's just vaporware and you haven't put too much thought into what it should do
Their track record? Seems like this is the first you’re hearing of them, but this is just the latest (and yes, most ambitious) experiment. They’ve been successfully using technology to help normal people defeat abusive systems built by “the professions” for years.
So if a chat program can pass the bar exam, it's okay? Because I would bet that if a program can represent someone semi-competently in court, passing the bar exam which needs an objective marking criterion would be trivial by comparison.
Cart before the horse... you have to pass the bar before you get the chance to represent someone semi-competently in court. Generally, lawyers have 5 years of experience before they are considered competent enough to be semi-competently represent someone in court.
Most states also require a law degree in addition to passing the Bar.
But a fun fact is that magistrates generally aren't required to pass the Bar, nor hold a law degree. Most states require extremely basic courses of 40 or so hours of training. I know of a magistrate that has tried numerous times to pass the Bar and has failed. I'm not sure how much competence our system mandates.
If you make a ridiculous argument using confabulated case law as a lawyer, you can be subject to discipline by the state bar and even lose your law license. The legal system's time and attention is not free and unlimited and that's why you need a license to practice law.
The judges and so forth don't want to deal with a bunch of people talking nonsense. Who is the lawyer who is putting their reputation on the line for the AI's argument? The people doing this want to say nobody is at fault for the obviously bogus arguments it's going to spout. That's why it's unacceptable.
Well, the problem is that the defendant has a right to competent representation, and ineffective assistance of counsel fails to fulfill that right.
(Your hypothetical includes a fine, so it isn't clear whether the offense in your hypothetical is one with, shall we say, enhanced sixth amendment protections under Gideon and progeny, or even one involving a criminal offense rather than a simple civil infraction, but...) in many cases lack of a competent attorney is considered structural error, meaning automatic reversal.
In practice, that means that judges (who are trying to prevent their decisions from being overturned) will gently correct defense counsel and guide them toward competence, something that frustrated me when I was a prosecutor but which the system relies upon.
Seems like the solution is clear then. If the judge gently corrects defense counsel and guides them towards competence, they can just do the same with AI. Then the company can use that data to improve it! Eventually it will be perfect with all the free labor from those judges.
>Judge: that case does not exist. Ask it about this case instead
>AI: I apologize for the mistake, that case is more applicable. blah blah blah. Hallucinates an incorrect conclusion and cites another hallucinated case to support it.
Judge: The actual conclusion to the case was this, and that other case also does not exist.
Isn't that the same thing? Seems fine to me, I know the legal system is already pretty overwhelmed but eventually it might get so good everyone could be adequately represented by a public defender.
Speaking of, I remember reading most poor people can only see the free lawyer they've been assigned for a couple minutes and they barely review the case? I don't understand how that is okay, as long as technically they're competent even if the lack of time makes them pretty ineffective...
Ehhh... the judge's patience for that kind of thing is not unlimited. At some point they're going to reopen the inquiry about counsel (there's also the issue that an AI probably can't be your counsel of record, since it hasn't passed the bar; more likely the court would view it as you representing yourself with the assistance of research software).
BREAK BREAK
My jurisdiction (the military justice system) is a bit of an oddball, but I generally (softly) disagree with your last paragraph.
In jurisdictions with good criminal justice systems, most cases don't take that long to review. Possession of marijuana case, the officer stopped you for a broken taillight, smelled or claimed to smell marijuana, asked you if it was okay if he takes a look. You said okay. He finds a tiny amount of a green leafy substance in a small plastic bag. He says, "Hey look, marijuana is not that big a deal but lots of this stuff is laced with other things. This is just marijuana, right?" and you respond "Yes, sir, just marijuana, I don't mess with any real drugs." Prosecution is offering a diversion deal with probation and no permanent criminal record.
The correct answer in that case is to take the deal. We are not going to win by arguing that Raich was wrongly decided or that the officer lied about smelling it (because we're probably talking state court, so Raich doesn't matter, and you consented to the search, so the pretext, even if it was pretextual, also doesn't matter). We also aren't going to win attacking the chain of evidence, because the drug lab results don't matter, because you admitted it's marijuana.
In that case, yes, I'm going to take all of about 4 minutes to strongly advise the client to take the deal.
Oh, that's actually a relief that the reason they take so little time with clients before telling them to take the deal is simply because the cases are generally clear cut. Although like many things, I'll bet this can vary by region significantly.
The rest of my comment I dropped the '/s' for. I think it's wild some people think current LLMs can replace lawyers... The absolute best I would think they could currently do is maybe speed up research for paralegals. I was just imagining expecting a judge to QA a private companies software and thought it was really funny.
A huge, huge number of cases are extremely cut and dry. Probably 80+% of the misdemeanor docket is drugs and traffic. You know what evidence is required to prove you were driving while your license was suspended? That you were driving (police cam) and that your license was suspended (certified DMV record).
I know it sounds scary when stats get thrown around like "Over 90% of defendants plead guilty without a trial!" but that's usually pursuant to a generous deal from an overworked prosecutor's office who absolutely does not want to do a freaking jury trial because you were desperate to get to work and couldn't get a ride.
Obviously, if the underlying reason your license was suspended is extreme (vehicular manslaughter, 3rd+ DUI) then their tune will probably change, but that's an extreme minority of cases. Those are the ones that go to trial.
"Counsel, I'm unfamiliar with the case you've cited. Have you brought a copy for the court? No? How about a bench brief? Very well. I am going to excuse the panel for lunch and place the court in recess. Members of the jury, please return at 1:00. Counsel, please arrive with bench briefs including printouts of major cases at 12:30. Court stands in recess." bang
That’s not how the legal system works. You aren’t slipping anything through. Either the judge knows the case, they don’t know all the cases, or the judge will research or clerks will research and you will be sanctioned if you try to do so thing unethical.
IANAL, but I'd think in this case this is prosecutor's job.
Also, the original post is about the traffic ticket. I'm pretty sure if the judge hears a reference to something he had never heard before, he'll be like "huh? wtf?"
If this is the case, the lawyers should have nothing to fear, and the plaintiff nothing to lose but a parking ticket. I say we stop arguing and run the experiment.
As noted by several lawyers when some of the details of this experiment were revealed: The AI already committed a cardinal sin of traffic court in that it subpoenaed the ticketing officer.
Rule 1 of traffic court: If the state's witness (the ticketing officer) doesn't show, defendant gets off. You do not subpoena that witness, thereby ensuring they show up.
If the AI or its handlers cannot be trusted to pregame the court appearance even remotely well then no way in hell should it be trusted with the actual trial.
You want to run this experiment? Great, lets setup a mock court with a real judge and observing lawyers and run through it. But don't waste a real court's time or some poor bastard's money by trying it in the real world first.
A reminder that "Move fast and break things" should never apply where user's life or liberty is at stake.
While I agree in absolute terms, in legal terms it is problematic because it sets precedent, which is what the law often runs off of. Better to not breach that line until we're sure it can perform in all circumstances, or rules have been established which clearly delineate where AI assistance/lawyering is allowed and where it isn't.
> noted by several lawyers when some of the details of this experiment were revealed: The AI already committed a cardinal sin of traffic court in that it subpoenaed the ticketing officer
Mike Dunford is a practicing attorney. Embedded tweet is of a non-lawyer who screenshotted Joshua Browser DoNotPay CEO) saying the subpoena had been sent. He has since deleted those tweets as DNP backs away from this plan.
Non-lawyers aren't banned from giving legal advice because lawyers are trying to protect their jobs, they're banned from giving legal advice because they're likely to be bad at it, and the people who take their advice are likely to be hurt.
Yes, in this case, it would just be a parking ticket, but the legal system runs on precedent and it's safer to hold a strict line than to create a fuzzy "well, it depends on how much is at stake" line. If we know that ChatGPT is not equipped to give legal advice in the general case, there's no reason to allow a company to sell it as a stand-in for a lawyer.
(I would feel differently about a defendant deciding to use the free ChatGPT in this way, because they would be deliberately ignoring the warnings ChatGPT gives. It's the fact that someone decided to make selling AI legal advice into a business model that makes it troubling.)
>> Non-lawyers aren't banned from giving legal advice because lawyers are trying to protect their jobs, they're banned from giving legal advice because they're likely to be bad at it, and the people who take their advice are likely to be hurt.
But why would the opposing side's lawyers care about this? They presumably want their client to win the lawsuit.
I only have immediate knowledge of UK law, but lawyers will generally have a duty to the court to act with independence in the interests of justice. This tends to mean that in situations where one side are self-represented or using the services of ChatGPT, etc. the opposing side is under a duty not to take unfair advantage of the fact that one side is not legally trained.
They don't have to help them, but they can't act abusively by, for example, exploiting lack of procedural knowledge.
If they deliberately took advantage of one side using ChatGPT and getting it wrong because the legal foundation of knowledge isn't there for that person, that could be a breach of their duty to the court and result in professional censure or other regulatory consequences.
When did the opposing side's lawyers say anything about this? Are you confused? Law is a regulated profession. The lawyers pointing out that this is illegal aren't on the other side of the case...
Well, it is supposed to be a Justice system, and not a game. While it is very easy to forget that, and many of the participants in it clearly don't behave as such, the outcome of it should be to be just.
Ultimately though the argument you have set up here seems to make it all but impossible for AI to displace humans in the legal profession. If the argument is "precedent rules" then "only humans can be lawyers" is precedent.
I'm not sure if this particular case with this particular technology made sense - but I do think we need to encourage AI penetration of the legal profession, in a way that has minimal downside risk. (For defendants and plaintiffs, not lawyers.) It would be hugely beneficial for society if access to good legal advice was made extremely cheap.
No, if in a hypothetical future we have technology that is capable of reliably performing the role, I don't have a problem with it. This tech is explicitly founded on LLMs, which have major inherent weaknesses that make them unsuitable.
They are not scared that it will fail. They are scared that it will succeed. And there's a great reason to allow a company to sell a stand-in for a lawyer. Cost. This isn't targeted at people who can afford lawyers, it's targeted at people who can't, for now at least.
It's naive to think that a company would develop an AI capable of beating a lawyer in court and then sell it cheaply to poor people to beat traffic tickets. If anyone ever manages to develop an AI that is actually capable of replacing a lawyer, it will be priced way, way out of reach of those people. It will be sold to giant corporations so that they can spend $500k on licence fees rather than $1 million on legal fees. (And unless those corporations can get indemnities from the software vendor backed by personal guarantees they'd still be getting a raw deal.)
These people are being sold snake oil. Cheap snake oil, maybe, but snake oil nonetheless.
Lawyers aren't scared at all. It's traffic court, you are really overstating things. If it was a serious case, it'd be even more ridiculous to put more on the line by being represented by a computer algorithm that isn't subject to any of the licensing standards of an atty, none of the repercussions, and being run by a business that is disclaiming all liability for their conduct.
You know what an attorney can't do? Disclaim malpractice liability!
It'd be wondrous if the esteemed minds of hackernews could put their brain cycles towards actually applying common sense and other things rather than jerking off to edgy narratives about disruption while completely disregarding the relevant facts to focus on what they find politically juicy ("lawyers are scared it will succeed". It's a tautological narrative you are weaving for yourself that completely skirts past all the principles underlying the legal profession and it's development over hundreds of years.
Considering it's so bad it came to people's attention when it sent a subpoena to make sure someone came to testify against its client when he might have had a default judgement in his favour if they hadn't, I think the people who can't afford the lawyers have a lot more to be scared of than the lawyers...
And the reason lawyers are expensive is because bad legal advice usually costs far more in the long run.
>They are not scared that it will fail. They are scared that it will succeed.
Not really. There are more lawyers than legal jobs. A lot of lawyers are toiling for well under 100k a year. People pay 1500 dollars an hour for some lawyers and 150 an hour for others due to perceived (and actual) quality differences. Adding a bunch more non-lawyers isn't going to impact the demand for the 1500 dollars an hour lawyers.
Legal work is expensive because ANY sort of bespoke professional work is expensive. Imagine if software developers had to customize their work for each customer.
> are not scared that it will fail. They are scared that it will succeed
Lawyers make heavy use of automated document sifting in e.g. e-discovery.
Junior lawyers are expensive. Tech that makes them superfluous is a boon to partners. When we toss the village drunk from the bar, it isn’t because we’re scared they’ll drink all the booze.
We can certainly run the experiment, just like we can let a kid touch a hot pan on the stove.
Like the kid, the experiment is not to add knowledge to the world. Every adult knows touching a hot pan results in a burn. Just like everyone who understands how current LLMs work knows that it will fail at being a lawyer.
Instead the point of such an experiment is to train the experimenter. The kid learns not to touch pans on the stove.
In this case it’s not fair to metaphorically burn desperate legal defendants so that the leaders and investors in an LLM lawyer company learn their lesson. It’s the same reason we don’t let companies “experiment” with using random substances to treat cancer.
I mean, why not run it as an experiment? Fake Parking ticket, fake defendant, pay a judge to do the fake presiding. If the actual goal was to test it, it would be trivially easy to do. The goal here wasn't to test it, it was to get publicity.
Exactly. I asked it for books on Hong Kong history and it spit out five complete fabrications. The titles were plausible and authors were real people, but none of them had written the books listed.
That case, to me, is indicative of a larger problem - it's 75 pages of arcane justifications, and yet I already knew how all of the justices had voted just from reading the premise, because like every Supreme Court case in a politicized area it was decided by personal conviction and the rest is post-hoc rationalization.
There is no hallucination on the part of the humans involved, only intellectual dishonesty.
We are unique in the universe but not important to its existence. Our automated inference technologies are accurately representing is; I read half a dozen STEM papers last year, even more during lockdown. Comma splices, grammatical errors everywhere. ChatGPT is us in the aggregate.
Even the geniuses of our species are imperfect and hallucinate being better than they are given their accomplishments relative to the laymen.
The court of law is itself an ephemeral hallucination which fails all the time; given the number of people proven innocent, it’s been suggested through analysis up to 25% or more may be incarcerated incorrectly. Drug laws are just one instance of humans hallucinating correct application of courts. YT broke a while back when it AI got hung up on circular logic in a debate about copyright (easily googled).
The burden of proof of “correctness” is on humans to prove their society is not merely a titillating hallucination.
We made computing machines before we had all the abstract semantics to describe it. Do those semantics mean anything to the production of computing machine or are they just a jargon bubble for a minority to memorize and capitalize on relative to those who have no idea wtf they’re talking about.
LOL - you gotta be kidding. In software, we strive for that - by running tests. In the law, there are no tests.
Not saying that absolute precision isn't required. I know lots of cases were an extra comma, a wrong date, or a signature from the wrong person has cost someone tens of millions of dollars. I would argue that AI-based tools could prevent such HUMAN mistakes.
> I would argue that AI-based tools could prevent such HUMAN mistakes.
Still you won’t find a useful/sober lawyer who would argue a case in front of a judge based on a made up precedent which never existed in reality.
Making a mistake as you put it, in humans, is quite different than “hallucinations” of an LLM. The practical AI tool that is good for preventing such human mistakes (precisely) doesn’t exist _yet_.
I agree in many cases, except traffic court isn’t real court. It’s mostly procedural. Hell in my state, in many cases the judge is a person who wins an election and takes a 20 hour course, and prosecutions are conducted by police officers.
Lawyers like to wax about the interests of justice. Reality is for crap like this, it’s a revenue funnel where there’s a pecking order of people who get breaks. Some small fraction have an actual legal dispute of facts.
I’d argue that you could best serve the interests of justice by bringing the whole process online. You’d eliminate the “friends and family” discount and have a real tribunal process for cases that need it, instead of the cattle call and mockery that these courts are.
NYC actually does a decent job with this by outsourcing the whole thing to DMV and administrative law judges on a tight leash. It’s mostly justice by flowchart. That pushes bullshit down to the police.
>>The legal profession is perhaps the closest one to computer programming
There is a ton of people using ChatGPT for programming.... so much so that I wonder if we will have a crisis in skills as people forgo how to write code.
sysadmin circles have tons people celebrating how they will not have to learn powershell now as an example
ChatGPT has no concept of a model for code, no understanding of syntax or logic or even language keywords. It just pulls info out of it's backside that sounds believable like a lot of people do, it's just better at it. I suspect the immediate future will be resplendent with tales of AI-generated code causing catastrophes.
i use copilot every day, and it's never written more than one line at a time that didn't need adjusting. If you are letting an AI write code for you without knowing what it does, you should not be in programming. I would probably just fire someone if their response to me was ever "well the AI wrote that bit of code that deleted our DB, i didnt know what it did"
There’s a lot of people doing a big circle jerk over ChatGPT with wild ideas of singularity and oddly eagerly awaiting the end of white collar work. Whatever. I agree that programmers being obsessed with it can lead to skill atrophy. But, in reality, there are many people that are very technical and are not becoming reliant on these things for coding.
I agree, but it doesn't have to be that way. I've been learning a couple new languages and frameworks lately and found it really accelerates my learning, increases enjoyment, and is good at explaining the silly mistakes that I make.
So it can enhance skills just as much as it can atrophy them.
And I'm okay with some skills atrophying... I hate writing regular expressions, but they're so useful for some situations. It's a shame chatGPT fails so hard at them, otherwise I would be content to never use a regex again.
> Defendant (as dictated by AI): The Supreme Court ruled in Johnson v. Smith in 1978...
> Judge: There was no case Johnson v. Smith in 1978.
> Defendant (as dictated by AI): Yes, you are right, there was no case Johnson v. Smith in 1978.
That's hilarious, watch some of the trails on courtTV on youtube. The trials are are as culturally biased as you can get. And these ones are the ones we get to see. Judges are not some logical Spock! Free of influence, politics, and current group think. But people who think about their careers, public opinion and know who pays their paycheck. And these are the the competent ones.
I remember judge Judy proclaiming "If it doesn't make sense, it's not true!!!",
while screaming at some goy. This is pretty much the level of logic you can expect from a judge.
Maybe that's where we're headed. It looks like it's becoming more and more okay to just make things up to please your tribe. Why shouldn't that seep into the courtroom. I hope this doesn't happen.
Ironically, I recently saw a convo on Twitter where someone was showing off a ChatGPT generated legal argument, and, it had done exactly that, hallucinated a case to cite.
In the court itself there would definitely be no way to trust them right now, but I could see AI being a useful research tool for cases. It could find patterns and suggest cases for someone that is qualified to look further into. No idea how hard it is for lawyers to find relevant cases now, but seems like it could be a tough problem.
Yes, absolutely. As a senior software engineer, Copilot has been invaluable in helping me to do things faster. But having an expert human in the loop is key: someone has to know when the AI did it wrong.
What's so bad about this experiment isn't that they tried to use AI in law, it's that they tried to use AI without a knowledgeable human in the loop.
Yes it's true that LLMs hallucinate facts, but there are ways to control that. Despite the challenges they can spit out perfectly functional code to spec to boot. So for me it's not too much of a stretch to think that it'd do a reasonably good job at defending simple cases.
Yes, you'd do better, but you'd still have a LLM that is designed to predict the most likely next words. It would still hallucinate and invent case law, it would just be even harder for a non-lawyer to recognize the hallucination.
Eh, what if it was trained on all the previous cases ever to have existed? I think it could be pretty good, as long as it detects novelty as a to flag and confirm error case.
That's not the point. LLMs work by predicting what text to generate next. It doesn't work by choosing facts, it works by saying the thing that sounds the most appropriate. That's why it's so confidently wrong. No amount of training will eliminate this problem: it's an issue with the architecture of LLMs today.
You could layer another system on top of the LLM generations that attempts to look up cases referenced and discards the response if they don't exist, but that only solves that particular failure mode.
There are other kinds of failures that will be much harder to detect: arguments that sound right but are logically flawed, lost context due to inability to read body language and tone of voice, and lack of a coherent strategy, to name a few.
All of these things could theoretically be solved individually, but each would require new systems to be added which have their own new failure modes. At our current technological level the problem is intractable, even for seemingly simple cases like this one. A defendant is better off defending themselves with their own preparation than they are relying on modern AI in the heat of the moment.
It’s bizarre that anyone that supposedly works in technology even thinks this is realistic. This betrays a large lack of knowledge of technology and a child like understanding of the legal system.
It fails at determining if a number is prime and provides bogus arguments to such effect. You think it would make sense for this to argue complex legal cases with strategy? This isn’t Go or chess.
The AI is also _deceptively_ "right". For example, it will cite precedent that has since been superseded.
A non-lawyer representing themselves in a criminal case would overlook that, make a bad/wrong/misinformed argument, and go to jail.
In other fields, it'll lie to you about the thickness of steel pipe required to transport water at a certain pressure, it'll refer to programming libraries that don't exist, and it'll claim something impossible in one breath and happily explain it as fact in the next.
At the same time, all these cases are on the internet somewhere. It wouldn’t be too tricky to make a lawyer gpt that is heavily trained on existing legal documents, and is only allowed to quote verbatim from real sources.
They will pretty soon be able to fact check everything they say when they gain real time internet connectivity.
But for now yeah you're right. A year-two from now this won't be true anymore.
Being connected to the internet may make their information more up to date but they won’t stop being confidently incorrect. These language models are amazing at token prediction but it’s a mistake to believe that they’re actually intelligent.
Give it an inane prompt to better see the difference, like "write me an essay about camels and their value in cybersecurity" which gives us the answer
"The camel is a valuable animal in the cybersecurity field due to its ability to store large amounts of data in its hump. Camels are able to cross vast distances with little water, which makes them ideal for carrying large amounts of data across networks. They are also very sturdy animals, able to withstand harsh conditions and even attacks from cyber criminals. In addition, their long eyelashes protect their eyes from sand and dust, making them perfect for working in dusty environments such as data centers."
Well it did what you asked him to do , you ordered it to write an essay about camels and their value in cybersecurity. It doesn't understand if you're joking or not or what the purpose of the whole thing.
I asked it this:
Are camels important to cyber security?
Answer:
"No, camels are not typically considered to be important for cyber security. Camels are domesticated mammals that are well-adapted to desert environments and are often used as a means of transportation or for their milk, meat, and hides. Cyber security, on the other hand, is the practice of protecting computer systems, networks, and data from unauthorized access, use, disclosure, disruption, modification, or destruction. The two are not related."
A human would say that’s ridiculous and tell you why. Google would give you a link to Apache Camel. ChatGPT tells you about how great a camels humps are for data storage.
That’s preposterous. An intelligent agent, human, mineral, or otherwise, would respond that this is a ridiculous idea and ideally explain the reasons that such is the case. Imagine you are a student and you asked this amazing AI sone question of similar if mildly ridiculed and turn imagine the student didn’t already know the answer. Would you think this kind of response would be an example of an intelligent AI?
If it cannot deal with such things without being prompted in such a way that the prompter knows the answer already, how could it deal with complex legal situations with actually intelligent adversaries?
This is overly optimistic. For one, fact checking is much harder than you think it is. Aside from that, there are also many additional problems with AI legal representation, such as lack of body language cues, inability to formulate a coherent legal strategy, and bad logical leaps. We're nowhere near to solving those problems.
AI hallucinations are going to be the new database query injection. Saying that real time internet connected fact checking will solve that is every bit as naive as thinking the invention of higher level database abstractions like an ORM will solve trivially injectable code.
We can't even make live fact checking work with humans at the wheel. Legacy code bases are so prolific and terrible we're staring down the barrel of a second major industry crises for parsing dates past 2037, but sure LLM's are totally going to get implemented securely and updated regularly unlike all the other software in the world.
I'd also argue that "hallucination" is, at least in some form, pretty commonplace in courtrooms. Neither lawyers' nor judges' memories are foolproof and eyewitness studies show that humans don't even realise how much stuff their brain makes up on the spot to fill blanks. If nothing else, I expect AI to raise awareness for human flaws in the current system.
That the legal system has flaws isn't a good argument for allowing those flaws to become automated. If we're going to automate a task, we should expect it to better, not worse or just as bad (at this stage it would definitely be worse).
> Defendant (as dictated by AI): The Supreme Court ruled in Johnson v. Smith in 1978...
> Judge: There was no case Johnson v. Smith in 1978.
LLMs hallucinate, and there is absolutely no space for hallucination in a court of law. The legal profession is perhaps the closest one to computer programming, and absolute precision is required, not a just-barely-good-enough statistical machine.