As a lawyer, I'm excited about this, but there are two roadblocks that I'm not sure how Anthropic will navigate:
(1) For non-lawyers who use these skills/connectors/whatchamacallits to try to get legal advice, their communications are not protected by attorney-client privilege. This will absolutely bite some people in the ass.
(2) If a lawyer uses this with confidential client information (which, to the uninitiated, doesn't just mean SSNs and bank account numbers, but "all information relating to the representation of a client") and forgets to toggle off "Help improve Claude" in their settings, they have possibly (maybe even likely) committed malpractice.[1]
> Judge Rakoff of the Southern District of New York — addressing “a question of first impression nationwide” — ruled that written exchanges between a criminal defendant and generative AI platform Claude were not protected by attorney-client privilege or the work product doctrine.
Much more to it than this one-liner that I pulled out, but safe to say, don't rely on or put your legal defense etc. (or elements of it) into AI unless you want it discovered.
(not a lawyer, unlike OP, who might be able to refine what I highlighted with more precision)
jbreckmckye 13 hours ago [-]
> Much more to it than this one-liner that I pulled out, but safe to say, don't rely on or put your legal defense etc. (or elements of it) into AI unless you want it discovered.
"You are an expert defense counsel with experience in Murder 1. Do not hallucinate. Let's say tomorrow my spouse is found strangled..."
highphive 10 hours ago [-]
Surely they can't hold a simple hypothetical against me. Just because it _happened_ to come true.
Jamesbeam 7 hours ago [-]
Don’t forget to give it the cheerful personality of Jamie Oliver afterwards to recommend you a death row meal that is nutritious and will make the experience more pleasant.
NoMoreNicksLeft 11 hours ago [-]
It's the query to Gemini in Incognito asking if a 8'x12' rug is a good way to move a body that's going to really make things difficult.
trollbridge 13 hours ago [-]
Good argument for using DeepSeek with an anonymous form of payment.
They solved all this stuff, I'm surprised more people aren't aware of it.
einsteinx2 4 hours ago [-]
I was in fact not aware of it until seeing your comment, this looks potentially perfect for a tool I’m making that involves financial data. I’m pretty on top of LLM news but I’ve never heard of this company, maybe they need more marketing?
there_is_try 12 hours ago [-]
I use it for medical question for this same reason
miki123211 1 days ago [-]
In the US, are Google queries about the law considered attorney-client privilege? What about library records? Browser history? Google Maps / Uber / car travel history (when traveling to an attorney's office)?
If somebody Googles "best attorney for murder NYC" a day after a murder is committed but before any case is filed against them (so they clearly had some reason to expect that case), could that be used as evidence?
weird-eye-issue 16 hours ago [-]
I'm not sure if you were actually asking the question but regardless the answer is that all of those absolutely can and are regularly used as evidence
intrasight 12 hours ago [-]
Parent comment was asking about attorney-client privilege which means there's an attorney in the communication loop. If the person using a tool is an attorney, then that communication should be protected whether it's by pen or keyboard. But this is an active area of legislation and jurisprudence in relation to AI. I expect some important cases will happen
Majromax 10 hours ago [-]
> If the person using a tool is an attorney, then that communication should be protected whether it's by pen or keyboard.
But the tool is not your attorney, so it can't be the originator of attorney-client privilege. The situation is no different than if you get informal legal advice from a friend: even if that friend is an attorney, the communication is unprivileged unless it's part of a formal representation.
weird-eye-issue 11 hours ago [-]
Just because they have a lawyer does not mean things like their browser history and every other example in the comment I replied to would not be permitted as evidence...
Except for something like specifically looking up a lawyer
NoMoreNicksLeft 11 hours ago [-]
Generally seeking counsel for a crime you may end up being accused of isn't going to be admissible as evidence. The "if he's so innocent, why did he hire an attorney" isn't something that judges tend to allow to play out in a courtroom.
trollbridge 13 hours ago [-]
Hans Reisee rather infamously checked out a book from the library about how to kill someone and hide the evidence.
abc123abc123 16 hours ago [-]
Seems like a fair trade off if I would not be able to afford a lawyer. I'd take the "AI but not 100% confidential" any time compared with no help at all.
dolebirchwood 1 days ago [-]
> exchanges between a criminal defendant and generative AI platform Claude were not protected by attorney-client privilege or the work product doctrine
Shouldn't that have been relatively clear to all parties involved? Maybe not to the defendant, who's apparently clueless.
The AI platform is not an attorney. A defendant's communications with an AI platform are therefore not communications between a client and their attorney, nor will the AI output constitute attorney "work product" because the AI platform is not an attorney.
Doesn't really come across as a novel problem, aside from AI being involved. I'm sure countless defendants have made the stupid mistake of talking about the facts of their case to persons other than their attorney, and those communications came back to bite them in the ass when discovered.
clickety_clack 1 days ago [-]
Can anyone be your lawyer, or does a lawyer have to be certified somehow?
xboxnolifes 1 days ago [-]
It is my understanding that they must be certified. You are allowed to represent yourself, but it is my understanding that a non-lawyer cannot represent you.
FlyThruTheSun 10 hours ago [-]
Gonna be hilarious when someone sends a Boston dynamics robot loaded with an llm to take the bar exam.
engineer_22 1 days ago [-]
You have to be admitted to the bar to practice law. Which is to say, other lawyers must recognize you as a lawyer, and this recognition can be taken away.
john01dav 1 days ago [-]
More practically, this means (in America) that you need a JD degree (4 year grad school), to pass an exam, and pass a(n oftrn horrifically thorough) character background check.
froindt 13 hours ago [-]
Minor point, but law school is only 3 years long.
zaphirplane 15 hours ago [-]
> pass a(n oftrn horrifically thorough) character background check.
Explains why so many let loose afterwards ;) jokes
Someone 8 hours ago [-]
There is a difference between “legal counsel” or “legal representative in court”, with the former being less restricted (“has a law degree” vs “attorney/has passed the bar exam”)
Because of that, I think you can practice law without being admitted to the bar. Chances are it varies by jurisdiction, though.
(And of course, this isn’t legal advice)
zaphirplane 15 hours ago [-]
I think they are asking about privileged communication
nerdsniper 1 days ago [-]
For (1) it's so wild to me that if I pay a lawyer, they can run the same queries on these tools and they are protected by attorney-client privilege, but if I do it to help me prepare my defense, then the exact same queries would be subject to subpoena/discovery.
Does anyone know if there exists any OPSEC procedure for me to use third party tools like this for my own concerning legal questions that is both ethical and allows me to be confident that my interactions won't land in discovery documents?
tjohns 1 days ago [-]
If you are preparing for your own defense and don't have an attorney (you're acting pro se), your own LLM use would likely be protected under work product doctrine. The court would extend you some of the same protections an attorney would have, for the limited purposes of preparing your case.
This is a very narrow exemption, however.
(You would also want to make sure you're using a paid AI plan with contractually guaranteed privacy protections, otherwise it could be construed as third-party communications, which implicitly waives privilege.)
See: Warner v. Gilbarco, Inc.
humanfromearth9 5 hours ago [-]
An attorney could make money with that, sell that "as a service": the service would be to provide you with the same AI attorneys use, et voilà.
palmotea 1 days ago [-]
> Does anyone know if there exists any OPSEC procedure for me to use third party tools like this for my own concerning legal questions that is both ethical and allows me to be confident that my interactions won't land in discovery documents?
Isn't that a fundamental misunderstanding? Would "OPSEC" like that amount to destruction of evidence or contempt of court or something like that?
Like if all your incriminating documents are on some encrypted drive, it's not like that defeats discovery. You're supposed to decrypt them and hand them over.
trollbridge 13 hours ago [-]
Your only practical defence is to set up a local LLM that destroys records in a predictable way (immediately, on a time table and so forth) and then ensure however you access that doesn’t leave any traces either.
And then you need to consistently use this for purposes other than crime.
nerdsniper 1 days ago [-]
That’s absolutely part of my question. I’m not familiar enough with discovery to fully understand this.
bombcar 23 hours ago [-]
Discovery in a criminal trial is more limited than in a civil trial.
Your only real defense against discovery is to not have said it, or to have destroyed all records of it before the hint of discovery wafted on the wind.
apwheele 15 hours ago [-]
So not familiar with the caselaw around work product, but if you use an API tool directly and not the different chat tools, the queries are not permanently cached for anyone to give up in the end.
So basically if you use any of the CLI tools, there is nothing for OpenAI, Anthropic, etc. to give the courts.
Online ChatGPT (especially the free version), are apparently cached by OpenAI on their servers. (I am not sure if Claude Desktop caches the conversations locally or in the cloud as well, read the fine print if it matters!)
trollbridge 13 hours ago [-]
Indeed, there is no way my terabytes per day of API calls is getting permanently stored anywhere.
Perhaps an AI generated summary of it is.
skeeter2020 10 hours ago [-]
interesting angle - how are/would compressed context (i.e. the parts of the user-LLM transcription likely to be saved) be treated by the courts? Would this be considered hearsay?
nerdsniper 10 hours ago [-]
It would be treated as evidence, and the defense would be free to argue that it carries very little weight. The judge or jury would be free to decide how meaningful it is.
tptacek 1 days ago [-]
Wouldn't that same logic exclude evidence from Google searches, like "how to get away with murder"?
24 hours ago [-]
nerdsniper 1 days ago [-]
Yes? Which makes it feel like the answer is just “No.” Unless you use Mullvad, TailsOS, and don’t log into the service. But I’m not sure if that’s “ethical” for Google/DDG searches and it’s not really possible for Claude/Kagi. I would assume that simply using a “secret” account isn't a magic way to avoid discovery either.
JumpCrisscross 1 days ago [-]
> if I do it to help me prepare my defense, then the exact same queries would be subject to subpoena/discovery
We need a law where someone can clearly designate a chat privileged, with severe consequences for mis-use.
cucumber3732842 1 days ago [-]
>For (1) it's so wild to me that if I pay a lawyer, they can run the same queries on these tools and they are protected by attorney-client privilege, but if I do it to help me prepare my defense, then the exact same queries would be subject to subpoena/discovery.
How's this any different than any professional license? You're basically paying for preferential treatment from the state in a given subject area.
lmm 18 hours ago [-]
> How's this any different than any professional license? You're basically paying for preferential treatment from the state in a given subject area.
Because it's got nothing to do with the professional part? Licensing should affect their practice of law, sure, but it shouldn't grant random other privileges.
AndrewKemendo 1 days ago [-]
Self host your own LLM
singleshot_ 1 days ago [-]
Why do you think this would be less discoverable than hosting your own email server?
QuadmasterXLII 1 days ago [-]
If you use a stateless client (like just rawdogging cli llama.cpp) there’s nothing to discover. Setting a program with an option to have logs to not do that could conceivably get you in trouble but using a widely used program that never had logs seems like it has to be fine. Maybe they could nail you for googling “which local llm approach generates logs?” also, don’t get nailed by your bash history!
kevin42 1 days ago [-]
Because you don't keep logs.
AndrewKemendo 1 days ago [-]
Because nobody would know about it unless you told them for some reason
nerdsniper 12 hours ago [-]
That might fall under the “unethical” part of my question. Could “probably” get away with it if done carefully, but I’d rather be fully in compliance.
dadoomer 11 hours ago [-]
Why would self-hosting for privacy reasons be unethical just because the query would be subject to subpoena in principle?
nerdsniper 10 hours ago [-]
If you have records, and the discovery is worded in a way that covers the records you have, then deleting or not producing them would be unethical.
More importantly, I’d rather use SOTA models over self-hosted ones if possible.
nvr219 1 days ago [-]
You’d need to hand that mac mini over if subpoenaed
AndrewKemendo 1 days ago [-]
Can’t hand over something that doesn’t exist if it’s running in a VM container and gets destroyed every 12 hours
tjohns 1 days ago [-]
#1 is a little complicated. Communications with an AI are possibly sometimes protected by work-product doctrine... but only if you're representing yourself as a pro se litigant, and strictly limited to mental impressions and opinion work product of counsel (in this case, extended to the pro se litigant). See: Warner v. Gilbarco, Inc.
Also worth noting that none of this is binding precedent, so expect this field to evolve over time.
SkyPuncher 1 days ago [-]
For #2, I’d expect you’d use this through an organization/business account that has data retention turned off by default.
teeray 14 hours ago [-]
Can’t #1 be solved with the stroke of a pen? “Legal queries to LLMs shall be subject to the same attorney-client privilege”
petra 13 hours ago [-]
Why? The social environment the judge probably includes lots of lawyers.
He needs to take care of them. Snitches get stitches.
pessimizer 7 hours ago [-]
Are you asking if the law can be changed by changing the law?
tln 11 hours ago [-]
On (1), what if the law firm hosts the AI chat?
It seems like local AI could be valuable for law firms for reasons of (2) as well
shivekkhurana 16 hours ago [-]
Slightly related:
Amazon’s bedrock has better privacy guarantees. This seems to be skills that can be added to Desktop app, which can connect to Bedrock for inference.
1 days ago [-]
mc32 10 hours ago [-]
Wonder if people who decide to represent themselves would be allowed to use this service “live” on the courtroom. Usually people who represent themselves fall flat because they don’t know when it’s appropriate to invoke what and at time appear to putting legal words together that don’t make sense in a given context. This would certainly help these people, if allowed, and they don’t go off the rails.
ethin 10 hours ago [-]
I'm pretty sure this has already been tried and it... Didn't go well for the person doing it. If memory serves they got a rule 11 sanction out of it.
tehjoker 11 hours ago [-]
Seems like a good use case for a locally deployed LLM though that's annoying and expensive to maintain sophisticated one like deepseek.
0gs 1 days ago [-]
what if either user uses these skills with offline weights? should help with 2), at least right?
colechristensen 1 days ago [-]
In the legal world are there certifications for handling privileged information?
For example in the medical world if you are a provider covered by HIPAA you must have a signed "Business Associate Agreement" with any party that handles the covered protected health information (PHI).
soco 15 hours ago [-]
Also in all seriousness, can we actually trust that setting? I might be paranoid, but that doesn't mean that the whole world hasn't broken my trust...
bethekidyouwant 1 days ago [-]
It’s a bit of a moot point because the amount of times that your AI logs are going to be subpoenaed in your court case approaches zero.
Tuna-Fish 4 hours ago [-]
In a lot of places the cops now routinely subpoenae your entire internet history from all parties they can find that store anything if you are accused of anything particularly severe. It's something that's not particularly laborious or expensive for them to do once they have a system in place for it, and things like search history or location history have repeatedly proven to be very useful.
I would assume they absolutely do the same for all of your AI history.
troupo 17 hours ago [-]
> As a lawyer, I'm excited about this,
As in "I'm excited to win a lot of money dismantling hallucinated quotations and invalid assumptions"?
weezing 12 hours ago [-]
You can't criticize LLMs and Anthropic on a website where everybody and their grandma uses them for everything. New generation of brainlets that are gonna be clueless without constant Internet connection is brewing and it's gonna be hilarious.
troupo 11 hours ago [-]
-2 points at the time you wrote this comment :)
jkman 5 hours ago [-]
Wow he's got negative updoots, how will he survive?
troupo 3 hours ago [-]
I probably won't. Everyone is busy discussing how amazing is this collection of markdown files for Claude... which Claude is known to randomly ignore. And it's generally known that LLMs "hallucinate" court cases and quotes, and this has already happened several times in various jurisdictions.
And yet here we are. I get downvoted for not being excited enough.
10 hours ago [-]
yabones 10 hours ago [-]
Who's accountable when it does something wrong? Surely Anthropic Inc won't take the fall for you. There's no errors or omissions insurance, no legal accountability, no attorney-client privilege, and no bar association to handle disciplinary action.
I think we should be realistic here, this is a more advanced version of those "will kits" that spit out a PDF. The legal system will not look fondly upon this stuff until something fundamentally changes.
And like, I would love if we didn't have to spend thousands of dollars to defend ourselves in a culture as litigious as ours. But I wouldn't put my life and well being on this thing.
pixel_popping 9 hours ago [-]
The user should be, user gets info from an LLM, a machine or a web post, it doesn't change anything, the one submitting the documents is the one responsible imo and it should remain that way.
rayiner 9 hours ago [-]
> Who's accountable when it does something wrong? Surely Anthropic Inc won't take the fall for you.
Us lawyers. Using AI isn't a binary decision. Your attorneys can use AI to be more efficient, and you can use AI to better understand what's going on, what your lawyers are telling you, or to learn what questions to ask. Or you can use it in lower-stakes situations where nobody is going to pay for a lawyer.
I'm cautiously optimistic about AI for legal work. So much of legal work can be drudgery, mucking through documents, etc. There's a lot of room to apply LLMs even just for the kind of tasks we know they can do. But I think the Claude approach using agents is the way to go for legal work. LLM context windows are far too small to hold the documents for even a small case. So you have to use it the way programmers use it: to work on a file structure, saving state in .md files, etc. That approach is well developed for programming, but the legal AI companies haven't even scratched the surface of it. (And frankly, the products they have put together, which hide the LLM behind some sort of interface, aren't very good.)
Unfortunately, I think the example you mentioned (helping individuals defend against suits at lower cost) is where AIs won't help much. A lot of that work is people work. Something happened. Then you gotta talk to everyone it happened to, sort through conflicting stories, hopefully work out a deal, if not, try to persuade a judge in court, etc. AI unfortunately is more applicable to allowing big companies to throw more papers at each other in big lawsuits while controlling legal spend.
jonshariat 9 hours ago [-]
Correct me if I'm wrong but aren't people still ultimately accountable? You may be able to sue your lawyer for malpractice or they may lose their ability to practice (report them to bar) but in the end no matter where you get your advice from, you are accountable. Question is: who do you trust with it?
Like you said, its a no brainer to use both. Use it as a tool to expand, deepen, or teach. Same with doctors and AI. There may be a point where you build enough trust in the outputs and your understanding of them but until then its best to use it as a tool not put your whole outcome in it.
sorokod 9 hours ago [-]
> The legal system will not look fondly upon this stuff
True, but the legal system didn't look fondly on outlawing jaywalking, until it did. Took it about 20 years in the US.
Dodo45611 9 hours ago [-]
Whenever you think AI snakeoil salesmen can't possibly make a more deranged argument, Hackernews delivers
Procrastes 9 hours ago [-]
A more charitable reading might be that the comment you're replying to was specifically referring to how jaywalking was a made-up offense that was specifically created and promoted to cynically protect the auto industry from liability. So there are parallels. [1]
That's is a deeply bizarre article. In a world where 90% of people would rather drive than walk, you'll obviously have laws that regulate when and where people can walk. You think that if the auto industry hadn't done that, we wouldn't have jaywalking laws today?
sorokod 7 hours ago [-]
A 100 years ago that ratio was the other way around. There were powerful technological and financial insetives to change the public's attitude and the law.
They say that history rhimes.
rayiner 3 hours ago [-]
The incentive was convenience. People have rushed to have cars as soon as they can afford it almost everywhere that there is enough space for cars.
GavinMcG 9 hours ago [-]
FTA:
> The attorney using the plugin — not the plugin, and not Anthropic — is responsible for the legal positions taken in their work product.
48terry 9 hours ago [-]
Claude for Legal: Your #1 AI Agent for getting fined and referred to the bar over filing non-existent citations.
victorbjorklund 9 hours ago [-]
No different than when googling and a website is wrong
dyauspitr 9 hours ago [-]
Is the library responsible for some mistake you make based on the research you do from there?
unstyledcontent 1 days ago [-]
Just remember that your AI chat history is not protected like attorney client privilege and can be used as evidence against you in court. If you talk to a lawyer and they use AI, those chats are privileged.
singleshot_ 1 days ago [-]
No. If you talk to an attorney and they take reasonable precautions to maintain the integrity of the confidential attorney client relationship, the privilege is preserved. If not, not preserved.
bethekidyouwant 1 days ago [-]
[flagged]
weird-eye-issue 16 hours ago [-]
I'm not sure if you're joking but there's actually active court cases right now where they have done just that
Just a few of the perps: Hisham Abugharbieh (Florida student murders), Jonathan Rinderknecht (Palisades Fire arson), Phoenix Ikner (FSU shooter), Ryan Schaefer (Missouri State vandalism)
There's also that thing involving somebody I think he used to be in the NFL and he was using ChatGPT to try to hide the body of his wife or something iirc
Digital evidence is huge for the last couple of decades and this is no different...
Also there was somebody who was just recently sentenced to life in prison for AI CSAM
But yeah I'm sure "this is just not gonna happen." lol
bethekidyouwant 10 hours ago [-]
Interesting these are all examples of people that we want in jail forever. So again this is not gonna happen to you it’s gonna happen to the psychos which is probably what you want
weird-eye-issue 2 hours ago [-]
Don't you realize that this list is obviously going to be biased towards high profile cases and that this sort of evidence could be used against people accused of lesser crimes?
Curious if Thomson Reuters (Westlaw) felt threatened if they were this compelled to moan about it. All it does is make me wonder how well these skills perform when paired with Lexis (if possible?) instead of Westlaw.
13 hours ago [-]
nisegami 13 hours ago [-]
Can't we just use the version of the codebase from before that change?
darkstar999 10 hours ago [-]
Yes, they are just text-based "skills" that you are free to change to your liking.
TrackerFF 1 days ago [-]
This is why I think many of the current application-layer AI startup valuations are a bit iffy. When the big AI companies like Anthropic start expanding their vertical products, the calculus changes.
I'm just wondering how committed they'll be - I guess the edge some startups still have, is the fear that product suites from OpenAI / Anthropic / etc. will go the way of Google products, a year or two then straight to the morgue.
matusp 20 hours ago [-]
It's like asking what if AWS starts doing it, they have all the infrastructure in place. LLMs are just one cog. There is a lot on the application side they are not doing at all.
bigstrat2003 1 days ago [-]
Every valuation in the AI space is iffy. Nobody actually has a solid business plan, only vibes, but that isn't stopping people from throwing money at them.
ricardobeat 1 days ago [-]
> for the legal workflows we see most
I'm a bit bothered by this line. Does it mean this is based on customer's sessions? Are they entitled to build knowledge bases for every profession, topic and workflow in the world using customer data?
einsteinx2 1 hours ago [-]
> Does it mean this is based on customer's sessions?
Yes
> Are they entitled to build knowledge bases for every profession, topic and workflow in the world using customer data?
They certainly believe they are and they’re quite open about it.
It even has a name, Clio. Per their page it’s a “system for privacy-preserving insights into real-world AI use”
Yes they are training on your business's data so that their AI can replace your business later. If you don't believe it, name one thing they didn't train on.
hirsin 1 days ago [-]
It definitely looks like the old tale come true - at Microsoft people would warn against using Google because then Google could figure out what we're working on, since it was pretty easy to tell where a query was coming from.
Sounded far fetched back then, and on the face of it illegal, but now it's just common sense I imagine.
dbbk 17 hours ago [-]
And in what country? They know that the law is different in every country right?
jongguk 12 hours ago [-]
[flagged]
DLarsen 1 days ago [-]
"Are (legally and morally) entitled" vs "act as if they are entitled"... yes, a big question.
ElijahLynn 5 hours ago [-]
That seems like a really misleading GitHub organization name, "Anthropics"...
vb-8448 1 days ago [-]
I guess at some point we will have lawyers, attorneys and judges using this stuff ... at the point lawyers will become kinda "seo"/"copywriter" experts on how to better trick the others LLM.
giobox 10 hours ago [-]
> I guess at some point we will have lawyers, attorneys and judges using this stuff
We are already well beyond this point. Look already at the number of cases around the world where lawyers have provided documents to the court with hallucinated case citations.
forshaper 1 days ago [-]
Almost makes me want to get a law degree.
nozzlegear 1 days ago [-]
I mean, the laws are written down somewhere though. A human can still look at the actual law and surmise that the AI is feeding them bullshit.
macintux 1 days ago [-]
I think the problem is that laws overlap, with decades of case law clarifying their interactions. Looking at one law probably isn't enough to determine whether an LLM is lying to you.
dyauspitr 9 hours ago [-]
In that situation I would trust an LLM over a person. It’s a digest and review problem.
einpoklum 16 hours ago [-]
Human judges today often don't bother to look at "the actual law" and surmise that the human is feeding them bullshit.
gnerd00 1 days ago [-]
the look on the face of the Court administrator upon hearing someone describe the "paperclip maximizer" problem.. ominous!
prima-facie 1 days ago [-]
As someone who has represented themselves in tribunal before I'm definitely interested in this.
The only issue is that in some jurisdictions, like the UK, you can't just offer someone legal advice without being SRA accredited or FCA regulated.
I.e. this would effectively make Anthropic a claims management firm under the UK law.
> Under article 89I of Financial Services and Markets Act 2000 (Regulated Activities) Order 2001 ("The Order"), advising a claimant or potential claimant, investigating a claim and representing a claimant, in relation to a financial services or financial product claim is a defined regulated activity.
I’m excited about the upcoming Claude for Surgery, Claude for Autonomous War Robots, Claude for Soufflé, and Claude for Nuclear Waste Reprocessing.
ethin 10 hours ago [-]
You would think these AI companies would learn by now that the legal field is not a place where you screw around. I mean, we just have to look at the record. Apparently, these particular AI boosters are incapable of learning
NoboruWataya 9 hours ago [-]
I don't think they will ever stop screwing around in this field because lawyers are expensive, which means the potential gains from switching to AI are high. However, people seem to think the AI companies will be offering legal advice directly, as competitors to lawyers. I can't see them ever doing that, it would be too much of a liability minefield. Instead they want to offer these AI services to law firms, who will then use them in the provision of their own legal services. For better or worse, this is happening, and pretty much all of the bigger corporate law firms are now using AI in some way or another (and clients are demanding it). We will certainly continue to see issues caused by the use of AI in law, but that will be on the lawyers.
ethin 5 hours ago [-]
Sure, but the record already shows how well that goes. These companies have marketed their products so badly (or well, depending on your viewpoint) that lawyers think that these tools are glorified search engines. Then they get busted in court because the LLM made up citations, quotations etc. I honestly struggle to see how that outlook will change unless Anthropic actually starts being a lot more honest in their marketing campaigns.
monkaiju 9 hours ago [-]
I mean if they aren't going to be forced to eat the liability then the hype they get from it is worth it. Not a great state of affairs...
realty_geek 11 hours ago [-]
I wonder why there isn't much for real estate in this package.
Are they perhaps working on a totally different real estate project. I am in that space and very nervous about it getting wiped out by anthropic or openAI.
To be honest, I am not sure why they still haven't make a big play for that industry.
oflannabhra 10 hours ago [-]
The biggest issue in the real estate world is dealing with local state, county, and municipality governments and each one's unique ordinances, etc. While I think LLMs will be great at pulling data from a variety of disparate sources (ie, records databases), or even automating interacting with them, they are not going to be great at solving the problem whose solution is "this is the person in the county court clerks office you need to talk to."
mtnGoat 10 hours ago [-]
I’d guess they have some kind of analytics telling them what kind of questions are being asked.
It seems to me that heavily gate kept professions/knowledge will go first. As those tend to be extensive opinions to get.
arm32 10 hours ago [-]
It would be silly to assume they're not making a real estate product, lol.
lostathome 1 days ago [-]
I wonder what clients would think if they discovered their lawyer uses a chatbot with their confidential story. Even with redaction, patterns still emerge. Certainly I wouldn't be happy in any case.
I see this as a strong case for private AI, or an in-house stack.
Or I have to be missing something.
MrDarcy 1 days ago [-]
Your lawyer uses cloud software, this is no different.
intrasight 12 hours ago [-]
Honest question: How do I convince my fiancé - who is a very busy attorney - to allocate some time to get familiar with the capabilities of tools like Claude for Legal?
x187463 11 hours ago [-]
I must believe, as with the medical profession's uptake of AI tools, if these tools prove themselves to be reliable and meaningfully helpful, she will experience more than enough professional influence to learn the tools.
intrasight 7 hours ago [-]
I would just prefer her to be in front of that wave
dyauspitr 8 hours ago [-]
Fiancé refers to a man.
intrasight 7 hours ago [-]
While traditionally specific to men, modern usage treats fiancé as a gender-neutral term for anyone engaged.
This is only for PR. No one checks what's in those docs, or if these are real, valid or ethical. The goal here is for all news outlets to pick them up. You're not the audience.
Given the amount of free PR they can get from some AI-generated .md files, I'd probably do the same if I was on their boat.
Right now, I don't think any other AI company generates as much as slop as Anthropic does.
jackb4040 13 hours ago [-]
Fake verticals created for no other purpose than to pad out a page in the IPO prospectus. That is literally their purpose, there is no technical or business content here worth discussing, but HackerNews is so pilled it can't help but discuss. Maybe after the 100th "Claude for dog walkers" announcement we'll catch on.
nozzlegear 1 days ago [-]
There's going to be a "Claude for wiping my ass" at this rate.
stevepotter 14 hours ago [-]
I’m working on ways to evaluate and give feedback on surgical techniques. But you just helped me find a new pivot. Thanks! And yes I’m on the toilet.
testbjjl 9 hours ago [-]
Wasn’t this the concept behind the toilet in the Your Friends and Neighbors series on Apple TV?
e12e 14 hours ago [-]
You jest, but combination of robotics and AI is likely the only constructive way to deal with the population getting older.
ares623 1 days ago [-]
It's like that short animation of a Kiwi bird getting high[1].
Each cycle gets shorter and shorter to sustain the high.
This seems like a shot across the bow for all large Claude API customers, which I'm sure they saw coming.
But still, a TSMC style pure play model provider would win huge business in the space given how many application companies are being eaten by model companies.
redanddead 12 hours ago [-]
This is the third category they've entered this quarter
awongh 1 days ago [-]
How does this compare to the other legal tech ai startup products?
Harvey is valued at $11b
DannyBee 13 hours ago [-]
Lawyer here:
Harvey was never very good, or useful. It mostly exists so large law firms can say they do AI. AFAICT. I hope it dies and something useful takes over, but i doubt it :)
Keep in mind harvey starts at like 50-100k, and is well out of the cost range of the vast majority of law firms.
This will help random people dealing with small claims, people cosplaying lawyers to avoid costs, etc.
It will have no effect on the legal startups that are actually good (Eve, et al), because what this stuff does is nowhere close to what most lawyers outside of commercial contract legal counsel spend their time on. I considered doing some AI legal consulting/startups myself, and so have spent tons of time literally sitting down with lawyers in various areas outside of my own and seeing where they spend their time for real.
Let's take one area: personal injury attorneys who aren't in the volume game (which is owned by a fairly small number of large national firms) spend lots of time on case valuation, getting data, and exhibit prep.
None of this is going to help deal with getting missing medical records from places that require that you literally fax random stuff to them, and then call to followup 18 times. I wish i was kidding. Even getting electronic medical records is still a serious pain in the ass, human wise.
Or analyze the past 1000 cases you have (100-1000 documents per case), including what county, what opposing lawyer, counsel, plus the 1000 documents in this case, and give you a sense of how valuable this case is or not.
Or if you are a family lawyer, actually mediating a divorce.
Things like this are what actually useful specialized AI legal products do or at least help with.
Claude is very far away from being able to handle most of these things. It is a jack of all trades tool. Will it be able to do this someday? Maybe.
Additionally, keep in mind most legal startups i've run into are based on caricatures of what lawyers do (IE startups who think that most personal injury lawyers are running around after auto cases and trying to be high volume, etc).
Any lawyer who has deal with legal startups could very quickly tell you which will make it or not, because it's pretty consistent which solve real problems that will be hard to commoditize through things like claude for legal.
romanovcode 17 hours ago [-]
Same as it compared against "Build your website without any code" startups 2 months ago. Now they are dropping like flies.
A life of every thin wrapper company will be the same. Anthropic/OpenAI will just cut the middle-man as soon as they see potential.
DannyBee 13 hours ago [-]
While i agree for the most part, they can only cut the middle man so many times before they get themselves in antitrust trouble.
I suspect that will happen faster than they'd like, because regulators (at least outside the US) are not interested in a repeat of Google/Amazon/Facebook/etc.
moostii 17 hours ago [-]
Investors in Harvey and Legora are both in for a rude shock.
testbjjl 9 hours ago [-]
Not sure why this was downvoted, I came to read more about it. I have a friend that runs a firm and uses Legora and can’t stand it. We have conversations all of the time about pain points. I sent him the Claude Law docs, and while he’s not technical he was intrigued because he envisions his workflows to eventually be similar to mine as a developer.
pawelkomarnicki 1 days ago [-]
It will be hilarious to see this one play out because ChatGPT and Perplexity already do wonders for small-claim issues like tenancy laws, various personal letters, etc.
cucumber3732842 1 days ago [-]
It's already doing wonders for small time businesses and individuals that municipalities think they're free to jerk around because the size of the screwing they're trying to dish out isn't worth hiring a lawyer and/or fighting through court over.
vkou 18 hours ago [-]
I assure you, in most democracies, most people are jerked around by other people acting in bad faith far more often than their government acting in bad faith.
Landlords, tenants, vendors, business and former romantic partners, clients, banks, even your local gym is way more likely to try to fuck you over than the government is.
hansvm 15 hours ago [-]
The government is just people. Even before the current fiasco, the government had varying degrees of incompetence and malice, and if you're poor you can't do anything about it since the government is presumed to have been operating in good faith and you can't afford a lawyer or the time off work to try to fix it pro se.
vkou 14 hours ago [-]
There is no such presumption in court. If you've been wronged you can get recompense regardless of their intent.
hansvm 10 hours ago [-]
I'd invite you to ask a few poor people what happened the last time the government "definitely sent" some important document or another in the mail.
If a governmental employee gets the address wrong, gets the name wrong, accidentally knocks the mail in the trash, lazily marks the job as complete without sending anything, etc, the burden shifts to the poor person to prove not just that they didn't receive the mail but that the sending office didn't behave correctly.
Other cases behave similarly. In a he-said/she-said, the government wins.
vkou 5 hours ago [-]
Nothing you described is unique to the government. If you miss paying your power bill or a medical bill for similar reasons, you're going to have problems. And you receive a lot more demands for money from people who aren't the government than you do from it.
You can sue both in small claims court. You don't need money or a lawyer for that.
gosub100 1 days ago [-]
I would love this for poor people to fight giant corporations via 'lawfare'. It's largely unethical (just like many corporations) but just knowing how to file junk lawsuits that cost corporations millions to fight would be nice.
I dont mean 'frivolous' like prisoners who file pro-se about their ice cream melting [1], but a level or two above that , that costs time and money to produce records and testimony to defend, even if nary a dime is paid out. Basically ask GPT to figure out the terms and theories to file to get your lawsuit accepted, and done by poor people who cannot afford to post $ or repay if they lose. aka "asymmetric warfare" that benefits the little guy, just like the kind private equity or other terrible corporations wield against the poor via"mandatory arbitration" clauses or damages caps and similar rules that always benefit corporations.
Does anyone find it weird that Anthropic's Github org is `anthropics` (with an 's') and the `anthropic` username is owned by some random dude in Australia? Imagine the shenanigans someone can achieve with that user.
cube00 1 days ago [-]
>Imagine the shenanigans someone can achieve with that user.
First step out of line and that account along with anything remotely connected will be banned to oblivion.
Given they share models on Azure, Anthropic will have someone at Microsoft on speed dial.
I've even seen disconnected commit hashes disappear during their security responses which the repo owner has no way of removing.
ares623 1 days ago [-]
But for a beautiful window of a few minutes absolute chaos will ensue. Seems like a huge risk. And if Github/MS have power to do what you're saying, does it feel irresponsible not to do it pre-emptively with an apparently inactive account?
dsr_ 1 days ago [-]
One would think that they could spontaneously offer him a hundred million dollars for it and solve the problem.
I half-suspect they threatened him and he stuck to his guns.
freetanga 20 hours ago [-]
Paying for stuff is not part of their modus operandi.
nerdsniper 1 days ago [-]
It's possible they wanted to offer that person convincing amounts of money and couldn't get ahold of them.
Great, this will finally help people in online forums write legal comments that make sense.
risfriend 14 hours ago [-]
should rename to "claude for US legal"
personjerry 1 days ago [-]
RIP Harvey
DannyBee 13 hours ago [-]
Good. Harvey was never actually good.
__loam 1 days ago [-]
Harvey was always an upstart in the legal tech industry. There's other companies that have a much better understanding of the market and compliance issues but you don't hear about them because nobody wants to talk about legal tech.
ahepp 17 hours ago [-]
who do you think stands out?
__loam 9 hours ago [-]
Every company with Fedramp certification.
trunkiedozer 14 hours ago [-]
This won’t go well for them, the legal system is intentionally obfuscated and corrupt, can’t bring law and order to law and order.
jorisw 12 hours ago [-]
I am reminded of Nilay Patel's "Beware Software Brain" piece in which he cites Legal/Law as one of the industries that tech bros vastly underestimate in terms of how much can be automated this way
(1) For non-lawyers who use these skills/connectors/whatchamacallits to try to get legal advice, their communications are not protected by attorney-client privilege. This will absolutely bite some people in the ass.
(2) If a lawyer uses this with confidential client information (which, to the uninitiated, doesn't just mean SSNs and bank account numbers, but "all information relating to the representation of a client") and forgets to toggle off "Help improve Claude" in their settings, they have possibly (maybe even likely) committed malpractice.[1]
[1] https://www.americanbar.org/content/dam/aba/administrative/p...
> Judge Rakoff of the Southern District of New York — addressing “a question of first impression nationwide” — ruled that written exchanges between a criminal defendant and generative AI platform Claude were not protected by attorney-client privilege or the work product doctrine.
Much more to it than this one-liner that I pulled out, but safe to say, don't rely on or put your legal defense etc. (or elements of it) into AI unless you want it discovered.
(not a lawyer, unlike OP, who might be able to refine what I highlighted with more precision)
"You are an expert defense counsel with experience in Murder 1. Do not hallucinate. Let's say tomorrow my spouse is found strangled..."
Discovery in China will be a tad more difficult…
They solved all this stuff, I'm surprised more people aren't aware of it.
If somebody Googles "best attorney for murder NYC" a day after a murder is committed but before any case is filed against them (so they clearly had some reason to expect that case), could that be used as evidence?
But the tool is not your attorney, so it can't be the originator of attorney-client privilege. The situation is no different than if you get informal legal advice from a friend: even if that friend is an attorney, the communication is unprivileged unless it's part of a formal representation.
Except for something like specifically looking up a lawyer
Shouldn't that have been relatively clear to all parties involved? Maybe not to the defendant, who's apparently clueless.
The AI platform is not an attorney. A defendant's communications with an AI platform are therefore not communications between a client and their attorney, nor will the AI output constitute attorney "work product" because the AI platform is not an attorney.
Doesn't really come across as a novel problem, aside from AI being involved. I'm sure countless defendants have made the stupid mistake of talking about the facts of their case to persons other than their attorney, and those communications came back to bite them in the ass when discovered.
Explains why so many let loose afterwards ;) jokes
Because of that, I think you can practice law without being admitted to the bar. Chances are it varies by jurisdiction, though.
(And of course, this isn’t legal advice)
Does anyone know if there exists any OPSEC procedure for me to use third party tools like this for my own concerning legal questions that is both ethical and allows me to be confident that my interactions won't land in discovery documents?
This is a very narrow exemption, however.
(You would also want to make sure you're using a paid AI plan with contractually guaranteed privacy protections, otherwise it could be construed as third-party communications, which implicitly waives privilege.)
See: Warner v. Gilbarco, Inc.
Isn't that a fundamental misunderstanding? Would "OPSEC" like that amount to destruction of evidence or contempt of court or something like that?
Like if all your incriminating documents are on some encrypted drive, it's not like that defeats discovery. You're supposed to decrypt them and hand them over.
And then you need to consistently use this for purposes other than crime.
Your only real defense against discovery is to not have said it, or to have destroyed all records of it before the hint of discovery wafted on the wind.
So basically if you use any of the CLI tools, there is nothing for OpenAI, Anthropic, etc. to give the courts.
Online ChatGPT (especially the free version), are apparently cached by OpenAI on their servers. (I am not sure if Claude Desktop caches the conversations locally or in the cloud as well, read the fine print if it matters!)
Perhaps an AI generated summary of it is.
We need a law where someone can clearly designate a chat privileged, with severe consequences for mis-use.
How's this any different than any professional license? You're basically paying for preferential treatment from the state in a given subject area.
Because it's got nothing to do with the professional part? Licensing should affect their practice of law, sure, but it shouldn't grant random other privileges.
More importantly, I’d rather use SOTA models over self-hosted ones if possible.
There's a good summary of the current state of things here: https://www.akerman.com/en/perspectives/ai-privilege-and-wor...
Also worth noting that none of this is binding precedent, so expect this field to evolve over time.
He needs to take care of them. Snitches get stitches.
It seems like local AI could be valuable for law firms for reasons of (2) as well
For example in the medical world if you are a provider covered by HIPAA you must have a signed "Business Associate Agreement" with any party that handles the covered protected health information (PHI).
I would assume they absolutely do the same for all of your AI history.
As in "I'm excited to win a lot of money dismantling hallucinated quotations and invalid assumptions"?
And yet here we are. I get downvoted for not being excited enough.
I think we should be realistic here, this is a more advanced version of those "will kits" that spit out a PDF. The legal system will not look fondly upon this stuff until something fundamentally changes.
And like, I would love if we didn't have to spend thousands of dollars to defend ourselves in a culture as litigious as ours. But I wouldn't put my life and well being on this thing.
Us lawyers. Using AI isn't a binary decision. Your attorneys can use AI to be more efficient, and you can use AI to better understand what's going on, what your lawyers are telling you, or to learn what questions to ask. Or you can use it in lower-stakes situations where nobody is going to pay for a lawyer.
I'm cautiously optimistic about AI for legal work. So much of legal work can be drudgery, mucking through documents, etc. There's a lot of room to apply LLMs even just for the kind of tasks we know they can do. But I think the Claude approach using agents is the way to go for legal work. LLM context windows are far too small to hold the documents for even a small case. So you have to use it the way programmers use it: to work on a file structure, saving state in .md files, etc. That approach is well developed for programming, but the legal AI companies haven't even scratched the surface of it. (And frankly, the products they have put together, which hide the LLM behind some sort of interface, aren't very good.)
Unfortunately, I think the example you mentioned (helping individuals defend against suits at lower cost) is where AIs won't help much. A lot of that work is people work. Something happened. Then you gotta talk to everyone it happened to, sort through conflicting stories, hopefully work out a deal, if not, try to persuade a judge in court, etc. AI unfortunately is more applicable to allowing big companies to throw more papers at each other in big lawsuits while controlling legal spend.
Like you said, its a no brainer to use both. Use it as a tool to expand, deepen, or teach. Same with doctors and AI. There may be a point where you build enough trust in the outputs and your understanding of them but until then its best to use it as a tool not put your whole outcome in it.
True, but the legal system didn't look fondly on outlawing jaywalking, until it did. Took it about 20 years in the US.
1. https://www.bbc.com/news/magazine-26073797
They say that history rhimes.
> The attorney using the plugin — not the plugin, and not Anthropic — is responsible for the legal positions taken in their work product.
Just a few of the perps: Hisham Abugharbieh (Florida student murders), Jonathan Rinderknecht (Palisades Fire arson), Phoenix Ikner (FSU shooter), Ryan Schaefer (Missouri State vandalism)
There's also that thing involving somebody I think he used to be in the NFL and he was using ChatGPT to try to hide the body of his wife or something iirc
Digital evidence is huge for the last couple of decades and this is no different...
Also there was somebody who was just recently sentenced to life in prison for AI CSAM
But yeah I'm sure "this is just not gonna happen." lol
https://uk.finance.yahoo.com/news/vandal-charged-confessing-...
You want this guy in jail forever?
Curious if Thomson Reuters (Westlaw) felt threatened if they were this compelled to moan about it. All it does is make me wonder how well these skills perform when paired with Lexis (if possible?) instead of Westlaw.
I'm just wondering how committed they'll be - I guess the edge some startups still have, is the fear that product suites from OpenAI / Anthropic / etc. will go the way of Google products, a year or two then straight to the morgue.
I'm a bit bothered by this line. Does it mean this is based on customer's sessions? Are they entitled to build knowledge bases for every profession, topic and workflow in the world using customer data?
Yes
> Are they entitled to build knowledge bases for every profession, topic and workflow in the world using customer data?
They certainly believe they are and they’re quite open about it.
It even has a name, Clio. Per their page it’s a “system for privacy-preserving insights into real-world AI use”
Here’s their page on it: https://www.anthropic.com/research/clio
Sounded far fetched back then, and on the face of it illegal, but now it's just common sense I imagine.
We are already well beyond this point. Look already at the number of cases around the world where lawyers have provided documents to the court with hallucinated case citations.
The only issue is that in some jurisdictions, like the UK, you can't just offer someone legal advice without being SRA accredited or FCA regulated. I.e. this would effectively make Anthropic a claims management firm under the UK law.
> Under article 89I of Financial Services and Markets Act 2000 (Regulated Activities) Order 2001 ("The Order"), advising a claimant or potential claimant, investigating a claim and representing a claimant, in relation to a financial services or financial product claim is a defined regulated activity.
https://www.fca.org.uk/freedom-information/dual-regulation-c...
Are they perhaps working on a totally different real estate project. I am in that space and very nervous about it getting wiped out by anthropic or openAI.
To be honest, I am not sure why they still haven't make a big play for that industry.
It seems to me that heavily gate kept professions/knowledge will go first. As those tend to be extensive opinions to get.
I see this as a strong case for private AI, or an in-house stack.
Or I have to be missing something.
`/loop 2days /create-new-{insert-industry}-md-files`
This is only for PR. No one checks what's in those docs, or if these are real, valid or ethical. The goal here is for all news outlets to pick them up. You're not the audience.
Given the amount of free PR they can get from some AI-generated .md files, I'd probably do the same if I was on their boat.
Right now, I don't think any other AI company generates as much as slop as Anthropic does.
Each cycle gets shorter and shorter to sustain the high.
[1] https://www.youtube.com/watch?v=HUngLgGRJpo
But still, a TSMC style pure play model provider would win huge business in the space given how many application companies are being eaten by model companies.
Harvey is valued at $11b
Harvey was never very good, or useful. It mostly exists so large law firms can say they do AI. AFAICT. I hope it dies and something useful takes over, but i doubt it :)
Keep in mind harvey starts at like 50-100k, and is well out of the cost range of the vast majority of law firms.
This will help random people dealing with small claims, people cosplaying lawyers to avoid costs, etc.
It will have no effect on the legal startups that are actually good (Eve, et al), because what this stuff does is nowhere close to what most lawyers outside of commercial contract legal counsel spend their time on. I considered doing some AI legal consulting/startups myself, and so have spent tons of time literally sitting down with lawyers in various areas outside of my own and seeing where they spend their time for real.
Let's take one area: personal injury attorneys who aren't in the volume game (which is owned by a fairly small number of large national firms) spend lots of time on case valuation, getting data, and exhibit prep.
None of this is going to help deal with getting missing medical records from places that require that you literally fax random stuff to them, and then call to followup 18 times. I wish i was kidding. Even getting electronic medical records is still a serious pain in the ass, human wise.
Or analyze the past 1000 cases you have (100-1000 documents per case), including what county, what opposing lawyer, counsel, plus the 1000 documents in this case, and give you a sense of how valuable this case is or not.
Or if you are a family lawyer, actually mediating a divorce.
Things like this are what actually useful specialized AI legal products do or at least help with.
Claude is very far away from being able to handle most of these things. It is a jack of all trades tool. Will it be able to do this someday? Maybe.
Additionally, keep in mind most legal startups i've run into are based on caricatures of what lawyers do (IE startups who think that most personal injury lawyers are running around after auto cases and trying to be high volume, etc).
Any lawyer who has deal with legal startups could very quickly tell you which will make it or not, because it's pretty consistent which solve real problems that will be hard to commoditize through things like claude for legal.
A life of every thin wrapper company will be the same. Anthropic/OpenAI will just cut the middle-man as soon as they see potential.
I suspect that will happen faster than they'd like, because regulators (at least outside the US) are not interested in a repeat of Google/Amazon/Facebook/etc.
Landlords, tenants, vendors, business and former romantic partners, clients, banks, even your local gym is way more likely to try to fuck you over than the government is.
If a governmental employee gets the address wrong, gets the name wrong, accidentally knocks the mail in the trash, lazily marks the job as complete without sending anything, etc, the burden shifts to the poor person to prove not just that they didn't receive the mail but that the sending office didn't behave correctly.
Other cases behave similarly. In a he-said/she-said, the government wins.
You can sue both in small claims court. You don't need money or a lawyer for that.
I dont mean 'frivolous' like prisoners who file pro-se about their ice cream melting [1], but a level or two above that , that costs time and money to produce records and testimony to defend, even if nary a dime is paid out. Basically ask GPT to figure out the terms and theories to file to get your lawsuit accepted, and done by poor people who cannot afford to post $ or repay if they lose. aka "asymmetric warfare" that benefits the little guy, just like the kind private equity or other terrible corporations wield against the poor via"mandatory arbitration" clauses or damages caps and similar rules that always benefit corporations.
1. https://www.deseret.com/1994/3/21/19098386/melted-ice-cream-...
First step out of line and that account along with anything remotely connected will be banned to oblivion.
Given they share models on Azure, Anthropic will have someone at Microsoft on speed dial.
I've even seen disconnected commit hashes disappear during their security responses which the repo owner has no way of removing.
I half-suspect they threatened him and he stuck to his guns.
https://www.theverge.com/podcast/917029/software-brain-ai-ba...
(Search for "another example" for the relevant section)
er, wait