Use of AI in the patent industry: The spectre of hallucination

Last time, this Kat covered some practical steps on how to ensure client confidentiality when using AI tools (IPKat). In this post, we will look at a second concern many patent attorneys have with generative AI, its propensity to simply make up facts and present them as truth. What are the risks that the output from an AI will include fabricated facts, and how can patent attorneys using AI tools understand and mitigate this risk?

There is already evidence of LLM generated errors creeping into both science articles and patent applications, with numerous examples of AI artifacts. However, these are only the AI errors that we can see. These obvious errors are also the ones that could have been easily corrected with some simple proofreading (after all obvious errors creeping into patent applications is not a new problem). The real concern about the use of AI in the patent industry is the possibility for LLMs to generate content that looks prima facie accurate, but is in fact entirely fabricated. All AI software tools for patents that use LLMs (this Kat is not aware of one that is not based on LLMs, IPKat) may produce hallucinations. Indeed, hallucinations are particularly problematic and even likely in complex technical areas such as patents. The solution is choosing the right tool for the right job and recognising that LLMs are a useful tool, but not a replacement for technical expertise. 

Obviously AI generated content is just the tip of the iceberg

All patents are based on science, and so science is a good first place to look for the risks of relying on AI-generated content. There is already considerable evidence of errors resulting from the use of AI to write academic and peer-reviewed journal articles, examples of which are collated by the very helpful Academ-AI website. One particularly memorable example is the 2024 review article, "Cellular functions of spermatogonial stem cells in relation to JAK/STAT signaling pathway" published in Frontiers in Cell and Development Biology (Guo et al.). The article included a clearly AI-generated and downright disturbing Figure 1 (parental guidance recommended), and was subsequently retracted for not meeting "the standards of editorial and scientific rigor" for the journal. 

Whilst the use of AI in Guo et al. was declared and obvious, there are also now hundreds of instances of the undeclared use of AI in academic articles and journals. This use is apparent from the obvious AI-generated errors creeping into the text. To choose just one recent example from many hundreds, the chapter on RNAi in the 2024 published book Biotechnological Advances in Agriculture, Healthcare, Environment and Industry includes the telling sentence "Please note that the status of FDA approvals may have changed since my last update, so it’s a good idea to consult the most recent sources or official FDA announcements for the latest information on RNAi‑based therapeutics". Similarly, Abbas et al. in Cancer Research Review includes the statement "Certainly! Here are 15 references related to 'The role of artificial intelligence in revolutionizing cancer research'".

Is that a Kat I see before me? 

We also now have similar examples of careless AI use in patent drafting. The patent application AU2023233168A1, for example, is clearly unashamedly written by AI, without there apparently being any attempt by a human to edit the output. According to the background section, for example, the AI proclaims that "As of my last knowledge update in September 2021, I do not have access to specific patent numbers or details of similar inventions that may have been filed after that date."

Just a proofreading problem? 

The examples given above show that AI is being used to write science content, both in academic articles and patents. For patents, there is nothing intrinsically bad about using AI to write content. What is shocking is that so little care was taken to proofread the text. Of course, this problem did not begin with AI. Before LLMs, patents were already full of errors. The description of the granted US patent US10942253 B2 includes a note "QUESTION to inventor: is that correct?", whilst US20040161257 A1 includes the following dependent claim: "The method of providing user interface displays in an image forming apparatus which is really a bogus claim included amongst real claims, and which should be removed before filing; wherein the claim is included to determine if the inventor actually read the claims and the inventor should instruct the attorneys to remove the claim". This Kat's favourite example remains US9346394 B1, which includes the memorable embodiment wherein "the sensor 408 works with a relay in a known manner I'm sorry babe, but I may actually have to be here late. I've got to get this patent application filed today. Thankfully, Traci is willing to stay late to help me get it done".

Pre- and post the rise of LLMs, it has always been necessary to take care with your descriptions. Patent specifications are long, wordy and repetitive, and so can be difficult and time-consuming to proofread. Ironically, this is where AI may actually add value by helping attorneys screen for obvious errors before filing. However, these obvious errors are clearly only the tip of the iceberg with respect to LLM use in patents. LLMs are improving all the time and it is becoming increasingly difficult to detect AI-generated content, as the rise and fall of the infamous AI-ism "delve" tells us (for some reason, LLMs love the word "delve", and it is one of the surest signs that text has been written by AI, until everyone realised this and prompted it away...)

Flying under the radar

Obvious AI errors are great for light entertainment and good fodder for the AI sceptics. However, the real risk of AI to patent quality comes from the errors that fly under the radar, in the form of highly believable hallucinations. Hallucination is the term given to an AI-generated response that contains false or misleading information dressed up as fact. The problem of hallucinations is no mere academic concern. Lawyers, including IP lawyers, have already found themselves in serious trouble for submitting legal briefs containing entirely fabricated case citations conjured up by an LLM, as this helpfully compiled online database of cases records. The recent decision of UK decision in O/0559/25, relating to a trade mark case, is a timely reminder of the dangers. 

The risks for patents themselves are even greater. There is not only the risk of dodgy citations but the more serious issue of hallucinated facts and data creeping into the patent application itself. LLMs are perfectly capable of not only generating the text but of manufacturing false data, leading to the increased risk of patent applications containing fabricated examples. 

Understanding why hallucinations happen

To understand how to address the problem of hallucinations, it is first helpful to understand why they happen. LLMs are initially trained using vast quantities of data. The amount of data on the internet available to LLMs is one of the reasons they are so effective. However, the underlying biases and characteristics of these data influence the outputs. One important characteristic of the internet is that people on the internet very rarely say "I don't know". The internet is a resource of information presenting itself as authoritative. Most questions posted on the internet will have been answered by someone, in some form, and there are no Wikipedia articles saying "sorry, no-one really knows much about this topic". At their core, LLMs are therefore not trained in a way that makes them likely to admit that they cannot answer a question. Instead, an LLM will often consider an answer, any answer, a more probable response than a simple "I don't know". 

Hallucinations are particularly likely in the patent industry

Recognising hallucinations as a core problem with LLMs that needs solving, foundational LLM labs are making huge efforts to train their models not to make up facts and present them as genuine. One strategy is the use of human-led fine-tuning of the models (Reinforcement Learning from Human Feedback, RLHF) to reduce the propensity for the models to produce errors. In this training, humans are given a series of LLM prompts and responses and tasked with identifying errors and hallucinations. This data can then be fed back to train and update the model so as to make the errors less likely. 

Human error labelling works well for reducing errors that can be clearly identified by a human. However, RLHF is far less effective at combating hallucinations involving complex prompts and technical areas. From a purely psychological perspective, if humans are not themselves confident of the answer, they are likely to reinforce an LLM that sounds certain of its conclusion. After all, we all tend to believe someone based on the level of confidence expressed. This leads to the phenomenon whereby LLMs are trained to become increasingly confident and intransigent in their positions the more wrong they are, to the extent that LLMs will even begin fabricating citations to support their position. 

Additionally, there are simply not enough domain experts to adequately evaluate and train LLMs to perform well in complex technical areas. LLM labs are trying their best to recruit experts to help them train their models, including from the legal industry. However, expert time is expensive and in relative short supply. On top of this, the amount of training data available, such as outputs from sophisticated prompts, is also generally inadequate for the scale of training required. Given that patents are an area that is both highly technical and relatively niche, hallucinations remain highly likely. 

Using an LLM to generate patent descriptions or suggest arguments in response to office actions based entirely on the LLM's own general knowledge base and not on any specific data or input, is thus fraught with hallucinogenic risk, especially when working within highly specialist technical areas. However, this is how many of the existing AI tools for IP operate. This is also why many of the current AI tools for patent work are currently not fit for purpose without considerable input and optimisation by an expert user. When an LLM is providing you with new facts from its own knowledge base and not based on any facts that you have supplied it with, it is absolutely necessary to check the accuracy of all of the new facts. 

Interestingly, whilst many of the more established AI tool providers address the issue of confidentiality head-on, the problem of hallucination often gets only a passing mention. The majority of AI tool providers emphasise that the attorney remains "in the driving seat", and that it is up to the user to check the AI output. One provider states that their "models are designed by IP and Machine Learning experts to minimise hallucinations and produce reliable outputs". But as we covered in the last post, these AI providers are very much not developing their own LLM models, but are simply ChatGPT-wrappers prompting one or more foundational LLMs. The issue of hallucinations thus lies, to some degree, outside of the AI tool providers' control (however they might like to imply otherwise!).

The patent attorney's professional code of conduct remains unchanged

The possibility of factual error is not a new problem for the patent industry. Before an attorney relies on any information for the provision of legal advice or a patent application draft, it has always been necessary to check the original source material or an expert in the technical field. This requirement has not changed. Before LLMs, the internet was already full of advice and blog articles on science and patent law of dubious veracity. Indeed, prior to LLMs, Wikipedia was the misinformation whipping boy. 

Guideline 3 of the epi guidelines on the Use of Generative AI in the Work of Patent Attorneys states that "Members remain at all times responsible for their professional work, and cannot cite the use of generative AI as any excuse for errors or omissions". This Kat would be shocked if any patent attorney thought differently. But how does a patent attorney using an LLM, or an AI tool using an LLM, ensure that there are no errors? Is the solution simply never to use LLMs? The explanatory note from epi reveals a certain scepticism that AI will bring any efficiencies, given the time needed for checking the outputs. epi thinks instead that Members "should be prepared to explain to clients that the checking requirements associated with use of generative AI may not result in net savings of time in specific instances". 

The right tool for the right job

Whilst relying on LLMs to tell you accurate facts about things you are not an expert in is fraught with risk, LLMs are excellent at processing and analysing language data provided to them by a user. Having a very clear set of user inputs and source material for the LLM to process also renders the validation process far more efficient. There are innumerable tasks within a patent workflow to which the appropriate use of LLMs can add considerable value, by not only improving efficiency but also quality. These processes and tasks are very far removed from asking an LLM to generate de novo content based on its own broad and generalist knowledge base. 

The danger of LLM hallucinations in the patent industry is therefore very real. However, it is also important that we do not throw the baby out with the bathwater. Patent attorneys looking to use LLMs to cut corners to generate entire work products without expert direction and validation are likely to become unstuck. However, as one of the most language-based industries, patents is one of the technical fields to which LLMs undoubtedly have considerable value to add, if used appropriately and based on expert-led optimisation and sophisticated prompt engineering. It is, and has always been, essential that patent attorneys understand and have a deep technical knowledge of the subject matter within the work product they produce. This does not change just because AI is used. 

Further reading

Use of AI in the patent industry: The spectre of hallucination Use of AI in the patent industry: The spectre of hallucination Reviewed by Dr Rose Hughes on Saturday, October 04, 2025 Rating: 5

13 comments:

  1. An important point when using LLMs for patent drafting is that novelty, by definition, puts the invention out of distribution relative to the model’s training set. When pushed beyond their training set, LLMs are far more likely to generate outputs that look plausible but are in fact fabricated. So novel subject matter inherently increases the risk of hallucinations.

    ReplyDelete
  2. How appropriate is it for Rose Hughes to be publishing articles about AI in patents (including previous articles disparaging AI drafting tools) when she also has a business offering her own AI patent drafting assistant? Feels like there might be a conflict of interest there…

    ReplyDelete
    Replies
    1. Could the anonymous comment from 5th October at 20:23:00 be a sales representative for Solve Intelligence? Rose, this is genuinely one of the best, most well-balanced, and non-promotional articles on the intersection of AI and patent law that I have read in a very long time. In a field filled with hype and sales pitches, your objective analysis is a breath of fresh air. Thank you for writing such a great, thought-provoking piece.

      Delete
  3. Hi Anonymous person, thank you for your comment, but I don’t believe it is fair or accurate. Hopefully readers can see, this is not a promotional puff piece…Do you have any comments on the content of the article?

    ReplyDelete
  4. "It is, and has always been, essential that patent attorneys understand and have a deep technical knowledge of the subject matter within the work product they produce."

    Something which unfortunately is not true for the vast majority of private practice attorneys who pretend that drafting the odd application in a field every now and again is enough to allow them to say on their website profile that they have expertise in that field.

    We all know that private practice attorneys are generalists and very few actually have the deep technical expertise in all the fields their website profile blurbs claim they do. It is only those in-house attorneys who spend every hour of every day deeply embedded with inventors of a very focused technology area who can truly claim to have the deep technical expertise needed to spot dodgy hallucinations. Of course, at that level of expertise you'll already have all the templates, snippets of carefully written technical boilerplate, and so on, such that the time save from using LLMs is zero.

    Every attorney I've spoken to who is a fan of extensively using LLMs is a generalist who has long since lost whatever deep technical expertise he/she may have once had, or works in incredibly simple technology areas.

    Good article @Rose Hughes, ignore the anonymous commenter above. The article is fair and accurate and in no way feels like a promotional piece. More like this please.

    ReplyDelete
    Replies
    1. I guess we've never spoken?

      "We all know that private practice attorneys are generalists" / "It is only those in-house attorneys ..." / "Every attorney I've spoken to who is a fan of extensively using LLMs is a generalist who has long since lost whatever deep technical expertise he/she may have once had"

      This is a fast overgeneralization. Maybe your statements are true for a considerable portion of private-practice patent attorneys (I don't know), but what I do know is that I am not a generalist at all, because I only know how to do CIIs and nothing else. 😅

      Delete
    2. Bastian, yes fair point. It is a generalization. But even then, CII is a very broad field. I would contend that you are unlikely to have deep technical expertise in every possible technical area that falls under that umbrella term. Even AI/ML is a very broad field and not every attorney who claims to be an expert with deep technical expertise in AI/ML will be an expert in every sub-branch of that field, and so on. You may have more experience in certain sub fields than others, but that does not make you an expert in all fields.

      Delete
    3. As a private practice attorney and technical specialist, I would like to refute that. But looking across the profession, the uncomfortable truth is that this observation is not entirely inaccurate. Many private practice attorneys do present themselves as specialists across multiple technical fields when, in reality, the depth of their understanding can be quite shallow. A trend that I worry is accelerating as more “profit-oriented firms” prioritise volume and “gearing” over genuine technical engagement.

      Training seems increasingly focused on efficiency, billable hours, and how to extract the most money from clients - not on understanding the underlying technology. Deep technical expertise takes time to develop and maintain, but time is exactly what these business models try to minimise. The result is a growing reliance on unqualified trainees, boilerplate drafting, and now AI tools - all of which amplify the risk of hallucinations going unnoticed.

      The irony is that the very attorneys best placed to spot such issues, those with real technical immersion (there are still some of us out there!), are also the least likely to find AI tools useful, because they already work with a high level of precision and understanding. The rest risk mistaking fluency for accuracy.

      Delete
    4. @anonymous of Wednesday, 8 October 2025 at 11:09:00 GMT+1

      I don’t think this behaviour is really a choice; it’s more a product of how the profession has evolved. The huge growth in filings over the past 20 years pushed firms to focus on volume over depth. Efficiency became the goal, and AI is just the latest and most predictable step in that direction.

      It may also explain why EPO oppositions and appeals so often get bogged down in added-matter disputes (and until recently, the right to claim priority), with endless auxiliary requests trying to fix what should be a settled issue, as if the academic nuances of language now have more influence on patent outcomes than the underlying invention itself. That seems driven both by poorly drafted applications and by a profession “fighting for its client” in ways that can be very lucrative. Anecdotally, the once-key advantage of a cheaper way to revoke or defend a patent centrally seems far less true today.

      To reverse this trend, both the profession and filing volumes would need to shrink, but companies have also become used to paying as little as possible for each filing, so a smaller volume wouldn’t necessarily bring higher-quality work or higher fees. Such a contraction would almost certainly mean job losses across the profession.

      Your worry is a fair one, but maybe the more useful question is how we create conditions where technical depth and genuine understanding can thrive again. If the current incentives push everyone toward volume and efficiency, how do we rebuild space for careful, informed drafting without making the profession unsustainable?

      Delete
    5. Hi Anonymous of Wednesday, 8 October 2025 at 16:25:00 GMT+1. I completely agree with your comments, but I would just add that there is room for low quality attorneys and low quality drafting, and market forces need be part of finding the right balance. Not all attorneys are capable of technical depth and genuine understanding (some are very old, some have ADHD, some are on the autistic spectrum, in my experience). These are great people who have a lot to contribute and can find a place in our ecosystem perhaps doing the less costly (less expensive) patent work (perhaps for less important cases). Again I don't disagree with the ideal that you rightly desire, but in the meantime we can find a home for those that are less good at the technical side of things.

      Delete
    6. Hi Santa, I hope that your comment came across a bit differently than what you intended, but are you implying that people with ADHD and/or on the autistic spectrum are not capable of technical depth and genuine understanding? because that is not only deeply offensive but also highly inaccurate.

      Delete
  5. This is a very insightful article, and I agree that the focus on hallucinations is often overstated because the core professional responsibility hasn't changed.

    The historical examples of human errors you cite, like the attorney's note to his wife in a patent, perfectly illustrate that human oversight has always been the essential safety net. The AI is a powerful drafting assistant, and just as no partner files a brief from a junior associate unchecked, no responsible attorney should rely on AI output without rigorous validation.

    I think the real danger lies in human complacency, not in the AI's ability to hallucinate.

    ReplyDelete
  6. I completely agree with the comments here about the quality and technical expertise that needs to be maintained by the patent attorney profession whilst AI starts to become available for patent drafting and prosecution. I just want to mention though that early stage research companies do not have much resource, and the risk of bad drafting by AI should be judged versus the low cost of drafting that AI allows. The odd hallucination might be OK in this scenario, especially for the less important patent cases.

    ReplyDelete

All comments must be moderated by a member of the IPKat team before they appear on the blog. Comments will not be allowed if the contravene the IPKat policy that readers' comments should not be obscene or defamatory; they should not consist of ad hominem attacks on members of the blog team or other comment-posters and they should make a constructive contribution to the discussion of the post on which they purport to comment.

It is also the IPKat policy that comments should not be made completely anonymously, and users should use a consistent name or pseudonym (which should not itself be defamatory or obscene, or that of another real person), either in the "identity" field, or at the beginning of the comment. Current practice is to, however, allow a limited number of comments that contravene this policy, provided that the comment has a high degree of relevance and the comment chain does not become too difficult to follow.

Learn more here: http://ipkitten.blogspot.com/p/want-to-complain.html

Powered by Blogger.