Kama Thuo, PLLC
- Engineering & Technology Law Firm -
    • Patents
    • AI Counseling
    • Wireless Telecom
    • Trademarks
    • Software Licensing
    • Trade Secrets
    • Patent Analysis
    • Patent Prosecution
    • Patent Licensing
    • Patent Enforcement
    • AI for Patent Law
    • AI Legal Evaluation
    • AI Legal Compliance
    • AI Legal Compliance Tools
    • Wireless Telecom
  • About
    • Client Login
  • Contact Us
  1. Kama Thuo, PLLC | News & Insights
  2. AI Law
  3. AI for Patent Law

AI in Patent Analysis & Tech Transactions

Details
Category: AI for Patents
  • AI

In the rapidly evolving field of AI, Kama Thuo, PLLC offers comprehensive legal services to help clients navigate the intricate landscape of AI development and application. This includes strategic counsel on patenting AI technologies, managing AI risks, and ensuring compliance with privacy and other legal requirements. Kama Thuo focuses on fostering innovation while protecting intellectual assets by helping clients leverage AI technologies to gain competitive advantages and achieve long-term success.

Navigate the below articles to learn more about:

  • Machine Learning in Patent Analysis.
  • Best AI Foundation Models for Patent Analysis.
  • How to Train AI Models for Patent Analysis.

See also patent analysis preferred non-lawyer vendors and tools:

  • AI Automation Vendor: Rfwel Engr AI Group
  • Patent Analysis AI Tool: patanal.ai 
  • Patent Analytics Vendor: Patent Analytics, Inc
  • Wireless Technology Consultants: Rfwel Engr WDI Research

 

Which foundation AI models are best for patent analysis?

Details
Category: AI for Patents

Foundation models are a type of artificial intelligence (AI) model that is trained on a massive amount of unlabeled data. This allows them to learn general patterns and relationships in the data, which can then be fine-tuned for specific tasks such as patent prior-art searches, patent classification, and other patent analysis tasks. They are called "foundation" models because they can be used as a basis for building other AI models.

Key characteristics of foundation models:

  • Trained on massive datasets: Foundation models are trained on very large datasets, often containing billions or even trillions of data points. This allows them to learn complex patterns and relationships that would be difficult to learn from smaller datasets.
  • General purpose: Foundation models are designed to be general purpose, meaning they can be used for a wide range of tasks. This is in contrast to traditional AI models, which are often designed for specific tasks.
  • Adaptable: Foundation models can be adapted to specific tasks by fine-tuning them on smaller datasets that are specific to the task. This allows them to be used for a wide range of tasks without having to be retrained from scratch each time.
  • Powerful: Foundation models have demonstrated impressive performance in natural language processing which is at the heart of any patent analysis task.

Foundations models that are best suited for patent analysis are those that can process and understand large amounts of textual data. These include, for example:

 

Transformer-based models:

  • BERT (Bidirectional Encoder Representations from Transformers): This model is adept at understanding the context and nuances of language, making it well-suited for analyzing complex patent claims and identifying relevant prior art. Includes BERT variants such as SciBERT (trained on scientific text), LegalBERT (trained on legal text), and RoBERTa (Robustly Optimized BERT Pre-training Approach - an improved version of BERT with more training data and longer training time.)
  • GPT-3, GPT-4 (Generative Pre-trained Transformer): While primarily known for text generation, OpenAi's GPTs' powerful language understanding capabilities can be leveraged for prior art search tasks, especially when generating relevant keywords or summarizing patent documents.
  • T5 (Text-to-Text Transfer Transformer): T5 is a versatile model that can be trained for various NLP tasks in a text-to-text format. For example, T5's flexible architecture can be fine-tuned for specific prior art search tasks, such as retrieving documents with similar claims or classifying patents into relevant categories.
  • LLaMA 2 (Large Language Model Meta AI): Llama 2 is designed to be a powerful, scalable model suitable for a variety of natural language processing tasks, including those involving legal and technical text. Compared to some larger models, Llama 2 can be more computationally efficient.
  • Falcon is an open-source family of state-of-the-art large language models (LLMs) developed by the Technology Innovation Institute (TII) in Abu Dhabi. It comes in different sizes (7B, 40B, 180B parameters), catering to diverse computational needs and use cases. The Falcon-40B model, in particular, has been praised for its performance and relatively small size.

 

Sentence Transformer models:

  • These models are specifically designed for generating meaningful sentence embeddings, which can be used to calculate semantic similarity between patent claims and prior art documents. This allows for more accurate and efficient identification of relevant prior art.
  • For example, SBERT (Sentence-BERT) is a modification of BERT that is optimized for generating sentence embeddings, making it ideal for semantic similarity tasks. See also SciBERT-Sentence, LegalBERT-Sentence. 

 

Other techniques:

  • Word2Vec and GloVe: While not strictly foundation models, these word embedding algorithms can be used as a starting point for building custom prior art search systems.
    • They can help identify semantically similar words and phrases, which can be used to broaden the search and uncover relevant prior art that might otherwise be missed.
    • They are used to represent words as vectors and capture semantic relationships between words based on their context in a large corpus of text.
    • Word2Vec is generally faster, focuses on local context windows, and uses a shallow neural network, while GloVe considers global co-occurrence statistics and uses matrix factorization.
  • GloVe (Global Vectors for Word Representation):
    • GloVe is a count-based model. It constructs a large matrix of word co-occurrences, where each cell represents how frequently a pair of words appears together within a certain window in a corpus. It then applies matrix factorization techniques to reduce this high-dimensional co-occurrence matrix into a lower-dimensional representation. The output is a dense vector representations of words where the distance between vectors captures semantic similarity.
    • Training GloVe utilizes a weighted least squares model to learn embeddings that minimize the difference between the dot product of word vectors and the logarithm of their co-occurrence probability.
    • GloVe captures both global statistics and local context, leading to potentially more meaningful representations. It is efficient for creating large vocabulary embeddings.
  • Word2Vec (Word to Vector):
    • Word2Vec is a predictive model that learns word embeddings by trying to predict a target word given its context (Continuous Bag of Words or CBOW) or predict context words given a target word (Skip-gram).
    • Word2Vec focuses on local context, considering only the words surrounding the target word within a fixed window size.
    • Word2Vec is computationally efficient and can be trained on large corpora relatively quickly.

 

Fine-tuning for patent data:

It's important to note that while these foundation models are powerful, they generally need to be fine-tuned on patent-specific data to achieve optimal performance for prior art searching. This involves training the models on large datasets of patents and non-patent literature to improve their understanding of patent-specific language and terminology.

Machine Learning for Patent Analysis

Details
Category: AI for Patents
  • AI

Machine learning, a subset of artificial intelligence, involves training models (typically foundation models) to learn from data and make predictions or decisions without being explicitly programmed. There are various types of learning paradigms in machine learning where each paradigm has distinct characteristics and application to patent analysis.

Supervised Learning: Best for tasks where labeled data is available and necessary for training, such as patent classification, summarizing patents, and claim construction. These models use labeled datasets to learn mappings from inputs to outputs and make accurate predictions on new data. Learn More.

Unsupervised Learning: Ideal for tasks with large unlabeled datasets, like prior art searching, where the goal is to discover hidden patterns or similarities. Clustering and topic modeling can help uncover relationships between documents without the need for labeled data. Learn More.

Semi-Supervised Learning: Suitable for tasks where labeled data is scarce but can be augmented with unlabeled data, such as infringement identification. This approach leverages a small amount of labeled data to guide the learning process on a larger set of unlabeled data, improving model accuracy. Learn More.

Reinforcement Learning: Useful for optimizing workflows and decision-making processes in patent analysis, where sequential actions and resource allocation are crucial. It can dynamically adjust patent classification, optimize prior-art searching strategies, automate infringement identification, and refine claim construction based on feedback and rewards. Learn More.

Transfer Learning: Involves using pre-trained models on large datasets and fine-tuning them on smaller, related datasets, making it particularly useful for patent classification and summarization when labeled data is limited. Learn More.

Federated Learning: A decentralized approach where multiple entities collaborate to train models without sharing their data, enhancing model performance through diverse datasets while maintaining data privacy. This is ideal for collaborative patent analysis across organizations. Learn More.

Active Learning: An iterative process where the model selects the most informative samples for labeling, minimizing labeling effort while maximizing performance. It can be applied to challenging cases in patent classification, claim construction, prior-art searching, and infringement analysis to improve accuracy with fewer labeled examples. Learn More.


Supervised Learning

In supervised learning, the model is trained on a labeled dataset. Each training example is paired with an output label, and the model learns to map inputs to the corresponding outputs.

Characteristics:

  • Requires a large amount of labeled data for training.
  • The goal is to learn a mapping from inputs to outputs and make predictions on new, unseen data.
  • Common algorithms include: Linear regression, logistic regression, decision trees, support vector machines, and neural networks.

 

Application of Supervised Learning to Patent Analysis

 
Patent Classification

Patent classification which entails categorizing patents into technology domains (different from existing classifications such as CPC) typically relies on a pre-labeled dataset where patents are already categorized. Supervised learning models can be trained on this labeled data to accurately predict the category of new patents.

Uses algorithms like Support Vector Machines (SVM) or Neural Networks to classify patents based on labeled training data where each patent is associated with a category.

Patent Summaries

Summarization tasks benefit from supervised learning because the model needs to learn the mapping between the full text and its summary, requiring labeled pairs of patent documents and their summaries.

Uses models like Sequence-to-Sequence (Seq2Seq) with Attention Mechanisms or Transformers (e.g., BERT, GPT) that are trained on labeled datasets where summaries are provided for each patent.

Claim Construction

Claim construction, which involves construing the meaning of specific terms and phrases used in claims in light of the specification, benefits from supervised learning on labeled datasets where the meaning of claim terms has been annotated. The meaning of claim terms used for the labeled dataset can be obtained from court opinions (Markman rulings), PTAB decisions, or from experts (a POSITA and/or legal practitioner).

Uses Natural Language Processing (NLP) models like BERT or RoBERTa fine-tuned on labeled datasets where claims are annotated with their construction.


Unsupervised Learning

In unsupervised learning, the model is trained on unlabeled data. The model tries to learn the underlying structure or distribution in the data without explicit labels. Useful in clustering and dimensionality reduction tasks. 

Characteristics:

  • Does not require labeled data.
  • The goal is to discover hidden patterns, groupings, or features in the data.
  • Common algorithms include: K-means clustering, hierarchical clustering, principal component analysis (PCA), and autoencoders.

 

Application of unsupervised Learning to Patent Analysis

 
Prior-Art Searching

Prior art searching involves large datasets where labeled data might not be readily available. Clustering and topic modeling can help uncover relationships and similarities between documents based on their content. Unsupervised learning for prior-art searching models can use clustering or topic modeling (e.g., Latent Dirichlet Allocation - LDA) to group similar documents based on textual content. This helps in discovering prior art without needing labeled data.


Semi-supervised Learning

Semi-supervised learning combines elements of both supervised and unsupervised learning. The algorithm is typically trained on a small amount of labeled data and a large amount of unlabeled data. Useful in scenarios where labeled data is expensive or time-consuming to obtain.

Characteristics:

  • Uses labeled and unlabeled data to improve learning efficiency.
  • Leverages the small labeled dataset to guide the learning process on the larger unlabeled dataset.
  • Common algorithms include variants of supervised learning algorithms modified to handle unlabeled data, self-training, co-training, and generative models.

 

Application of semi-supervised Learning to Patent Analysis

 
Infringement identification/ searching for licensing targets

Infringement identification often has limited labeled data due to the complexity and specificity of each case. Semi-supervised learning can leverage a small amount of labeled data along with a larger corpus of unlabeled data to improve the model's accuracy. Labeled examples of mapping between patent claims and product literature can be used to train models like semi-supervised neural networks or self-training algorithms.


Reinforcement Learning

Reinforcement learning (RL) is a type of machine learning where an agent learns to make decisions by performing actions in an environment to maximize some notion of cumulative reward. RLHF (Reinforcement Learning from Human Feedback) incorporates human feedback in the rewards function, so the ML model can perform tasks more aligned with human goals, wants, and needs.

Characteristics:

  • The agent interacts with the environment and receives feedback in the form of rewards or penalties.
  • The goal is to learn a policy that maximizes the cumulative reward over time.
  • Common algorithms include: Q-learning, Deep Q-Networks (DQN), and Proximal Policy Optimization (PPO).

 

Application of Reinforcement Learning to Patent Analysis

 
Patent Classification

Reinforcement learning can be used to dynamically adjust the patent classification process based on feedback. For example, a model could learn to prioritize certain sections of the patent specification that, read together with the claims, are more indicative of specific categories, thereby improving classification accuracy over time.

Prior-Art Searching

An RL agent can be trained to optimize the search strategy by exploring various paths and techniques to identify relevant prior art more efficiently. The agent receives rewards for finding high-relevance documents, leading to more effective search processes.

Infringement Identification

Similarly, RL can be applied to automate and optimize the comparison of patent claims with details regarding a target product or process. The agent can learn to focus on key features or sections of the product literature that are more likely to indicate infringement, improving the efficiency and accuracy of the identification process.

Claim Construction

Reinforcement learning can be utilized to refine the process of claim interpretation by updating the model based on new District Court, ITC, and PTAB decisions. The agent can receive rewards for accurately construing claims based on expert feedback, thereby improving its ability to interpret complex claim language over time.


Transfer Learning

Transfer learning involves taking a pre-trained model on a large dataset and fine-tuning it on a smaller, related dataset. This approach is particularly useful when there is limited labeled data available for the specific task.

 

Application of Transfer Learning to Patent Analysis

 
Patent Classification & Summarization

Pre-trained language models like BERT or GPT can be fine-tuned on patent datasets to improve classification accuracy and generate high-quality summaries, even with limited labeled data.

Learn More


Federated Learning

Federated learning is a decentralized approach where multiple entities collaborate to train a model without sharing their data. Each entity trains a local model on its own data and only shares model updates with a central server.

Different organizations (e.g., law firms, patent analytics firms, research institutions, etc.) can collaborate to train robust patent analysis models while keeping their proprietary data private. This approach enhances model performance through diverse datasets without compromising data privacy.

 


Active Learning

Active learning is an iterative process where the model actively selects the most informative samples for labeling. This approach minimizes the labeling effort while maximizing model performance.

 Active learning can be used to identify the most challenging cases of patent classification, claim construction, prior-art or infringement anlysis, etc., for expert review. By focusing on these cases, the model can improve its accuracy with fewer labeled examples.

 


Summary

Kama Thuo, PLLC collaborates with various AI vendors to efficiently tackle AI-based patent analysis matters. The table below highlights a hypothetical collaboration on typical patent analysis tasks.

 

AI-based patent analysis
Patent Analysis Task
Legal Analysis 
(KTH patents & AI counseling)
Patent Analytics Vendor
(Preferred: Patent Analytics, Inc)
AI Automation Vendor
(Preferred: Rfwel Engineering, LLC)
Invalidity Search Provide legal opinions on patent invalidity Generate invalidity search reports using patent databases & AI tools Train/fine-tune AI models to identify correct prior art and different types of NPL prior art such as system art. 
Freedom-to-operate/ Clearance Search Analyze legal risks for potential infringement and evaluate design-around modifications Compile relevant patent data and generate clearance reports Fine-tune AI models to streamline clearance searches
Patent Acquisition Search Legal due diligence for patent acquisitions Provide data on patent portfolios and potential acquisition targets Automate patent portfolio analysis using AI and integrate workflows
Infringement Identification/ Licensing Target Search Legal assessment of potential infringements and licensing viability Extract and compile relevant patent and product literature Integrate tools to search product documentation and apply to models to compare against target claims.
Patent Classification Analyze proposed technology classes and methodology to capture claim scope and pre-sort based on existing CPCs Generate and optimize labeling datasets for patent classification Prepare and label patents for supervised learning models
Patent/Family Summaries Analyze and revise generated summaries for reinforcement learning Generate detailed patent family reports Fine-tune NLP models to create concise patent summaries
Patent Excavation Study Provide legal insights and implications of findings, for example, in monetization Prepare patent portfolio for study Enhance data mining processes with AI-driven techniques
Technology Landscape Study Analyze and provide legal context for technology trends Generate reports on technology trends and patent landscapes Integrate data sources with AI-based tools to discover and analyze trend data
Claim Construction Interpret and provide legal opinions on likely claim meanings Gather relevant official claim constructions Apply relevant claim constructions to fine-tune NLP models to predict reasonable construction for new terms
Claim Amendment Study/Targeted Prosecution Legal strategy for claim amendments and continued prosecution and analyze potential reads on target products or standards Assist with prosecution history extraction and data pre-processing; provide reports on read on products or standards Integrate with APIs on prosecution data and other analysis AI tools

 

Contact us for any questions or clarifications.

Preferred support vendors:

AI Automation Vendor (Rfwel Engr AI Group) | Patent Analysis AI Tool (patanal.ai) | Patent Analytics Vendor (Patent Analytics, Inc) | Wireless Technology Consultants (Rfwel Engr WDI Research)

 

 

  • Kama Thuo Home
  • News & Insights
  • Professionals
  • Sitemap

Popular Tags

Patent Enforcement 4 AI 4 Amazon IP Disputes 3 Equipment Certification 3 Compliance 3 Spectrum 3 Patent documents 2 Antennas 2

Other Articles

  • Understanding HOA Restrictions on External Antennas: Legal Considerations for Homeowners and HOAs
  • Overview of FCC's Frequency Licensing in the United States
  • FCC ID Labeling Requirements
  • Spectrum Allocation in Part 88 Unmanned Aircraft Control Service (UACS)
  • Spectrum Allocation in Part 96 Citizens Broadband Radio Service (CBRS)
  • Broadband Data Collection (BDC) Certification Criteria
  • Professionals
  • Contact Us