The disclosed implementations relate generally to valuation of patents, and more specifically to systems and methods, for patent valuation using artificial intelligence. This application claims the benefit of U.S. Provisional Application No. 63/169,137 filed 31 Mar. 2021, the entire disclosure of which is hereby incorporated by reference herein.
Patents are valuable assets for companies worldwide. Mergers and acquisitions are frequently hinged on valuing patents, issued or pending. Even before a patent is published, or sometimes even before they are written, companies have a vested interest in estimating the value of patents, before investing sizable amounts of money to draft and prosecute patents. Conventional methods of valuing patents have relied on manual reviews. With the advent of machine learning, some techniques have emerged for automating patent valuation. But such traditional methods have relied on bibliographic entries of published patents as explanatory variables. Some techniques only process numeric data, such as number of claims, number of independent claims, number of characters in the claims, and so on, as explanatory variables. Some conventional methods use number of citations for valuing patents. However, since the number of citations requires a certain amount of time to become stable, it is not a reliable indication of the value of a recently filed patent. At the same time, the most valuable patents for strategic reasons include those that have just been applied for or have not yet been published. There are currently no known tools that can predict value metrics for patents, such as the number of citations, the citation score based on citation information including the number of citations, or the patent score calculated based on various indexes related to values of patents.
In addition to the problems set forth in the background section, there are other reasons where an improved system and method of patent valuation using machine learning are needed. For example, there are machine learning-based techniques that use text data of the patent as explanatory variables to determine if a patent is necessary. But machine learning techniques that are performed after natural language processing of text result in poor classification accuracy. Even though recurrent neural networks are able to partially overcome such problems, deep learning typically requires a large amount of labeled data and substantial computing power.
The present disclosure describes a system and method that addresses some of the shortcomings of conventional methods and systems. The disclosure describes techniques for using patent classification codes (and other non-numeric data) in patent applications, to create a sparse matrix that is suitable for efficient machine learning. One or more machine learning models are trained to predict patent valuation based on user-provided metrics. For example, the system uses information on the presence or absence of patent classification codes assigned to other patents in a training dataset, to train the machine learning models. This information is encoded into a large sparse matrix, which is used for efficient training of the machine learning models trained to evaluate or predict the value of the patents. In this way, the techniques enable fast and accurate evaluation or prediction of patents with limited computing power and without the need for large amounts of labeled data.
In accordance with some implementations, a method for training a machine learning model for patent valuation executes at a computing system. Typically, the computing system includes a single computer or workstation, or plurality of computers, each having one or more CPU and/or GPU processors and memory. The method of machine learning modeling implemented does not generally require a computing cluster or supercomputer.
The method includes obtaining an ordered master list of n distinct patent classification codes c1, c2, . . . , cn used by a patent office to characterize patent subject matter. Note that this is not necessarily all of the classification codes used by the patent office. The method also includes performing a sequence of steps for each of a plurality of patents issued by the patent office. The sequence of steps includes obtaining a respective set of one or more patent classification codes assigned to the respective patent by the patent office. Each of the patent classification codes in the respective set matches a respective patent classification code in the master list. The sequence of steps also includes forming a respective training vector comprising n elements. The ith element in the respective training vector is a categorical variable that specifies whether the patent classification code ci is included in the respective set of one or more classification codes. The sequence of steps also includes receiving a respective user-specified value metric for the respective patent. The method also includes generating a training data table comprising the training vectors, and training a machine learning model according to the training data table. The machine learning model is configured to predict value metrics for patents according to their corresponding patent classification codes.
In some implementations, the method further includes, for each of the plurality of patents issued by the patent office: extracting additional information from the respective patent, where the additional information includes non-numeric data; converting the non-numeric data in the respective patent to additional categorical variables; and generating the training data table further based on the additional categorical variables.
In some implementations, the additional information includes classification codes used by the patent office to search for prior art related to the respective patent.
In some implementations, the method further includes, after training the machine learning model: obtaining a new patent issued by the patent office; obtaining a new set of one or more patent classification codes assigned to the new patent by the patent office, where at least one patent classification code cn+1 in the new set does not match any patent classification code in the master list; updating each training vector in the training data table to include another element corresponding to the at least one patent classification code cn+1; updating the training data table to include a new training vector comprising n+1 elements for the new patent. The ith element in the new training vector is a categorical variable that specifies whether the patent classification code ci is included in the new set of one or more classification codes; and retraining the machine learning model according to the updated training data table. In this way, when a new classification code that has not been used in the training data is found, the new classification code is added to the sequence in the training data, and patents with the new classification code are added to the training data.
In some implementations, the machine learning model is further configured to output a respective confidence level for each predicted value metric. The respective confidence value for a respective patent is based on a percentage of patent classification codes for the respective patent that are included in the training data table. In some implementations, to calculate the respective confidence level, the method uses information on the presence or absence (e.g., represented using binary values 1 and 0) of elements of categorical variables, such as patent classification codes and applicant names. It is equally important to have information that the element is present (1) as well as information that the element is absent (0). Therefore, if there is no information on either the presence or absence (1, 0), machine learning would not be able to make predictions accurately. However, some predictions will be made even though they are not accurate. In that case, the predicted value will be unreliable, because the prediction does not include a basis for the prediction. Although it is only a matter of comparison, it is still meaningful to compare the reliability or confidence level, by referring to the richness of the patent classification codes. In some implementations, the method evaluates the accuracy of the prediction results using metrics, such as R2 (sometimes called R-squared). Some implementations output particular bibliographic information that contributed to the accuracy of the prediction by referring to feature importance (e.g., Gini importance). In many cases, some bibliographic data of patents to be predicted is missing, but the importance and the degree of missing data can be used to evaluate the reliability.
In some implementations, the method further includes, in accordance with a determination that the respective confidence level is below a minimum confidence level threshold: obtaining a new patent issued by the patent office and obtaining a new set of one or more patent classification codes assigned to the new patent by the patent office. At least one patent classification code cn+1 in the new set does not match any patent classification codes in the master list. The method extends the master list to incorporate the one or more patent classification codes assigned to the new patent, including the at least one patent classification code cn+1; updates each training vector in the training data table to include another element corresponding to the at least one patent classification code cn+1; updates the training data table to include a new training vector comprising n+1 elements for the new patent, where the ith element in the new training vector is a categorical variable that specifies whether the patent classification code ci is included in the new set of one or more classification codes; and retrains the machine learning model according to the updated training data table.
In some implementations, the machine learning model includes one or more of: Support Vector Regression, Light Gradient Boosted Machine regression, Random forest regression, Binary or multi-valued classifiers, and Deep neural networks.
In some implementations, the master list of patent classification codes is extracted from the plurality of patents.
In another aspect, a method is provided for using machine learning for patent valuation, and the method executes at a computing system. Typically, the computing system includes a single computer or workstation, or plurality of computers, each having one or more CPU and/or GPU processors and memory. The method of using machine learning does not generally require a computing cluster or supercomputer. The method also includes obtaining an ordered master list of n distinct patent classification codes c1, c2, . . . , cn used by a patent office to characterize patent subject matter. The method also includes obtaining a patent that requires valuation. The method also includes obtaining a set of one or more patent classification codes for the patent. One or more of the patent classification codes in the set match a respective patent classification code in the master list. The method also includes forming an input vector comprising n elements. The ith element in the input vector is a categorical variable that specifies whether the patent classification code ci is included in the set of one or more classification codes. The method also includes predicting and outputting a value metric for the patent according to a trained machine learning model that has been trained to predict value metrics for patents according to their respective patent classification codes and user-supplied value metrics.
In some implementations, the method further includes extracting additional information from the patent, where the additional information includes non-numeric data. Forming the input vector further includes converting the non-numeric data in the patent to additional categorical variables, and including the additional categorical variables in the input vector.
In some implementations, the method further includes, in accordance with a determination that the patent is an unpublished patent that lacks patent classification codes: estimating the set of one or more patent classification codes for the patent based on one or more attributes of the patent, such as inventor information, applicant information, and subsidiary information.
In some implementations, the trained machine learning model includes one or more of: Support Vector Regression; Light Gradient Boosted Machine regression; Random forest regression; Binary or multi-valued classifiers; and Deep neural networks.
In some implementations, the master list of patent classifications and the trained machine learning model are compressed and stored in a compressed file. The method further includes decompressing the compressed file to retrieve the trained machine learning model and the master list of patent classifications.
In some implementations, the master list of patent classifications and the trained machine learning model are serialized into one or more byte streams. The method further includes deserializing the one or more byte streams to retrieve the master list of patent classifications and the trained machine learning model.
In some implementations, a computing system includes one or more computers. Each of the computers includes one or more processors and memory. The memory stores one or more programs that are configured for execution by the one or more processors. The one or more programs include instructions for performing any of the methods described herein.
In some implementations, a non-transitory computer readable storage medium stores one or more programs configured for execution by a computing system having one or more computers, each computer having one or more processors and memory. The one or more programs include instructions for performing any of the methods described herein.
In another aspect, a method is provided for training a machine learning model for patent valuation. The method includes obtaining a list of patent applications for training the machine learning model. The method also includes: for each of a plurality of patent applications in the list of patent applications: vectorizing a respective set of text strings for the respective patent application to form a training vector consisting of m dimensions using natural language processing. Each element in the training vector is a numerical value representing one of m distinct textual features of the text strings; and receiving a respective user-specified value metric for the respective patent. The method also includes generating a training data table comprising the training vectors, and training the machine learning model according to the training data table, the machine learning model configured to predict value metrics for patent applications according to the numerical values of textual features.
In some implementations, the method further includes for each of the plurality of patent applications: extracting additional information from the respective patent application. The additional information includes numeric data and non-numeric data; converting the non-numeric data in the respective patent to categorical variables; and generating the training data table further based on the numeric data and the categorical variables.
In some implementations, the additional information includes classification codes used by a patent office to search for prior art related to the respective patent application.
In some implementations, the additional information includes bibliographic data for the respective patent application.
In some implementations, the method further includes, after training the machine learning model: obtaining a new patent application; vectorizing a respective set of text strings for the new patent application to form a training vector consisting of m dimensions using natural language processing. Each element in the training vector is a numerical value representing one of m distinct textual features of the text strings; receiving a new user-specified value metric for the new patent application; updating the training data table to include the training vector; and retraining the machine learning model according to the updated training data table.
In some implementations, the machine learning model includes one or more models selected from the group consisting of: Support Vector Regression; Light Gradient Boosted Machine regression; Random forest regression; Binary or multi-valued classifiers; and Deep neural networks.
In another aspect, a method is provided for using machine learning for patent valuation. The method includes obtaining a patent application that requires valuation; obtaining a set of text strings for the patent application; vectorizing the set of the text strings for the patent application into an input vector consisting of m dimensions using natural language processing. Each element in the input vector is a numerical value representing one of m distinct textual features of the text strings; and predicting and outputting a value metric for the patent application according to a trained machine learning model that has been trained to predict value metrics for patent applications according to numerical values of textual features and user-supplied value metrics.
In some implementations, the method further includes: extracting additional information from the patent application. The additional information includes numeric data and non-numeric data. Forming the input vector further includes converting the non-numeric data in the patent application to categorical variables, and including the numeric data and the categorical variables in the input vector.
In some implementations, the additional information includes classification codes used by the patent office to search for prior art related to the patent application.
In some implementations, the additional information includes bibliographic data for the patent application.
In some implementations, the trained machine learning model includes one or more models selected from the group consisting of: Support Vector Regression; Light Gradient Boosted Machine regression; Random forest regression; Binary or multi-valued classifiers; and Deep neural networks.
In another aspect, a system is provided for patent valuation using machine learning. The system includes a database storing targeted patent data records used for training, testing, and/or validation of one or more machine learning models. The system also includes a training data extraction module for extracting data for training a machine learning model from the targeted patent data records. The extracted data includes one or more objective variables to predict value of patent applications. The system also includes a patent text data extraction module for extracting text data contained in documents for patent applications. The system also includes a training data generation module for generating training data for training one or more machine learning models. The training data generation module forms training vectors by converting the text data into a vector using natural language processing. The system also includes a module to train one or more machine learning models using the training data generated by the training data generation module. The system also includes a test data extraction module, which extracts test data from the targeted patent data records. The system also includes a patent text data extraction module, which extracts text data from patent applications in the test data extracted by the test data extraction module. The system also includes a test vector generation module for generating test vectors based on the text data extracted by the patent text data extraction module. The system also includes a module for predicting and outputting a value metric for a patent application using the test vectors according to a trained machine learning model.
Thus methods and systems are disclosed that facilitate patent valuation using artificial intelligence.
For a better understanding of the disclosed systems and methods, as well as additional systems and methods, reference should be made to the Description of Implementations below, in conjunction with the following drawings in which like reference numerals refer to corresponding parts throughout the figures.
Reference will now be made to implementations, examples of which are illustrated in the accompanying drawings. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it will be apparent to one of ordinary skill in the art that the present invention may be practiced without requiring these specific details.
In some implementations, the module 106 also obtains a list of patent classification codes used by a patent office to characterize patent subject matter. In some implementations, the list is an ordered master list of n distinct patent classification codes c1, c2, . . . , cn used by the patent office to characterize patent subject matter. In some implementations, the list includes patent classification codes of different types, such as IPC and CPC.
In some implementations, the module 106 is also configured to extract additional information from the respective patent, such as numeric data and/or non-numeric data. In some implementations, the module 106 is also configured to convert the non-numeric data in the respective patent to additional categorical variables (in addition to the categorical variables obtained from patent classification codes described below). The non-numeric information includes, for example, CPC, IPC, IPC8, F-term (assigned by Japan Patent Office), FI (assigned by Japan Patent Office), applicant name, and inventor name. Numeric information includes, for example, number of families, number of words of specification, number of words of claim 1, number of claims, number of inventors, number of assignees, and number of claimed priority. These examples are described below in reference to
The system 100 also includes a patent classification extraction module 108, which extracts patent classification codes from a given patent (e.g., one of the patents in the records 104). Examples of patent classification codes are described below in reference to
The system 100 also includes a training data generation module 110 for generating training data for training one or more machine learning models. In some implementations, the module 110 forms training vectors. In some implementations, each training vector corresponds to a respective patent issued by the patent office. In some implementations, each training vector includes n elements, where i-th element in the training vector is a categorical variable that specifies whether the patent classification code ci is included in the respective set of one or more patent classification code described earlier. In some implementations, a set of categorical variables included as elements of the training vector is a set of binary values. The set of binary values represents whether or not each patent classification code assigned to any patent is selected for training by the module 106. In other words, the set of categorical variables includes any patent classification code that appears at least once in the training data. Each patent in the training data is assigned a binary value depending on whether a patent classification code is assigned to the patent. For example, if a patent classification code ci which is assigned to Patent X is not assigned to another Patent Y in the data extracted by the module 106, the binary value of a patent classification code ci for Patent Y will be “0.” In some implementations, training data is at least a set of patent classification codes and it may additionally include other non-numerical data and numerical data, examples of which are described below in reference to
In some implementations, as described below in reference to
In some implementations, the module 110 also receives a user-specified value metric for a patent. For example, a user may provide (or assign) a particular value for a patent (or a group of patents) and/or patent classification codes. To further illustrate, the user may statically assign a first value (e.g., a high value) for one type of patents (e.g., pharmaceutical patents) and a second value (e.g., a low value) for another type of patents (e.g., business method patents). In this way, the user provides labels for different patents (as is commonly the case in supervised learning systems). In some implementations, the module 110 also generates a training data table (an example of which is shown in
The system also includes a module 112 to train one or more machine learning models 102 using the training data generated by the module 110. Examples of machine learning models are described below in reference to
The block diagram in
The system 100 also includes a module 122 for predicting and/or outputting a value metric 124 (sometimes called prediction output) for a patent according to a trained machine learning model (e.g., the one or more machine learning models 102), which has been trained to predict value metrics for patents according to their respective patent classification codes and user-supplied value metrics. To use the system 100 for predicting value of patents, the path shown for testing the system 100 (e.g., steps performed by the module 114, the module 116, the module 118, the module 120, and/or the module 122) is used, according to some implementations.
In some implementations, the system 100 also outputs data (shown in column 134) as to whether a concerned patent should be reviewed by a user for validating the predicted output 124 for the patent, to add any missing patent classification codes to the table generated by the module 110, and/or to retrain the machine learning models 102, for improving its accuracy and/or prediction quality. For example, if the confidence level 132 is below a threshold (e.g., 50%), then the patents (or patent applications) are marked for further analysis and/or inclusion in the training data table, and/or the machine learning models 102 are retrained using the updated training data table. As shown in
In some implementations, the module 110 determines that the patent corresponds to an unpublished patent that lacks patent classification codes, and receives user input regarding patent classification codes (e.g., IPC, CPC, FI, or F-term codes) for the patent.
As described above, the techniques described herein use patent classification cods for evaluating and/or predicting the value of a patent by regression. Patent classification codes have a deep hierarchical structure that can classify detailed technologies and can assign many types of technological classifications to each patent. The techniques described herein transform the patent classification codes into categorical variables, which are then transformed into a large sparse matrix with mostly zero-valued cells or elements. For example, when patent classification codes (e.g., CPC or F-term codes) that have a deep hierarchy and that are often assigned to many individual patents, a large sparse matrix is obtained. Since patent classifications are assigned by a Patent Office, a certain level of classification quality is guaranteed. Therefore, by using the patent classification codes as input data for categorical variables, the system described herein obtains high quality large scale sparse matrices for training purposes. This allows the system to obtain regression and classification results with small amounts of data, minimal computing power, and faster prediction accuracy, compared to conventional deep learning techniques.
Some implementations add other information (e.g., numbers and table data obtained using natural language processing) to further improve the accuracy. For example, some implementations extract n-grams, using natural language processing, based on textual information (e.g., abstracts and claims) in patent gazettes, and generate vectors (e.g., a vector of 1s and 0s, with a value of 1 if the patent to be processed contains the n-grams, and 0 if the patent does not contain the n-grams). The result of the classification is used as a categorical variable. Some implementations consolidate a plurality of such results to compute a categorical variable, and use the variable as an explanatory variable for training purposes.
The computing device 300 may include a user interface 306 comprising a display device 308 and one or more input devices or mechanisms 310. In some implementations, the input device/mechanism includes a keyboard. In some implementations, the input device/mechanism includes a “soft” keyboard, which is displayed as needed on the display device 308, enabling a user to “press keys” that appear on the display 308. In some implementations, the display 308 and input device/mechanism 310 comprise a touch screen display (also called a touch sensitive display).
In some implementations, the memory 314 includes high-speed random access memory, such as DRAM, SRAM, DDR RAM or other random access solid state memory devices. In some implementations, the memory 314 includes non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid state storage devices. In some implementations, the memory 314 includes one or more storage devices remotely located from the GPU(s)/CPU(s) 302. The memory 314, or alternatively the non-volatile memory devices within the memory 314, comprises a non-transitory computer readable storage medium. In some implementations, the memory 314, or the computer-readable storage medium of the memory 314, stores the following programs, modules, and data structures, or a subset thereof:
Each of the above identified executable modules, applications, or sets of procedures may be stored in one or more of the previously mentioned memory devices, and corresponds to a set of instructions for performing a function described above. The above identified modules or programs (i.e., sets of instructions) need not be implemented as separate software programs, procedures, or modules, and thus various subsets of these modules may be combined or otherwise re-arranged in various implementations. In some implementations, the memory 314 stores a subset of the modules and data structures identified above. Furthermore, the memory 314 may store additional modules or data structures not described above. The operations of each of the modules and properties of the data structures shown in
Although
In some implementations, the memory 314 also includes modules to train and execute models 102. In some implementations, machine learning algorithms used for creating the one or more machine learning models 102 include LightGBM regression, Random Forest Regression, Support Vector Regression, Linear Regression, Neural Networks, Deep Learning, and/or other regression algorithms.
In some implementations, the trained machine learning models 102 are compressed column-wise along with the master list of patent classification codes, before storing the same in the memory 314. For example, the system 100 compresses the data of the machine learning model and the master list of patent classifications (e.g., using two PKL files), into a single compressed file (e.g., one Zip file). An advantage for compressing the data in this manner is that the training model and the master list of patent classifications are stored as a single unit, facilitating data management and reducing cost of memory.
In some implementations, the master list of patent classifications and/or the trained machine learning models 102 are serialized into one or more byte streams. The system 400 includes one or more modules in the memory 314 used to decompress any compressed data, and/or for deserializing any byte streams, to retrieve the master list of patent classifications and/or the trained machine learning model. Storing the patent classification codes from the training data enables efficient mapping of new patent classification codes, in formulating training and/or input vectors, and targeting patent valuation (for the selected codes). Compressing the data has the obvious advantage of reducing storage costs, for moving data (e.g., to customer sites), as well as efficient data management. Serializing data enables data portability, and enables the system 400 and/or the associated data to be used across computing platforms.
Conventional regression typically uses a population where the explanatory and objective variables are obvious, finds the best regression algorithm, reduces the explanatory variables to the important ones in one or more ways, and tunes the parameters of the regression algorithm to the optimal values. Example techniques for reducing explanatory variables include dimensionality reduction, principal component analysis (PCA), and feature engineering. In the system disclosed herein, non-numeric information, such as patent classification code, are used for predicting the patent value. Conventional methods do not use the detailed information on the technical field. In addition, the IPC and numerical correspondence table are established based on the feature importance level extracted by feature engineering, and the IPC is converted into numerical values using this table as input data.
In some implementations, the machine learning algorithms include LightGBM regression, Random Forest Regression, Support Vector Regression, Linear Regression, neural networks, deep learning, and/or other regression algorithms. Various methods may be used for vectorization of text data, such as a method for vectorizing sentences based on the number of occurrences of words (e.g., tf-idf, or LSI), a method for using distributed representation of words (e.g., Word2Vec, VMD, or LC-RWMD), or a method for using distributed representation of documents (e.g., Dec2Vec, or Sent2Vec). BERT (Bidirectional Encoder Representations from Transformers) may also be adopted. For BERT, the training data generation module first uses a tokenizer (e.g., SentencePiece, or WordPiece) to split a text string into multiple units called “tokens.” The tokens include not only words but also some symbols, such as [CLS] for the beginning of the tokens and [SEP] for the end of them. Then, the module creates a 512-dimensional vector with numerical data assigned for each token. This vector is then passed to a “Transformer Encoder” to create a 768-dimensional vector. In one example, the input data has 512 tokens and the output vector data has 768 dimensions, but various implementations may use any other combinations of the number of tokens and the number of dimensions. Various implementations employ vectorization methods, other than BERT, which take text data as input and convert it into numerical values.
In another aspect, a method is provided for training a machine learning model for patent valuation. The method includes obtaining a list of patent applications (includes at least one of issued patents and unissued patent applications) for training the machine learning model. For each of a plurality of patents applications in the list of patent applications, the method includes: (i) vectorizing a respective data set of text strings for the respective patent application into a training vector consisting of m dimensions using natural language processing. The ith element in the training vector is a numerical value representing one of m distinct textual features (e.g., a token, a word, or a phrase) of the text strings; and (ii) receiving a respective user-specified value metric for the respective patent application. The method also includes generating a training data table comprising the training vectors. The same ith elements out of multiple training vectors of the patent applications in the training data table are possible to be numerical values that are generated based on different tokens for each row since the training vectors are created after splitting text strings into multiple different sets of tokens depending on the patent documents. The method also includes training a machine learning model according to the training data table. The machine learning model is configured to predict value metrics for patent applications according to the numerical values of textual features.
In another aspect, a method is provided for using machine learning for patent valuation. The method includes obtaining a patent document (which may be an issued patent or an unissued patent application) that requires valuation. The method also includes obtaining a set of one or more text strings for the patent application. The method also includes vectorizing a respective data set of the text strings into an input vector consisting of m dimensions using natural language processing. The ith element in the training vector is a numerical value representing one of m distinct textual features (e.g., a token, a word, or a phrase) of the text strings. The method also includes predicting and outputting a value metric for the patent according to a trained machine learning model that has been trained to predict value metrics for patents according to numerical values of textual features and user-supplied value metrics.
Additional columns 1108 may include vectorized data generated from the text data (e.g., 512 dimensions or 768 dimensions). Columns to the right of label 1114 are added and values are updated. Note that the “ . . . ” symbol drawn in
The terminology used in the description of the invention herein is for the purpose of describing particular implementations only and is not intended to be limiting of the invention. As used in the description of the invention and the appended claims, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, steps, operations, elements, components, and/or groups thereof.
The foregoing description, for purpose of explanation, has been described with reference to specific implementations. However, the illustrative discussions above are not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The implementations were chosen and described in order to best explain the principles of the invention and its practical applications, to thereby enable others skilled in the art to best utilize the invention and various implementations with various modifications as are suited to the particular use contemplated.
According to 17th aspect, a method for training a machine learning model for patent valuation, the method comprising:
According to 18th aspect, the method according to 17th aspect, further comprising for each of the plurality of patent applications:
According to 19th aspect, the method according to 18th aspect, wherein the additional information includes classification codes used by a patent office to search for prior art related to the respective patent application.
According to 20th aspect, the method according to 18th aspect, wherein the additional information includes bibliographic data for the respective patent application.
According to 21th aspect, the method of any of 17th to 20th aspect, further comprising:
According to 22th aspect, the method of any of 17th to 20th aspect, wherein the machine learning model includes one or more models selected from the group consisting of:
According to 23th aspect, a method for using machine learning for patent valuation, the method comprising:
According to 24th aspect, the method of 23th aspect, further comprising:
According to 25th aspect, the method of any of 24th aspect, wherein the additional information includes classification codes used by the patent office to search for prior art related to the patent application.
According to 26th aspect, the method of any of 24th aspect, wherein the additional information includes bibliographic data for the patent application.
According to 27th aspect, the method of any of 23th to 26th aspect, wherein the trained machine learning model includes one or more models selected from the group consisting of:
According to 28th aspect, a system for patent valuation using machine learning, the system comprising:
The present application is a continuation application based on PCT Patent Application No. PCT/JP2022/016882, filed on Mar. 31, 2022, the entire content of which is hereby incorporated by reference.
Number | Date | Country | |
---|---|---|---|
63169137 | Mar 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2022/016882 | Mar 2022 | US |
Child | 18372731 | US |