US7792353B2 - Retraining a machine-learning classifier using re-labeled training samples - Google Patents
Retraining a machine-learning classifier using re-labeled training samples Download PDFInfo
- Publication number
- US7792353B2 US7792353B2 US11/555,036 US55503606A US7792353B2 US 7792353 B2 US7792353 B2 US 7792353B2 US 55503606 A US55503606 A US 55503606A US 7792353 B2 US7792353 B2 US 7792353B2
- Authority
- US
- United States
- Prior art keywords
- samples
- training
- sample
- label
- labels
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
Definitions
- the present invention pertains to machine learning and is particularly applicable to systems, methods and techniques for retraining a machine-learning classifier using re-labeled training samples.
- Automated classification of digital information has application in a number of different practical situations, including image recognition (e.g., identifying which photographs from among thousands or millions in a database include a picture of a face or a picture of a particular face), text classification (e.g., determining whether a particular e-mail message is spam based on its textual content), and the like.
- image recognition e.g., identifying which photographs from among thousands or millions in a database include a picture of a face or a picture of a particular face
- text classification e.g., determining whether a particular e-mail message is spam based on its textual content
- SVM Support Vector Machine
- Na ⁇ ve Bayes Na ⁇ ve Bayes
- the training samples often are selected from the much larger group of samples to be classified. In some cases, the training samples are randomly selected. In others, the training samples are selected in a systematic manner according to pre-specified criteria. Active learning is one example of the latter approach.
- active-learning methods construct training sets iteratively, starting from a small initial set and then expanding that set incrementally by selecting examples deemed “most interesting” by the classifier at each iteration.
- the “most interesting” samples ordinarily are those that are closest to the decision boundary or where there otherwise is greater uncertainty as to whether the classification predicted by the classifier is correct.
- FIG. 1 is a block diagram illustrating a system according to a representative embodiment of the present invention.
- FIG. 2 is a flow diagram illustrating a machine-learning process according to a first representative embodiment of the present invention.
- FIG. 3 is a block diagram illustrating a system for machine learning according to a representative embodiment of the present invention.
- FIG. 4 illustrates a process for assigning labels according to a representative embodiment of the present invention.
- FIG. 5 illustrates an example of a user interface for presenting a training sample for confirmation/re-labeling according to a first representative embodiment of the present invention.
- FIG. 6 illustrates an example of a user interface for presenting a training sample for confirmation/re-labeling according to a second representative embodiment of the present invention.
- FIG. 7 illustrates an example of a user interface for presenting a training sample for confirmation/re-labeling according to a third representative embodiment of the present invention.
- FIG. 8 is a flow diagram illustrating a machine-learning process according to a second representative embodiment of the present invention.
- FIG. 9 is a block diagram illustrating the selection of samples from among labeled training samples and unlabeled samples, according to a representative embodiment of the present invention.
- FIG. 1 is a block diagram illustrating a system 1 according to a representative embodiment of the present invention.
- unlabeled samples 2 are input into an automated classifier 3 that then outputs a corresponding class prediction 4 for each such sample.
- the samples 2 can comprise, e.g., text, images, video, or signals representing any physical phenomenon (e.g., sound, pressure, radiation, temperature and/or light).
- Such samples 2 typically are represented for purposes of classification by automated classifier 3 as a set of feature values, as discussed in more detail below.
- Classifier 3 applies a predetermined algorithm based on a supervised learning technique (e.g., Support Vector Machines or Na ⁇ ve Bayes) in order to obtain class predictions 4 .
- a training module 5 sets the classification parameters (e.g., weights) of classifier 3 using a set of training samples 7 and class labels 8 that have been assigned to such training samples 7 , typically by a human being. While conventional techniques generally assume that the assigned classification labels 8 are correct, in the preferred embodiments of the present invention such labels 8 are repeatedly questioned and some of them may be submitted for confirmation/re-labeling if they do not appear to conform to the underlying model used by classifier 3 , as discussed in more detail below.
- a supervised learning technique e.g., Support Vector Machines or Na ⁇ ve Bayes
- FIG. 2 is a flow diagram illustrating a machine-learning process according to a representative embodiment of the present invention.
- the entire process illustrated in FIG. 2 is implemented entirely in software, e.g., by reading software code from a computer-readable medium.
- the process instead is implemented in any of the other ways discussed herein.
- the following discussion also references the block diagram of FIG. 3 , which shows one example of a system 40 for implementing the process (again, with the individual components preferably implemented in software).
- Initial training set 45 (shown in FIG. 3 ) is obtained.
- Initial training set 45 could have been generated in the conventional manner. That is, referring to FIG. 3 , various samples 7 are selected and designated for labeling. Ordinarily, the samples 7 are chosen so as to be representative of the types of samples which one desires to classify using the resulting machine-learning classifier (e.g., unlabeled samples 8 ). For example, if one wished to classify a large volume of text-based articles into different subject-matter categories, then the samples 7 preferably would constitute a representative cross-section of such articles. Any of a variety of different (e.g., conventional) techniques can be used for selecting the initial training samples 7 .
- the initial training samples 7 could have been generated in whole or in part by the techniques of the present invention.
- training set 45 subsequently is modified using the techniques of the present invention, as discussed in more detail below.
- the selected samples 7 are labeled via interface module 43 in order to generate the training set 45 , which includes the set of samples and their assigned labels. More preferably, module 43 provides a user interface which allows a user 44 to designate an appropriate label for each presented training sample 7 . In one representative embodiment, module 43 displays to user 44 (here, a human domain expert) each sample 7 together with a set of radio buttons. The user 44 then clicks on one of the radio buttons, thereby designating the appropriate label for the sample 7 .
- user 44 here, a human domain expert
- the training set 45 typically includes some representation of the samples 7 , rather than the actual samples themselves, together with the samples' assigned labels. That representation ordinarily is a set of values for certain features, where the individual features have been chosen for their believed ability to help predict correct labels for the subject sample types, and where the same feature set is used across all of the samples 2 and 7 . Thus, a vector of values corresponding to such features represents a single sample (in this case, one of the samples 7 ). However, for the purpose of the present description, both the samples and their representations are referred to as the “samples”. Any of a variety of different (e.g., conventional) techniques for selecting the feature set can be used.
- step 12 the classifier 3 is trained, e.g., using training module 5 (shown also in FIG. 3 ).
- the training involves attempting to find an optimal (according to some underlying criteria) mapping from the supplied feature set values for the samples 7 to the corresponding classification labels 8 , so that the resulting classifier 3 can receive new unlabeled samples 2 and provide classification labels 4 for them based on its best guess in view of the feature set values for such unlabeled samples 2 .
- Any of a variety of different (e.g., conventional) training techniques can be used by training module 5 .
- portion 49 of system 40 can be implemented as a conventional automated classifier-training system.
- previously labeled samples 7 are selected from training set 45 and submitted for confirmation/re-labeling, e.g., as set forth in the following discussion.
- confirmation/re-labeling refers to the process of submitting an existing training sample for labeling so that its previously assigned label is either confirmed or contradicted, e.g., by a domain expert or other person.
- a request for confirmation/re-labeling can include, e.g.: (i) a mere request to indicate whether the previously assigned classification label is correct; and/or (ii) a request to designate a different label if the previously assigned classification label is incorrect.
- a “reply classification label” is received.
- Such a reply classification label can comprise a mere designation (explicit or implicit) that the previously assigned classification label is correct and/or a different classification label which is believed to be more appropriate than the previously assigned classification label.
- the confirmation/re-labeling process optionally includes providing the user with the previously assigned classification label and/or a label that has been predicted for the subject training sample. Accordingly, if the user believes such label to be correct, he or she can implicitly designate it as such, e.g., by not selecting another label.
- one or more training sample(s) are selected from training set 45 and designated as selected training samples 77 .
- processing block 50 which includes prediction/modeling module 52 and comparison/selection module 53 , with module 53 optionally considering certain other inputs 54 .
- labels are predicted for some or all of the samples in training set 45 using prediction/modeling module 52 .
- module 52 uses the classifier training module 5 (e.g., by applying module 5 ten times in a known-manner 10-fold cross-validation so that each item evaluated is not directly included in the training set of the classifier).
- module 52 uses the production classifier 3 (i.e., which is used to generate predictions 4 for unlabeled input samples 2 ); or (ii) a different classifier is used by prediction/modeling module 52 .
- the labels predicted by module 52 are supplied to comparison/selection module 53 .
- module 53 compares such predicted labels to the labels 8 that were assigned via interface module 43 . Then, based on that comparison, one or more sample(s) 77 are selected from training set 45 .
- the comparison and selection in module 53 can be performed in any of a variety of different ways. Often, a classifier generates a soft classification score and then makes a hard classification decision based on that score. For example, with respect to a binary classification task, a classifier might generate a soft classification score that falls within the range of [0, 1], with 0 representing the classifier's maximum confidence that the sample should not be classified into the subject category and 1 representing the classifier's maximum confidence that it should be classified into the category.
- the classification labels 8 assigned through interface module 43 may be soft (e.g., with a particular value corresponding to how well the subject sample fits within a particular category), in most cases the assigned classification labels 8 will constitute hard classification decisions (e.g., a binary decision of 0 or 1).
- module 53 selects one or more of the samples from training set 45 that have the largest disparity between the label assigned through module 43 and the label predicted by module 52 , designating such selections as selected samples 77 .
- the measure of disparity can comprise, e.g., a simple difference, a ratio, or any other selected function.
- Disparity measures of the foregoing types generally focus on individual training samples. However, disparity also can be evaluated in a broader sense, e.g., by considering a model that has been generated (e.g., in module 52 ) using the current training set 45 and identifying one or more samples from the training set 45 that appear to be inconsistent with such model, in such case designating the identified samples as selected samples 77 . In this embodiment, rather than identifying individual samples that are farthest from their predicted labels, the system 40 identifies, e.g., a cluster of training samples within a region or subspace of the overall feature set that does not appear to conform to the model.
- other considerations also (or instead) are taken into account in designating the selected training samples 77 in step 14 (of FIG. 2 ).
- One example is the effect that changing the assigned classification label 8 (e.g., to the label predicted by module 52 ) would have on classifier 3 . This consideration is discussed in more detail below.
- comparison/selection module 53 also takes into account other inputs 54 (e.g., relating to the unlabeled samples 2 and their currently predicted labels 4 ), which also are discussed in more detail below.
- the sample from training set 45 that optimizes the specified selection criterion (which can include a set of different sub-criteria) preferably is chosen.
- the selection criterion includes one or more sub-criteria biasing toward selection of related samples in training set 45 .
- Such sub-criteria are implemented, in representative embodiments, by clustering the training samples based on their feature sets (e.g., using known clustering techniques). It is noted that the number of samples from training set 45 to be selected or the criterion for determining such number preferably is specified by a user of system 40 .
- step 14 upon the conclusion of step 14 , one or more training sample(s) have been designated as selected training sample 77 .
- the samples 77 are submitted to the user 57 (which can be the same as user 44 , or a different person or automated process) for confirmation/re-labeling.
- a user interface module 55 shown in FIG. 3 preferably is used.
- the way in which the selected training samples 77 are presented to the user 57 by labeling interface module 55 preferably is different in different embodiments of the invention. Some examples are illustrated in FIGS. 5-7 .
- FIG. 5 illustrates an example in which an incoming e-mail message 80 is being labeled as spam 81 or a valid e-mail message 82 (i.e., a binary classification problem).
- the label 82 indicating a valid e-mail message has been pre-selected (e.g., by interface module 55 ), and the user 57 only needs to click on the radio button for spam label 81 if the user 57 disagrees with this pre-selection.
- the pre-selection is based either on the previously assigned classification label or the predicted class for the subject training sample 80 . In either event, the task of assigning classification labels generally will be easier for the user 57 .
- FIG. 6 illustrates a non-binary example in which an incoming e-mail message 90 is being labeled according to the category of subject matter (in this case, particular types of hardware) to which it most closely pertains.
- the choices are printer 91 , PC 92 , display 93 , server 94 , or keyboard/mouse 95 .
- a label in this case, label 93
- the user 57 only needs to click on the appropriate radio button 91 , 92 , 94 or 95 if he or she disagrees with the pre-selection.
- FIG. 7 illustrates a further example in which an incoming e-mail message 100 is being classified in the same manner as described above in connection with FIG. 6 , with labels 101 - 105 corresponding to labels 91 - 95 , respectively.
- labels 101 - 105 corresponding to labels 91 - 95 , respectively.
- the conflicting label 102 also is highlighted.
- label 102 is the previously assigned label and label 103 is the predicted label, or vice versa.
- the user 57 can do nothing if he or she agrees with the pre-selection, can click on the radio button associated with the conflicting label 102 (which has been highlighted and therefore is easy to identify) if that label is more appropriate, or can select a third label (from among labels 101 , 104 and 105 ) if the user 57 disagrees with both the pre-selection 103 and the alternate 102 .
- two or more alternates are highlighted where there are two or more close possibilities.
- the user 57 has the ability to designate more than one label (e.g., where the e-mail message pertains both to a PC 102 and to its display 103 ) and/or more than one label is pre-selected by interface module 55 .
- the selected training samples 77 provided for confirmation/re-labeling are presented in groups of related samples. More preferably, such samples preferably are presented in a logical order, e.g., according to their predicted labels, their previously assigned classification labels, a measure of the disparity between the previously assigned and the predicted labels, and/or the confidence of the classifier in the predicted labels. Depending upon the particular embodiments, the grouping or ordering is done spatially and/or temporally.
- users 44 and 57 generally are described above as being human beings, either or both can comprise automated processes, either in whole or in part.
- user 57 preferably has more expertise, and/or at least is given additional time to assign classification labels, than user 44 , with the desired goal being to obtain higher-quality classifications upon second or subsequent presentations.
- an automated process which is more processor-intensive than the production classifier
- a lower-level worker functions as user 44
- the labels 8 are assigned in whole or in part by an automated process functioning as user 44 that assigns labels based on predefined negative inferences. For example, in a binary classification system an e-mail message that has been labeled as reporting a problem with a user's PC is assigned a negative label with respect to a category of reporting a problem with a server.
- certain embodiments allow the user 44 and/or 57 to designate the degree of difficulty that the subject user had in classifying the subject samples (e.g., clicking on radio buttons for different levels of difficulty, such as easy, somewhat difficult, very difficult or extremely difficult).
- the provided information then can be incorporated into the prediction model and used, e.g., to inform the training module to put less weight on such samples; for this purpose, there exist several known training techniques that consider weighted training samples.
- the user interface for specifying degree of difficulty in assigning a label can be presented, in the various embodiments, for each labeling occurrence (including the initial labeling) or only when a training sample is resubmitted for labeling.
- the confirmation/re-labeling information is received for the submitted training examples and, in step 18 , that information is used to modify the training set 45 and then retrain the classifier 3 , e.g., using training module 5 and the revised training set 45 .
- that information is used to modify the training set 45 and then retrain the classifier 3 , e.g., using training module 5 and the revised training set 45 .
- training module 5 e.g., training module 5
- the revised training set 45 there are two possibilities for each resubmitted training sample 77 .
- the previously assigned classification label 8 for the training sample 77 could have been confirmed by user 57 .
- the training sample 77 preferably is returned to training set 45 with the same classification label 8 .
- the classification label 8 for the training sample 77 is tagged with a higher presumption of accuracy, so that it is less likely to be again selected for confirmation/re-labeling.
- the presumption is absolute so that the subject sample 7 is never again resubmitted.
- the previously assigned classification label 8 for the training sample 7 could have been contradicted by user 57 .
- the processing at this point preferably depends upon the particular embodiment, more preferably, on the presumptions that have been set by the users. Those presumptions, in turn, preferably track the manner in which classification labels are assigned, both by user 44 and by user 57 . For example, if (as shown in FIG.
- the users have set up the labeling process such that labeling by user 57 (i.e., with respect to samples 60 being submitted for confirmation/re-labeling 63 ) is expected to be much more reliable than labeling by user 44 (i.e., with respect to initial labeling 62 ), then the accuracy presumption 66 preferably is much higher than accuracy presumption 65 , with the result that the new label 88 assigned by user 57 replaces the previous label 8 .
- additional processing preferably is performed, e.g., to determine whether the sample really is difficult to classify (even by a human domain expert) or to resolve the discrepancy (e.g., by submitting the sample for an additional label assignment together with a request to assess the difficulty of classifying the sample 77 ).
- the user interface can also receive input indicating which sample/s are inherently difficult to classify, or even an indication that some should be removed from consideration entirely.
- the classifier 3 is retrained using the modified training set 45 (e.g., by training module 5 ), and the retrained classifier 3 is used to re-process at least some of the labeled training samples 45 , thereby obtaining new predictions 4 .
- processing returns to step 14 to select additional samples from training set 45 and repeat the foregoing process.
- samples are selected from both the set of unlabeled samples 2 and the set of training samples 45 , with the particular selections based on a criterion that attempts to maximize overall benefit to the training process 5 .
- FIG. 8 provides a flow diagram illustrating certain exemplary techniques embodying this approach, with additional references to the block diagram shown in FIG. 9 .
- the entire process illustrated in FIG. 8 is implemented in software, e.g., by reading and executing software code from a computer-readable medium.
- the same system 40 shown in FIG. 3 and discussed above
- the other inputs 54 preferably include information regarding the unlabeled samples 2 and their predicted labels 4 , together with information regarding particular unlabeled samples 2 that are potential candidates for labeling via interface module 55 .
- step 131 the current training set 45 is obtained, and in step 132 a prediction set 171 is obtained, the latter including the unlabeled samples 2 and the currently predicted labels 4 for such samples 2 .
- samples are selected 175 from among both training set 45 and prediction set 171 , e.g., by evaluating samples in both such sets and selecting one or more based on a specified selection criterion. As illustrated in FIG. 9 , this step is accomplished in the present embodiment by using a criterion that attempts to maximize the expected benefit from multiple different considerations, including but not limited to evaluation of: consistency 177 (e.g., questioning training label outliers), prediction confidence 178 (e.g., selection of samples with uncertain or unstable predictions), and variety 179 among the training samples.
- the prediction-confidence consideration 178 can be replaced with any known active-learning technique that evaluates the importance of labeling individual cases or any other technique that attempts to identify samples for labeling.
- Consistency 177 has been discussed above and generally is directed at ensuring that individual training samples or clusters of training samples are consistent with the prediction model that is being used by the production classifier 3 .
- this objective includes metrics designed to identify any samples in the training set 45 that can be considered to be outliers, i.e., samples whose assigned classification labels 8 are significantly different than what the classifier 3 (or prediction/modeling module 52 would predict for them).
- Conventional methods exist for identifying outliers (sometimes referred to in the literature as “noise labels”), typically for the purpose of ignoring them when training a classifier.
- the considerations pertaining to prediction confidence 178 and variety 179 pertain to selection of samples from the prediction set 171 , and information pertaining to such considerations generally is included within the information designated as other inputs 54 in system 40 (shown in FIG. 3 ).
- the objective pertaining to prediction confidence 178 biases toward selecting samples for which the classifier 3 is most uncertain regarding its predicted classification label 4 (such as boundary samples). For example, in the binary classification example discussed above, the bias is toward samples whose soft classification label is close to 0.5.
- Boundary samples can be selected, e.g., through conventional active-learning techniques. For example, boundary samples may be identified by selecting those examples having soft classification scores that are close to the decision threshold for specific hard classifications. Alternatively, boundary samples can be identified using a query-by-committee approach in which differently trained classifiers arrive at significantly different predictions (hard or soft) for the same sample. Conventionally, selection of boundary samples has been limited to previously unlabeled samples 2 . However, in the present embodiments it is applied to samples in both the training set 45 (disregarding the assigned labels) and the prediction set 171 .
- prediction-confidence objective 178 is limited to selecting samples only from the prediction set 171 (i.e., unlabeled samples 2 ).
- the prediction-confidence objective 178 is applied both to previously labeled training samples in set 45 and to unlabeled samples 2 .
- the previously assigned labels for such samples in set 45 preferably are ignored.
- the training samples in set 45 preferably are treated as unlabeled samples.
- the variety objective 179 attempts to ensure that the training samples are as representative as possible of the entire diversity of samples that have been, or are likely to be, input for classification. Variety preferably is achieved by determining how well the various unlabeled samples 2 are represented by the training samples 7 . In one embodiment of the invention, samples 2 are divided into clusters (e.g., using conventional clustering techniques), and each cluster is evaluated to determine whether it is proportionally represented by samples in training set 45 . Selection is then biased toward samples in under-represented clusters, e.g., by forcing selection from such under-represented clusters until balance is restored.
- a distance measure is constructed as a function of distance from the unlabeled samples 2 to the samples in training set 45 (e.g., using the sum of the distances from each unlabeled sample 2 to the nearest sample in training set 45 ), and then biasing selection toward samples having larger distance measures in their immediate vicinity (e.g., multiplying by a factor that is based on the distance measure).
- a measure of the effect of requesting that particular sample be labeled 180 (either for the first time or for confirmation) preferably is taken into account. For example, in one embodiment a fixed number of the closest boundary samples, the farthest outliers, and to the best candidates to achieve improved variety (e.g., 1, 2 or 3 of each) are identified and then the expected effect of requesting labeling 180 for each is evaluated.
- the measure of the effect can be determined by assuming the classification of each sample would be reversed (or, for non-binary classification problems, changed to the next closest category) if submitted for labeling 180 and then determining how big of an impact such a change would have on the classifier 3 (e.g., accuracy, F-measure, precision, recall, quantification, etc.).
- a cumulative score reflecting all three considerations is assigned to each sample, and the samples having the highest cumulative scores are evaluated for expected effect, e.g., in a similar manner. For samples that have not yet been labeled, no consistency measure is available; in such cases, for example, a default value of zero can be assumed when computing a weighted average of all three considerations.
- the cumulative score is adjusted to account for the effect measure, and the sample(s) having the highest adjusted scores are selected.
- the selections are constrained so that the samples are of the same type (e.g., clustered within the same region of the feature space), thereby resulting in a group of similar samples to be provided for labeling 180 .
- Such a constraint while potentially producing a less optimal sample set than otherwise would be obtained, often has the offsetting advantage that the user will find it easier to label a group of similar samples.
- certain embodiments incorporate a bias toward selecting samples from training set 45 whose assigned labels have a lower presumption of accuracy.
- a presumption might apply, e.g., in situations where the labels were assigned: (i) hastily by a human being, (ii) by a source of training labels that subsequently has been found to be less competent, or (iii) as a result of a negative inference, as discussed above.
- Such a bias is applied, e.g., by multiplying the cumulative adjusted score of each sample in training set 45 by a factor that is based on the presumed accuracy of its corresponding assigned classification label 8 .
- the change effect to classifier 3 is modified by the degree of expectation that the label in fact would be reversed.
- Such an expectation can be provided by a soft classification score output by the classifier 3 itself.
- the classifier 3 preferably is roughly calibrated to provide soft classification scores that correspond to such expectations, e.g., using calibration techniques known in the art, such as calibration based on a cross-validation calculation performed using later-identified actual classification labels.
- step 137 the samples selected in step 134 are presented to the user for labeling 180 .
- the samples are presented in an ordered manner, e.g., sorted by a determined degree of likelihood that the samples are a member of a particular class.
- step 138 the training set 45 is modified based on the labels received in step 137 , the classifier 3 is retrained based on the modified training set 45 , and at least some of the samples 2 and 7 are reprocessed using classifier 3 .
- This step is similar to step 18 in FIG. 2 , and the same considerations apply here.
- processing returns to step 134 in order to select 175 the next sample or set of samples and to repeat the foregoing process.
- 5-20 samples are selected 175 for labeling, sorted by their prediction strength (e.g. probability of belonging to the positive class according to the current classifier), and presented to the user 57 in a single screen. If the classifier 3 is reasonably accurate, the positive samples will be mostly gathered together, making it easier for the user 57 to group-select them and label them positive (same for the negative samples), with a few individual clicks to treat the exceptions.
- prediction strength e.g. probability of belonging to the positive class according to the current classifier
- Such devices typically will include, for example, at least some of the following components interconnected with each other, e.g., via a common bus: one or more central processing units (CPUs); read-only memory (ROM); random access memory (RAM); input/output software and circuitry for interfacing with other devices (e.g., using a hardwired connection, such as a serial port, a parallel port, a USB connection or a firewire connection, or using a wireless protocol, such as Bluetooth or a 802.11 protocol); software and circuitry for connecting to one or more networks (e.g., using a hardwired connection such as an Ethernet card or a wireless protocol, such as code division multiple access (CDMA), global system for mobile communications (GSM), Bluetooth, a 802.11 protocol, or any other cellular-based or non-cellular-based system), which networks, in turn, in many embodiment
- CDMA code division multiple access
- GSM global system for mobile communications
- Bluetooth Bluetooth
- 802.11 protocol any other cellular-based or non-cellular-based system
- the process steps to implement the above methods and functionality typically initially are stored in mass storage (e.g., the hard disk), are downloaded into RAM and then are executed by the CPU out of RAM.
- mass storage e.g., the hard disk
- the process steps initially are stored in RAM or ROM.
- Suitable devices for use in implementing the present invention may be obtained from various vendors. In the various embodiments, different types of devices are used depending upon the size and complexity of the tasks. Suitable devices include mainframe computers, multiprocessor computers, workstations, personal computers, and even smaller computers such as PDAs, wireless telephones or any other appliance or device, whether stand-alone, hard-wired into a network or wirelessly connected to a network.
- Suitable devices include mainframe computers, multiprocessor computers, workstations, personal computers, and even smaller computers such as PDAs, wireless telephones or any other appliance or device, whether stand-alone, hard-wired into a network or wirelessly connected to a network.
- any of the functionality described above can be implemented in software, hardware, firmware or any combination of these, with the particular implementation being selected based on known engineering tradeoffs. More specifically, where the functionality described above is implemented in a fixed, predetermined or logical manner, it can be accomplished through programming (e.g., software or firmware), an appropriate arrangement of logic components (hardware) or any combination of the two, as will be readily appreciated by those skilled in the art.
- the present invention also relates to machine-readable media on which are stored program instructions for performing the methods and functionality of this invention.
- Such media include, by way of example, magnetic disks, magnetic tape, optically readable media such as CD ROMs and DVD ROMs, or semiconductor memory such as PCMCIA cards, various types of memory cards, USB memory devices, etc.
- the medium may take the form of a portable item such as a miniature disk drive or a small disk, diskette, cassette, cartridge, card, stick etc., or it may take the form of a relatively larger or immobile item such as a hard disk drive, ROM or RAM provided in a computer or other device.
- the foregoing description refers to clicking or double-clicking on user-interface buttons, dragging user-interface items, or otherwise entering commands or information via a particular user-interface mechanism and/or in a particular manner. All of such references are intended to be exemplary only, it being understood that the present invention encompasses entry of the corresponding commands or information by a user in any other manner using the same or any other user-interface mechanism. In addition, or instead, such commands or information may be input by an automated (e.g., computer-executed) process.
- functionality sometimes is ascribed to a particular module or component. However, functionality generally may be redistributed as desired among any different modules or components, in some cases completely obviating the need for a particular component or module and/or requiring the addition of new components or modules.
- the precise distribution of functionality preferably is made according to known engineering tradeoffs, with reference to the specific embodiment of the invention, as will be understood by those skilled in the art.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
Claims (18)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/555,036 US7792353B2 (en) | 2006-10-31 | 2006-10-31 | Retraining a machine-learning classifier using re-labeled training samples |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/555,036 US7792353B2 (en) | 2006-10-31 | 2006-10-31 | Retraining a machine-learning classifier using re-labeled training samples |
Publications (2)
Publication Number | Publication Date |
---|---|
US20080103996A1 US20080103996A1 (en) | 2008-05-01 |
US7792353B2 true US7792353B2 (en) | 2010-09-07 |
Family
ID=39331533
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/555,036 Expired - Fee Related US7792353B2 (en) | 2006-10-31 | 2006-10-31 | Retraining a machine-learning classifier using re-labeled training samples |
Country Status (1)
Country | Link |
---|---|
US (1) | US7792353B2 (en) |
Cited By (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090092299A1 (en) * | 2007-10-03 | 2009-04-09 | Siemens Medical Solutions Usa, Inc. | System and Method for Joint Classification Using Feature Space Cluster Labels |
US20090116737A1 (en) * | 2007-10-30 | 2009-05-07 | Siemens Corporate Research, Inc. | Machine Learning For Tissue Labeling Segmentation |
US20090202144A1 (en) * | 2008-02-13 | 2009-08-13 | Museami, Inc. | Music score deconstruction |
US20100082627A1 (en) * | 2008-09-24 | 2010-04-01 | Yahoo! Inc. | Optimization filters for user generated content searches |
US20100312727A1 (en) * | 2008-12-19 | 2010-12-09 | Pottenger William M | Systems and methods for data transformation using higher order learning |
US20110007964A1 (en) * | 2009-07-10 | 2011-01-13 | Palo Alto Research Center Incorporated | System and method for machine-assisted human labeling of pixels in an image |
US20110007366A1 (en) * | 2009-07-10 | 2011-01-13 | Palo Alto Research Center Incorporated | System and method for classifying connected groups of foreground pixels in scanned document images according to the type of marking |
US20120246099A1 (en) * | 2011-03-23 | 2012-09-27 | Kabushiki Kaisha Toshiba | Learning device, learning method, and computer program product |
US20140040169A1 (en) * | 2012-07-31 | 2014-02-06 | George Forman | Active learning with per-case symmetrical importance scores |
US8649600B2 (en) | 2009-07-10 | 2014-02-11 | Palo Alto Research Center Incorporated | System and method for segmenting text lines in documents |
US8655724B2 (en) * | 2006-12-18 | 2014-02-18 | Yahoo! Inc. | Evaluating performance of click fraud detection systems |
US20140279742A1 (en) * | 2013-03-15 | 2014-09-18 | Hewlett-Packard Development Company, L.P. | Determining an obverse weight |
US20150100527A1 (en) * | 2010-07-16 | 2015-04-09 | First Wave Technology Pty Ltd | Methods and Systems for Analysis and/or Classification of Information |
EP2869237A1 (en) * | 2013-10-31 | 2015-05-06 | Kabushiki Kaisha Toshiba | Image display apparatus, image display method, and computer program product |
US20170316548A1 (en) * | 2015-09-06 | 2017-11-02 | Boe Technology Group Co., Ltd. | Super-Resolution Method and System, Server, User Device and Method Therefor |
WO2017201540A1 (en) * | 2016-05-20 | 2017-11-23 | Techcyte, Inc. | Machine learning classification of particles or substances in digital microscopy images |
US10078688B2 (en) | 2016-04-12 | 2018-09-18 | Abbyy Production Llc | Evaluating text classifier parameters based on semantic features |
CN109710788A (en) * | 2018-12-28 | 2019-05-03 | 斑马网络技术有限公司 | Image pattern mark and management method and equipment |
US10296848B1 (en) * | 2018-03-05 | 2019-05-21 | Clinc, Inc. | Systems and method for automatically configuring machine learning models |
US10303978B1 (en) | 2018-03-26 | 2019-05-28 | Clinc, Inc. | Systems and methods for intelligently curating machine learning training data and improving machine learning model performance |
WO2019199072A1 (en) * | 2018-04-11 | 2019-10-17 | Samsung Electronics Co., Ltd. | System and method for active machine learning |
US10572801B2 (en) | 2017-11-22 | 2020-02-25 | Clinc, Inc. | System and method for implementing an artificially intelligent virtual assistant using machine learning |
US10679150B1 (en) | 2018-12-13 | 2020-06-09 | Clinc, Inc. | Systems and methods for automatically configuring training data for training machine learning models of a machine learning-based dialogue system including seeding training samples or curating a corpus of training data based on instances of training data identified as anomalous |
US10754863B2 (en) | 2017-04-05 | 2020-08-25 | Yandex Europe Ag | Method and system for ranking a plurality of documents on a search engine results page |
US10789533B2 (en) | 2017-07-26 | 2020-09-29 | LogMeln, Inc. | Generating a consistently labeled training dataset by automatically generating and displaying a set of most similar previously-labeled texts and their previously assigned labels for each text that is being labeled for the training dataset |
US10878144B2 (en) | 2017-08-10 | 2020-12-29 | Allstate Insurance Company | Multi-platform model processing and execution management engine |
US20210110299A1 (en) * | 2019-10-15 | 2021-04-15 | Kinaxis Inc. | Interactive machine learning |
US20210109969A1 (en) | 2019-10-11 | 2021-04-15 | Kinaxis Inc. | Machine learning segmentation methods and systems |
US11010656B2 (en) | 2017-10-30 | 2021-05-18 | Clinc, Inc. | System and method for implementing an artificially intelligent virtual assistant using machine learning |
US11023824B2 (en) * | 2017-08-30 | 2021-06-01 | Intel Corporation | Constrained sample selection for training models |
US11030492B2 (en) * | 2019-01-16 | 2021-06-08 | Clarifai, Inc. | Systems, techniques, and interfaces for obtaining and annotating training instances |
US11048979B1 (en) * | 2018-11-23 | 2021-06-29 | Amazon Technologies, Inc. | Active learning loop-based data labeling service |
EP3905144A1 (en) * | 2020-04-30 | 2021-11-03 | UiPath, Inc. | Machine learning model retraining pipeline for robotic process automation |
US20210406472A1 (en) * | 2020-06-30 | 2021-12-30 | Hitachi, Ltd. | Named-entity classification apparatus and named-entity classification method |
US11270438B2 (en) | 2020-06-12 | 2022-03-08 | Samasource Impact Sourcing, Inc. | System and method for triggering machine learning (ML) annotation model retraining |
US11308117B2 (en) | 2017-12-07 | 2022-04-19 | Palantir Technologies Inc. | Relationship analysis and mapping for interrelated multi-layered datasets |
WO2022115199A1 (en) * | 2020-11-25 | 2022-06-02 | UiPath, Inc. | Robotic process automation architectures and processes for hosting, monitoring, and retraining machine learning models |
US11372900B1 (en) * | 2018-04-06 | 2022-06-28 | Corelogic Solutions, Llc | Artificial intelligence-based property data matching system |
US11416518B2 (en) * | 2018-10-30 | 2022-08-16 | Optum, Inc. | Machine learning for machine-assisted data classification |
WO2022221488A3 (en) * | 2021-04-14 | 2022-12-08 | Capital One Services, Llc | Data labeling for training artificial intelligence systems |
US11556810B2 (en) | 2019-07-11 | 2023-01-17 | International Business Machines Corporation | Estimating feasibility and effort for a machine learning solution |
US20230161971A1 (en) * | 2019-03-05 | 2023-05-25 | Kensho Technologies, Llc | Dynamically Updated Text Classifier |
US11715167B2 (en) | 2016-11-11 | 2023-08-01 | Palantir Technologies Inc. | Graphical representation of a complex task |
US11755949B2 (en) | 2017-08-10 | 2023-09-12 | Allstate Insurance Company | Multi-platform machine learning systems |
US11875367B2 (en) | 2019-10-11 | 2024-01-16 | Kinaxis Inc. | Systems and methods for dynamic demand sensing |
US12154013B2 (en) | 2019-10-15 | 2024-11-26 | Kinaxis Inc. | Interactive machine learning |
US12271920B2 (en) | 2019-10-11 | 2025-04-08 | Kinaxis Inc. | Systems and methods for features engineering |
US12321823B2 (en) | 2020-04-30 | 2025-06-03 | UiPath, Inc. | Machine learning model retraining pipeline for robotic process automation |
US12339926B1 (en) | 2020-07-01 | 2025-06-24 | Massachusetts Mutual Life Insurance Company | Systems and methods for risk factor predictive modeling with dynamic training |
US12346921B2 (en) | 2019-10-11 | 2025-07-01 | Kinaxis Inc. | Systems and methods for dynamic demand sensing and forecast adjustment |
Families Citing this family (165)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7801358B2 (en) * | 2006-11-03 | 2010-09-21 | Google Inc. | Methods and systems for analyzing data in media material having layout |
US8880529B2 (en) | 2007-05-15 | 2014-11-04 | Tivo Inc. | Hierarchical tags with community-based ratings |
WO2008144454A1 (en) * | 2007-05-15 | 2008-11-27 | Tivo Inc. | Media data content search system |
US9348499B2 (en) | 2008-09-15 | 2016-05-24 | Palantir Technologies, Inc. | Sharing objects that rely on local resources with outside servers |
US10515339B1 (en) * | 2008-09-25 | 2019-12-24 | Intuit Inc. | Error correction system for accountants |
US8219511B2 (en) * | 2009-02-24 | 2012-07-10 | Microsoft Corporation | Unbiased active learning |
US8725668B2 (en) | 2009-03-24 | 2014-05-13 | Regents Of The University Of Minnesota | Classifying an item to one of a plurality of groups |
US8935258B2 (en) * | 2009-06-15 | 2015-01-13 | Microsoft Corporation | Identification of sample data items for re-judging |
US8401979B2 (en) * | 2009-11-16 | 2013-03-19 | Microsoft Corporation | Multiple category learning for training classifiers |
JP5720579B2 (en) * | 2009-12-24 | 2015-05-20 | 日本電気株式会社 | Metric learning device, metric learning method, and program |
JP5898221B2 (en) * | 2010-10-19 | 2016-04-06 | スリーエム イノベイティブ プロパティズ カンパニー | Computer-aided assignment of evaluations for digital samples of manufactured web products |
US10043129B2 (en) | 2010-12-06 | 2018-08-07 | Regents Of The University Of Minnesota | Functional assessment of a network |
US8732574B2 (en) | 2011-08-25 | 2014-05-20 | Palantir Technologies, Inc. | System and method for parameterizing documents for automatic workflow generation |
US9967218B2 (en) * | 2011-10-26 | 2018-05-08 | Oath Inc. | Online active learning in user-generated content streams |
US9348677B2 (en) | 2012-10-22 | 2016-05-24 | Palantir Technologies Inc. | System and method for batch evaluation programs |
US10140664B2 (en) | 2013-03-14 | 2018-11-27 | Palantir Technologies Inc. | Resolving similar entities from a transaction database |
US8909656B2 (en) | 2013-03-15 | 2014-12-09 | Palantir Technologies Inc. | Filter chains with associated multipath views for exploring large data sets |
US8868486B2 (en) | 2013-03-15 | 2014-10-21 | Palantir Technologies Inc. | Time-sensitive cube |
US8799799B1 (en) | 2013-05-07 | 2014-08-05 | Palantir Technologies Inc. | Interactive geospatial map |
JP6465876B2 (en) | 2013-06-28 | 2019-02-06 | ディー−ウェイブ システムズ インコーポレイテッド | System and method for quantum processing of data |
US8938686B1 (en) | 2013-10-03 | 2015-01-20 | Palantir Technologies Inc. | Systems and methods for analyzing performance of an entity |
US10409926B2 (en) | 2013-11-27 | 2019-09-10 | Falkonry Inc. | Learning expected operational behavior of machines from generic definitions and past behavior |
US9105000B1 (en) | 2013-12-10 | 2015-08-11 | Palantir Technologies Inc. | Aggregating data from a plurality of data sources |
US9734217B2 (en) | 2013-12-16 | 2017-08-15 | Palantir Technologies Inc. | Methods and systems for analyzing entity performance |
US10579647B1 (en) | 2013-12-16 | 2020-03-03 | Palantir Technologies Inc. | Methods and systems for analyzing entity performance |
US10037128B2 (en) | 2014-02-04 | 2018-07-31 | Falkonry, Inc. | Operating behavior classification interface |
US10043112B2 (en) * | 2014-03-07 | 2018-08-07 | Qualcomm Incorporated | Photo management |
US8935201B1 (en) | 2014-03-18 | 2015-01-13 | Palantir Technologies Inc. | Determining and extracting changed data from a data source |
US9836580B2 (en) | 2014-03-21 | 2017-12-05 | Palantir Technologies Inc. | Provider portal |
US20160026923A1 (en) | 2014-07-22 | 2016-01-28 | Palantir Technologies Inc. | System and method for determining a propensity of entity to take a specified action |
US20160063394A1 (en) * | 2014-08-27 | 2016-03-03 | Microsoft Corporation | Computing Device Classifier Improvement Through N-Dimensional Stratified Input Sampling |
US9390086B2 (en) * | 2014-09-11 | 2016-07-12 | Palantir Technologies Inc. | Classification system with methodology for efficient verification |
US9229952B1 (en) | 2014-11-05 | 2016-01-05 | Palantir Technologies, Inc. | History preserving data pipeline system and method |
JP6337973B2 (en) * | 2014-11-27 | 2018-06-06 | 日本電気株式会社 | Additional learning device, additional learning method, and additional learning program |
US9483546B2 (en) | 2014-12-15 | 2016-11-01 | Palantir Technologies Inc. | System and method for associating related records to common entities across multiple lists |
US9335911B1 (en) | 2014-12-29 | 2016-05-10 | Palantir Technologies Inc. | Interactive user interface for dynamic data analysis exploration and query processing |
US11302426B1 (en) | 2015-01-02 | 2022-04-12 | Palantir Technologies Inc. | Unified data interface and system |
US9727560B2 (en) | 2015-02-25 | 2017-08-08 | Palantir Technologies Inc. | Systems and methods for organizing and identifying documents via hierarchies and dimensions of tags |
US9891808B2 (en) | 2015-03-16 | 2018-02-13 | Palantir Technologies Inc. | Interactive user interfaces for location-based data analysis |
US10628834B1 (en) | 2015-06-16 | 2020-04-21 | Palantir Technologies Inc. | Fraud lead detection system for efficiently processing database-stored data and automatically generating natural language explanatory information of system results for display in interactive user interfaces |
US10943181B2 (en) | 2015-06-26 | 2021-03-09 | Microsoft Technology Licensing, Llc | Just in time classifier training |
US9418337B1 (en) | 2015-07-21 | 2016-08-16 | Palantir Technologies Inc. | Systems and models for data analytics |
US9392008B1 (en) | 2015-07-23 | 2016-07-12 | Palantir Technologies Inc. | Systems and methods for identifying information related to payment card breaches |
US10121108B2 (en) | 2015-07-31 | 2018-11-06 | Bluvector, Inc. | System and method for in-situ classifier retraining for malware identification and model heterogeneity |
US9996595B2 (en) | 2015-08-03 | 2018-06-12 | Palantir Technologies, Inc. | Providing full data provenance visualization for versioned datasets |
WO2017030535A1 (en) * | 2015-08-14 | 2017-02-23 | Hewlett-Packard Development Company, L. P. | Dataset partitioning |
US9600146B2 (en) | 2015-08-17 | 2017-03-21 | Palantir Technologies Inc. | Interactive geospatial map |
US9671776B1 (en) | 2015-08-20 | 2017-06-06 | Palantir Technologies Inc. | Quantifying, tracking, and anticipating risk at a manufacturing facility, taking deviation type and staffing conditions into account |
US9485265B1 (en) | 2015-08-28 | 2016-11-01 | Palantir Technologies Inc. | Malicious activity detection system capable of efficiently processing data accessed from databases and generating alerts for display in interactive user interfaces |
US10706434B1 (en) | 2015-09-01 | 2020-07-07 | Palantir Technologies Inc. | Methods and systems for determining location information |
US20170068712A1 (en) | 2015-09-04 | 2017-03-09 | Palantir Technologies Inc. | Systems and methods for database investigation tool |
US9576015B1 (en) | 2015-09-09 | 2017-02-21 | Palantir Technologies, Inc. | Domain-specific language for dataset transformations |
GB201517462D0 (en) * | 2015-10-02 | 2015-11-18 | Tractable Ltd | Semi-automatic labelling of datasets |
JP6873120B2 (en) | 2015-10-27 | 2021-05-19 | ディー−ウェイブ システムズ インコーポレイテッド | Systems and methods for reducing degeneracy in quantum processors |
US10223429B2 (en) | 2015-12-01 | 2019-03-05 | Palantir Technologies Inc. | Entity data attribution using disparate data sets |
US10706056B1 (en) | 2015-12-02 | 2020-07-07 | Palantir Technologies Inc. | Audit log report generator |
US10114884B1 (en) | 2015-12-16 | 2018-10-30 | Palantir Technologies Inc. | Systems and methods for attribute analysis of one or more databases |
US10373099B1 (en) | 2015-12-18 | 2019-08-06 | Palantir Technologies Inc. | Misalignment detection system for efficiently processing database-stored data and automatically generating misalignment information for display in interactive user interfaces |
US10871878B1 (en) | 2015-12-29 | 2020-12-22 | Palantir Technologies Inc. | System log analysis and object user interaction correlation system |
US9792020B1 (en) | 2015-12-30 | 2017-10-17 | Palantir Technologies Inc. | Systems for collecting, aggregating, and storing data, generating interactive user interfaces for analyzing data, and generating alerts based upon collected data |
EP3398123B1 (en) * | 2015-12-31 | 2025-03-12 | KLA - Tencor Corporation | Accelerated training of a machine learning based model for semiconductor applications |
US11580375B2 (en) | 2015-12-31 | 2023-02-14 | Kla-Tencor Corp. | Accelerated training of a machine learning based model for semiconductor applications |
US10817796B2 (en) | 2016-03-07 | 2020-10-27 | D-Wave Systems Inc. | Systems and methods for machine learning |
US10698938B2 (en) | 2016-03-18 | 2020-06-30 | Palantir Technologies Inc. | Systems and methods for organizing and identifying documents via hierarchies and dimensions of tags |
US10068199B1 (en) | 2016-05-13 | 2018-09-04 | Palantir Technologies Inc. | System to catalogue tracking data |
US10007674B2 (en) | 2016-06-13 | 2018-06-26 | Palantir Technologies Inc. | Data revision control in large-scale data analytic systems |
US10545975B1 (en) | 2016-06-22 | 2020-01-28 | Palantir Technologies Inc. | Visual analysis of data using sequenced dataset reduction |
US10909130B1 (en) | 2016-07-01 | 2021-02-02 | Palantir Technologies Inc. | Graphical user interface for a database system |
EP3516599A4 (en) | 2016-09-26 | 2019-10-02 | D-Wave Systems Inc. | SYSTEMS, METHODS AND APPARATUS FOR SAMPLING FROM A SAMPLING SERVER |
US10552002B1 (en) | 2016-09-27 | 2020-02-04 | Palantir Technologies Inc. | User interface based variable machine modeling |
US9842338B1 (en) | 2016-11-21 | 2017-12-12 | Palantir Technologies Inc. | System to identify vulnerable card readers |
WO2018094496A1 (en) * | 2016-11-23 | 2018-05-31 | Primal Fusion Inc. | System and method for using a knowledge representation with a machine learning classifier |
US11544579B2 (en) | 2016-11-23 | 2023-01-03 | Primal Fusion Inc. | System and method for generating training data for machine learning classifier |
US11531852B2 (en) | 2016-11-28 | 2022-12-20 | D-Wave Systems Inc. | Machine learning systems and methods for training with noisy labels |
US11250425B1 (en) | 2016-11-30 | 2022-02-15 | Palantir Technologies Inc. | Generating a statistic using electronic transaction data |
GB201621434D0 (en) | 2016-12-16 | 2017-02-01 | Palantir Technologies Inc | Processing sensor logs |
US9886525B1 (en) | 2016-12-16 | 2018-02-06 | Palantir Technologies Inc. | Data item aggregate probability analysis system |
US10249033B1 (en) | 2016-12-20 | 2019-04-02 | Palantir Technologies Inc. | User interface for managing defects |
US10728262B1 (en) | 2016-12-21 | 2020-07-28 | Palantir Technologies Inc. | Context-aware network-based malicious activity warning systems |
US10360238B1 (en) | 2016-12-22 | 2019-07-23 | Palantir Technologies Inc. | Database systems and user interfaces for interactive data association, analysis, and presentation |
US11373752B2 (en) | 2016-12-22 | 2022-06-28 | Palantir Technologies Inc. | Detection of misuse of a benefit system |
EP3340106B1 (en) * | 2016-12-23 | 2023-02-08 | Hexagon Technology Center GmbH | Method and system for assigning particular classes of interest within measurement data |
US10721262B2 (en) | 2016-12-28 | 2020-07-21 | Palantir Technologies Inc. | Resource-centric network cyber attack warning system |
US10762471B1 (en) | 2017-01-09 | 2020-09-01 | Palantir Technologies Inc. | Automating management of integrated workflows based on disparate subsidiary data sources |
US10133621B1 (en) | 2017-01-18 | 2018-11-20 | Palantir Technologies Inc. | Data analysis system to facilitate investigative process |
US10509844B1 (en) | 2017-01-19 | 2019-12-17 | Palantir Technologies Inc. | Network graph parser |
US10515109B2 (en) | 2017-02-15 | 2019-12-24 | Palantir Technologies Inc. | Real-time auditing of industrial equipment condition |
US11138514B2 (en) * | 2017-03-23 | 2021-10-05 | Futurewei Technologies, Inc. | Review machine learning system |
US10581954B2 (en) | 2017-03-29 | 2020-03-03 | Palantir Technologies Inc. | Metric collection and aggregation for distributed software services |
US10866936B1 (en) | 2017-03-29 | 2020-12-15 | Palantir Technologies Inc. | Model object management and storage system |
US20180285775A1 (en) * | 2017-04-03 | 2018-10-04 | Salesforce.Com, Inc. | Systems and methods for machine learning classifiers for support-based group |
US10133783B2 (en) | 2017-04-11 | 2018-11-20 | Palantir Technologies Inc. | Systems and methods for constraint driven database searching |
CN107093210B (en) * | 2017-04-20 | 2021-07-16 | 北京图森智途科技有限公司 | A kind of laser point cloud labeling method and device |
US10482000B2 (en) * | 2017-04-24 | 2019-11-19 | Microsoft Technology Licensing, Llc | Machine learned decision guidance for alerts originating from monitoring systems |
US11880746B1 (en) * | 2017-04-26 | 2024-01-23 | Hrb Innovations, Inc. | Interface for artificial intelligence training |
US10563990B1 (en) | 2017-05-09 | 2020-02-18 | Palantir Technologies Inc. | Event-based route planning |
US10606872B1 (en) | 2017-05-22 | 2020-03-31 | Palantir Technologies Inc. | Graphical user interface for a database system |
US10795749B1 (en) | 2017-05-31 | 2020-10-06 | Palantir Technologies Inc. | Systems and methods for providing fault analysis user interface |
US10956406B2 (en) | 2017-06-12 | 2021-03-23 | Palantir Technologies Inc. | Propagated deletion of database records and derived data |
JP6909078B2 (en) * | 2017-07-07 | 2021-07-28 | 株式会社エヌ・ティ・ティ・データ | Disease onset prediction device, disease onset prediction method and program |
US11216762B1 (en) | 2017-07-13 | 2022-01-04 | Palantir Technologies Inc. | Automated risk visualization using customer-centric data analysis |
US10430444B1 (en) | 2017-07-24 | 2019-10-01 | Palantir Technologies Inc. | Interactive geospatial map and geospatial visualization systems |
US10628475B2 (en) * | 2017-10-03 | 2020-04-21 | International Business Machines Corporation | Runtime control of automation accuracy using adjustable thresholds |
CN107798390B (en) * | 2017-11-22 | 2023-03-21 | 创新先进技术有限公司 | Training method and device of machine learning model and electronic equipment |
US11314721B1 (en) | 2017-12-07 | 2022-04-26 | Palantir Technologies Inc. | User-interactive defect analysis for root cause |
US10877984B1 (en) | 2017-12-07 | 2020-12-29 | Palantir Technologies Inc. | Systems and methods for filtering and visualizing large scale datasets |
CN108009589A (en) * | 2017-12-12 | 2018-05-08 | 腾讯科技(深圳)有限公司 | Sample data processing method, device and computer-readable recording medium |
WO2019118644A1 (en) | 2017-12-14 | 2019-06-20 | D-Wave Systems Inc. | Systems and methods for collaborative filtering with variational autoencoders |
IL256480B (en) * | 2017-12-21 | 2021-05-31 | Agent Video Intelligence Ltd | System and method for use in training machine learning utilities |
US11263382B1 (en) | 2017-12-22 | 2022-03-01 | Palantir Technologies Inc. | Data normalization and irregularity detection system |
US10877654B1 (en) | 2018-04-03 | 2020-12-29 | Palantir Technologies Inc. | Graphical user interfaces for optimizations |
US10754822B1 (en) | 2018-04-18 | 2020-08-25 | Palantir Technologies Inc. | Systems and methods for ontology migration |
US10885021B1 (en) | 2018-05-02 | 2021-01-05 | Palantir Technologies Inc. | Interactive interpreter and graphical user interface |
US10754946B1 (en) | 2018-05-08 | 2020-08-25 | Palantir Technologies Inc. | Systems and methods for implementing a machine learning approach to modeling entity behavior |
CN110163376B (en) * | 2018-06-04 | 2023-11-03 | 腾讯科技(深圳)有限公司 | Sample detection method, media object identification method, device, terminal and medium |
US11119630B1 (en) | 2018-06-19 | 2021-09-14 | Palantir Technologies Inc. | Artificial intelligence assisted evaluations and user interface for same |
JP7139723B2 (en) * | 2018-06-27 | 2022-09-21 | 富士通株式会社 | Selection program, selection method and selection device |
CN110659657B (en) * | 2018-06-29 | 2024-05-24 | 北京京东尚科信息技术有限公司 | Method and device for training model |
US11386346B2 (en) | 2018-07-10 | 2022-07-12 | D-Wave Systems Inc. | Systems and methods for quantum bayesian networks |
US11775815B2 (en) | 2018-08-10 | 2023-10-03 | Samsung Electronics Co., Ltd. | System and method for deep memory network |
JP6981553B2 (en) * | 2018-09-04 | 2021-12-15 | 日本電気株式会社 | Identification system, model provision method and model provision program |
US11126638B1 (en) | 2018-09-13 | 2021-09-21 | Palantir Technologies Inc. | Data visualization and parsing system |
WO2020065627A1 (en) * | 2018-09-28 | 2020-04-02 | L&T Technology Services Limited | Method and device for creating and training machine learning models |
US10402691B1 (en) * | 2018-10-04 | 2019-09-03 | Capital One Services, Llc | Adjusting training set combination based on classification accuracy |
US11294928B1 (en) | 2018-10-12 | 2022-04-05 | Palantir Technologies Inc. | System architecture for relating and linking data objects |
US11582093B2 (en) * | 2018-11-05 | 2023-02-14 | Cisco Technology, Inc. | Using stability metrics for live evaluation of device classification systems and hard examples collection |
US11537875B2 (en) * | 2018-11-09 | 2022-12-27 | International Business Machines Corporation | Detecting and reducing bias in machine learning models |
US11461644B2 (en) | 2018-11-15 | 2022-10-04 | D-Wave Systems Inc. | Systems and methods for semantic segmentation |
US10824915B2 (en) * | 2018-12-04 | 2020-11-03 | Capital One Services, Llc | Artificial intelligence system for inspecting image reliability |
US11468293B2 (en) | 2018-12-14 | 2022-10-11 | D-Wave Systems Inc. | Simulating and post-processing using a generative adversarial network |
CN109492420B (en) * | 2018-12-28 | 2021-07-20 | 深圳前海微众银行股份有限公司 | Model parameter training method, terminal, system and medium based on federated learning |
CN109902582B (en) * | 2019-01-28 | 2023-08-08 | 舒糖讯息科技(深圳)有限公司 | Action classification method and device, storage medium and terminal equipment |
US11900264B2 (en) | 2019-02-08 | 2024-02-13 | D-Wave Systems Inc. | Systems and methods for hybrid quantum-classical computing |
US11625612B2 (en) | 2019-02-12 | 2023-04-11 | D-Wave Systems Inc. | Systems and methods for domain adaptation |
US12093787B2 (en) | 2019-04-10 | 2024-09-17 | D-Wave Systems Inc. | Systems and methods for improving the performance of non-stoquastic quantum devices |
US11221837B2 (en) * | 2019-04-11 | 2022-01-11 | Microsoft Technology Licensing, Llc | Creating and deploying packages to devices in a fleet based on operations derived from a machine learning model |
US11029936B2 (en) | 2019-04-11 | 2021-06-08 | Microsoft Technology Licensing, Llc | Deploying packages to devices in a fleet in stages |
US11176429B2 (en) * | 2019-05-13 | 2021-11-16 | International Business Machines Corporation | Counter rare training date for artificial intelligence |
US10698704B1 (en) | 2019-06-10 | 2020-06-30 | Captial One Services, Llc | User interface common components and scalable integrable reusable isolated user interface |
JP7346110B2 (en) * | 2019-07-08 | 2023-09-19 | キヤノン株式会社 | Systems, methods and programs |
WO2021035412A1 (en) * | 2019-08-23 | 2021-03-04 | 华为技术有限公司 | Automatic machine learning (automl) system, method and device |
CN110580290B (en) * | 2019-09-12 | 2022-12-13 | 北京小米智能科技有限公司 | Method and device for optimizing training set for text classification |
US11593569B2 (en) * | 2019-10-11 | 2023-02-28 | Lenovo (Singapore) Pte. Ltd. | Enhanced input for text analytics |
US11699132B1 (en) | 2019-10-23 | 2023-07-11 | Relativity Oda Llc | Methods and systems for facilitating family-based review |
US11651276B2 (en) * | 2019-10-31 | 2023-05-16 | International Business Machines Corporation | Artificial intelligence transparency |
US10846436B1 (en) | 2019-11-19 | 2020-11-24 | Capital One Services, Llc | Swappable double layer barcode |
CN110991551B (en) * | 2019-12-13 | 2023-09-15 | 北京百度网讯科技有限公司 | Sample processing method, device, electronic equipment and storage medium |
CN111539443B (en) * | 2020-01-22 | 2024-02-09 | 北京小米松果电子有限公司 | Image recognition model training method and device and storage medium |
US12299593B2 (en) | 2020-02-05 | 2025-05-13 | D-Wave Systems Inc. | Systems and methods for optimizing annealing parameters |
US11501216B2 (en) * | 2020-02-21 | 2022-11-15 | King.Com Ltd. | Computer system, a computer device and a computer implemented method |
US10963792B1 (en) * | 2020-03-26 | 2021-03-30 | StradVision, Inc. | Method for training deep learning network based on artificial intelligence and learning device using the same |
KR102485303B1 (en) * | 2020-10-15 | 2023-01-05 | 한화시스템 주식회사 | Apparatus and mehod for labeling data |
CN113408566B (en) * | 2020-11-17 | 2024-08-20 | 腾讯科技(深圳)有限公司 | Target detection method and related equipment |
US12182671B2 (en) * | 2021-01-26 | 2024-12-31 | International Business Machines Corporation | Optimizing a machine learning system |
US11593681B2 (en) * | 2021-03-25 | 2023-02-28 | CAMP Systems International, Inc. | Synthesizing disparate database entries for hardware component identification |
EP4278297A4 (en) * | 2021-03-25 | 2024-08-07 | Camp Systems International, Inc. | TRAINING A MACHINE LEARNING MODEL FOR HARDWARE COMPONENT IDENTIFICATION |
US20240177833A1 (en) * | 2021-03-26 | 2024-05-30 | Carnegie Mellon University | System, Method, and Computer Program Product for Segmenting an Image |
US12266158B2 (en) * | 2021-03-31 | 2025-04-01 | Ferrum Health Inc. | Systems and methods for evaluating the error rate of human-generated data |
US20220374116A1 (en) * | 2021-05-24 | 2022-11-24 | Clarifai, Inc. | Systems and methods for improved annotation workflows |
US11854264B2 (en) * | 2021-06-18 | 2023-12-26 | Kyndryl, Inc. | Speculative actions based on predicting negative circumstances |
CN113627563B (en) * | 2021-08-20 | 2024-12-31 | 作业帮教育科技(北京)有限公司 | Label marking method, device and medium |
US12271443B1 (en) * | 2021-09-23 | 2025-04-08 | Scale AI, Inc. | Automatic data curation |
CN114861780B (en) * | 2022-04-24 | 2024-11-08 | 中国邮政储蓄银行股份有限公司 | Data label marking method, device and processor |
US20240320206A1 (en) * | 2023-03-24 | 2024-09-26 | Gm Cruise Holdings Llc | Identifying quality of labeled data |
US20250173359A1 (en) * | 2023-11-27 | 2025-05-29 | Capital One Services, Llc | Systems and methods for identifying data labels for submitting to additional data labeling routines based on embedding clusters |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020164070A1 (en) * | 2001-03-14 | 2002-11-07 | Kuhner Mark B. | Automatic algorithm generation |
US20080027725A1 (en) * | 2006-07-26 | 2008-01-31 | Microsoft Corporation | Automatic Accent Detection With Limited Manually Labeled Data |
US7512582B2 (en) * | 2003-12-10 | 2009-03-31 | Microsoft Corporation | Uncertainty reduction in collaborative bootstrapping |
-
2006
- 2006-10-31 US US11/555,036 patent/US7792353B2/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020164070A1 (en) * | 2001-03-14 | 2002-11-07 | Kuhner Mark B. | Automatic algorithm generation |
US7512582B2 (en) * | 2003-12-10 | 2009-03-31 | Microsoft Corporation | Uncertainty reduction in collaborative bootstrapping |
US20080027725A1 (en) * | 2006-07-26 | 2008-01-31 | Microsoft Corporation | Automatic Accent Detection With Limited Manually Labeled Data |
Non-Patent Citations (3)
Title |
---|
International Conference in Machine Learning, Morgan Kauffman, San Francisco, CA, 2001. |
L. Li, A. Pratap, H.-T. Lin, and Y. S. Abu-Mostafa, "Improving Generalization by Data Categorization," in A. Jorge et al., eds., Knowledge Discovery in Databases: PKDD 2005, vol. 3721 of Lecture Notes in Artificial Intelligence, pp. 157-168. Springer-Verlag, 2005. |
Melville, P. and Mooney, R, "Diverse ensembles for active learning," Proc. of the 21st Int'l Conf. on Machine Learning (ICML, Banff), 584-591, 2004. |
Cited By (82)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8655724B2 (en) * | 2006-12-18 | 2014-02-18 | Yahoo! Inc. | Evaluating performance of click fraud detection systems |
US20090092299A1 (en) * | 2007-10-03 | 2009-04-09 | Siemens Medical Solutions Usa, Inc. | System and Method for Joint Classification Using Feature Space Cluster Labels |
US20090116737A1 (en) * | 2007-10-30 | 2009-05-07 | Siemens Corporate Research, Inc. | Machine Learning For Tissue Labeling Segmentation |
US8170330B2 (en) * | 2007-10-30 | 2012-05-01 | Siemens Aktiengesellschaft | Machine learning for tissue labeling segmentation |
US8494257B2 (en) * | 2008-02-13 | 2013-07-23 | Museami, Inc. | Music score deconstruction |
US20090202144A1 (en) * | 2008-02-13 | 2009-08-13 | Museami, Inc. | Music score deconstruction |
US20100082627A1 (en) * | 2008-09-24 | 2010-04-01 | Yahoo! Inc. | Optimization filters for user generated content searches |
US8793249B2 (en) * | 2008-09-24 | 2014-07-29 | Yahoo! Inc. | Optimization filters for user generated content searches |
US20100312727A1 (en) * | 2008-12-19 | 2010-12-09 | Pottenger William M | Systems and methods for data transformation using higher order learning |
US8572071B2 (en) * | 2008-12-19 | 2013-10-29 | Rutgers, The State University Of New Jersey | Systems and methods for data transformation using higher order learning |
US20110007964A1 (en) * | 2009-07-10 | 2011-01-13 | Palo Alto Research Center Incorporated | System and method for machine-assisted human labeling of pixels in an image |
US8452086B2 (en) | 2009-07-10 | 2013-05-28 | Palo Alto Research Center Incorporated | System and user interface for machine-assisted human labeling of pixels in an image |
US8442319B2 (en) * | 2009-07-10 | 2013-05-14 | Palo Alto Research Center Incorporated | System and method for classifying connected groups of foreground pixels in scanned document images according to the type of marking |
US8649600B2 (en) | 2009-07-10 | 2014-02-11 | Palo Alto Research Center Incorporated | System and method for segmenting text lines in documents |
USRE47889E1 (en) | 2009-07-10 | 2020-03-03 | Iii Holdings 6, Llc | System and method for segmenting text lines in documents |
US8768057B2 (en) | 2009-07-10 | 2014-07-01 | Palo Alto Research Center Incorporated | System and method for segmenting text lines in documents |
US20110007366A1 (en) * | 2009-07-10 | 2011-01-13 | Palo Alto Research Center Incorporated | System and method for classifying connected groups of foreground pixels in scanned document images according to the type of marking |
US10742579B2 (en) | 2010-07-16 | 2020-08-11 | Firstwave Technology Pty Ltd. | Methods and systems for analysis and/or classification of information |
US20150100527A1 (en) * | 2010-07-16 | 2015-04-09 | First Wave Technology Pty Ltd | Methods and Systems for Analysis and/or Classification of Information |
US11924151B2 (en) | 2010-07-16 | 2024-03-05 | Firstwave Technology Pty Ltd | Methods and systems for analysis and/or classification of electronic information based on objects present in the electronic information |
US9813367B2 (en) * | 2010-07-16 | 2017-11-07 | Firstwave Technology Pty Ltd | Methods and systems for analysis and/or classification of information |
US20120246099A1 (en) * | 2011-03-23 | 2012-09-27 | Kabushiki Kaisha Toshiba | Learning device, learning method, and computer program product |
US8805752B2 (en) * | 2011-03-23 | 2014-08-12 | Kabushiki Kaisha Toshiba | Learning device, learning method, and computer program product |
US20140040169A1 (en) * | 2012-07-31 | 2014-02-06 | George Forman | Active learning with per-case symmetrical importance scores |
US8825570B2 (en) * | 2012-07-31 | 2014-09-02 | Hewlett-Packard Development Company, L.P. | Active learning with per-case symmetrical importance scores |
US9053434B2 (en) * | 2013-03-15 | 2015-06-09 | Hewlett-Packard Development Company, L.P. | Determining an obverse weight |
US20140279742A1 (en) * | 2013-03-15 | 2014-09-18 | Hewlett-Packard Development Company, L.P. | Determining an obverse weight |
EP2869237A1 (en) * | 2013-10-31 | 2015-05-06 | Kabushiki Kaisha Toshiba | Image display apparatus, image display method, and computer program product |
US10296803B2 (en) | 2013-10-31 | 2019-05-21 | Kabushiki Kaisha Toshiba | Image display apparatus, image display method, and computer program product |
US20170316548A1 (en) * | 2015-09-06 | 2017-11-02 | Boe Technology Group Co., Ltd. | Super-Resolution Method and System, Server, User Device and Method Therefor |
US10565684B2 (en) * | 2015-09-06 | 2020-02-18 | Boe Technology Group Co., Ltd. | Super-resolution method and system, server, user device and method therefor |
US10078688B2 (en) | 2016-04-12 | 2018-09-18 | Abbyy Production Llc | Evaluating text classifier parameters based on semantic features |
WO2017201540A1 (en) * | 2016-05-20 | 2017-11-23 | Techcyte, Inc. | Machine learning classification of particles or substances in digital microscopy images |
US12079887B2 (en) | 2016-11-11 | 2024-09-03 | Palantir Technologies Inc. | Graphical representation of a complex task |
US11715167B2 (en) | 2016-11-11 | 2023-08-01 | Palantir Technologies Inc. | Graphical representation of a complex task |
US10754863B2 (en) | 2017-04-05 | 2020-08-25 | Yandex Europe Ag | Method and system for ranking a plurality of documents on a search engine results page |
US10789533B2 (en) | 2017-07-26 | 2020-09-29 | LogMeln, Inc. | Generating a consistently labeled training dataset by automatically generating and displaying a set of most similar previously-labeled texts and their previously assigned labels for each text that is being labeled for the training dataset |
US10878144B2 (en) | 2017-08-10 | 2020-12-29 | Allstate Insurance Company | Multi-platform model processing and execution management engine |
US12190026B2 (en) | 2017-08-10 | 2025-01-07 | Allstate Insurance Company | Multi-platform model processing and execution management engine |
US11755949B2 (en) | 2017-08-10 | 2023-09-12 | Allstate Insurance Company | Multi-platform machine learning systems |
US11023824B2 (en) * | 2017-08-30 | 2021-06-01 | Intel Corporation | Constrained sample selection for training models |
US11010656B2 (en) | 2017-10-30 | 2021-05-18 | Clinc, Inc. | System and method for implementing an artificially intelligent virtual assistant using machine learning |
US10572801B2 (en) | 2017-11-22 | 2020-02-25 | Clinc, Inc. | System and method for implementing an artificially intelligent virtual assistant using machine learning |
US11042800B2 (en) | 2017-11-22 | 2021-06-22 | Cline, Inc. | System and method for implementing an artificially intelligent virtual assistant using machine learning |
US11874850B2 (en) | 2017-12-07 | 2024-01-16 | Palantir Technologies Inc. | Relationship analysis and mapping for interrelated multi-layered datasets |
US11308117B2 (en) | 2017-12-07 | 2022-04-19 | Palantir Technologies Inc. | Relationship analysis and mapping for interrelated multi-layered datasets |
US10296848B1 (en) * | 2018-03-05 | 2019-05-21 | Clinc, Inc. | Systems and method for automatically configuring machine learning models |
US10303978B1 (en) | 2018-03-26 | 2019-05-28 | Clinc, Inc. | Systems and methods for intelligently curating machine learning training data and improving machine learning model performance |
US10679100B2 (en) | 2018-03-26 | 2020-06-09 | Clinc, Inc. | Systems and methods for intelligently curating machine learning training data and improving machine learning model performance |
US11373257B1 (en) | 2018-04-06 | 2022-06-28 | Corelogic Solutions, Llc | Artificial intelligence-based property data linking system |
US11372900B1 (en) * | 2018-04-06 | 2022-06-28 | Corelogic Solutions, Llc | Artificial intelligence-based property data matching system |
US20190318261A1 (en) * | 2018-04-11 | 2019-10-17 | Samsung Electronics Co., Ltd. | System and method for active machine learning |
US11669746B2 (en) * | 2018-04-11 | 2023-06-06 | Samsung Electronics Co., Ltd. | System and method for active machine learning |
WO2019199072A1 (en) * | 2018-04-11 | 2019-10-17 | Samsung Electronics Co., Ltd. | System and method for active machine learning |
US11809974B2 (en) | 2018-10-30 | 2023-11-07 | Optum, Inc. | Machine learning for machine-assisted data classification |
US11416518B2 (en) * | 2018-10-30 | 2022-08-16 | Optum, Inc. | Machine learning for machine-assisted data classification |
US11443232B1 (en) | 2018-11-23 | 2022-09-13 | Amazon Technologies, Inc. | Active learning-based data labeling service using an augmented manifest |
US11481906B1 (en) | 2018-11-23 | 2022-10-25 | Amazon Technologies, Inc. | Custom labeling workflows in an active learning-based data labeling service |
US11048979B1 (en) * | 2018-11-23 | 2021-06-29 | Amazon Technologies, Inc. | Active learning loop-based data labeling service |
US10679150B1 (en) | 2018-12-13 | 2020-06-09 | Clinc, Inc. | Systems and methods for automatically configuring training data for training machine learning models of a machine learning-based dialogue system including seeding training samples or curating a corpus of training data based on instances of training data identified as anomalous |
CN109710788A (en) * | 2018-12-28 | 2019-05-03 | 斑马网络技术有限公司 | Image pattern mark and management method and equipment |
US11030492B2 (en) * | 2019-01-16 | 2021-06-08 | Clarifai, Inc. | Systems, techniques, and interfaces for obtaining and annotating training instances |
US20210256326A1 (en) * | 2019-01-16 | 2021-08-19 | Clarifai, Inc. | Systems, techniques, and interfaces for obtaining and annotating training instances |
US11977847B2 (en) * | 2019-03-05 | 2024-05-07 | Kensho Technologies, Llc | Dynamically updated text classifier |
US20230161971A1 (en) * | 2019-03-05 | 2023-05-25 | Kensho Technologies, Llc | Dynamically Updated Text Classifier |
US11556810B2 (en) | 2019-07-11 | 2023-01-17 | International Business Machines Corporation | Estimating feasibility and effort for a machine learning solution |
US20210109969A1 (en) | 2019-10-11 | 2021-04-15 | Kinaxis Inc. | Machine learning segmentation methods and systems |
US12346921B2 (en) | 2019-10-11 | 2025-07-01 | Kinaxis Inc. | Systems and methods for dynamic demand sensing and forecast adjustment |
US11875367B2 (en) | 2019-10-11 | 2024-01-16 | Kinaxis Inc. | Systems and methods for dynamic demand sensing |
US11886514B2 (en) | 2019-10-11 | 2024-01-30 | Kinaxis Inc. | Machine learning segmentation methods and systems |
US12271920B2 (en) | 2019-10-11 | 2025-04-08 | Kinaxis Inc. | Systems and methods for features engineering |
US12154013B2 (en) | 2019-10-15 | 2024-11-26 | Kinaxis Inc. | Interactive machine learning |
WO2021072537A1 (en) * | 2019-10-15 | 2021-04-22 | Kinaxis Inc. | Interactive machine learning |
US12242954B2 (en) * | 2019-10-15 | 2025-03-04 | Kinaxis Inc. | Interactive machine learning |
US20210110299A1 (en) * | 2019-10-15 | 2021-04-15 | Kinaxis Inc. | Interactive machine learning |
EP3905144A1 (en) * | 2020-04-30 | 2021-11-03 | UiPath, Inc. | Machine learning model retraining pipeline for robotic process automation |
US12321823B2 (en) | 2020-04-30 | 2025-06-03 | UiPath, Inc. | Machine learning model retraining pipeline for robotic process automation |
US11270438B2 (en) | 2020-06-12 | 2022-03-08 | Samasource Impact Sourcing, Inc. | System and method for triggering machine learning (ML) annotation model retraining |
US20210406472A1 (en) * | 2020-06-30 | 2021-12-30 | Hitachi, Ltd. | Named-entity classification apparatus and named-entity classification method |
US12339926B1 (en) | 2020-07-01 | 2025-06-24 | Massachusetts Mutual Life Insurance Company | Systems and methods for risk factor predictive modeling with dynamic training |
WO2022115199A1 (en) * | 2020-11-25 | 2022-06-02 | UiPath, Inc. | Robotic process automation architectures and processes for hosting, monitoring, and retraining machine learning models |
WO2022221488A3 (en) * | 2021-04-14 | 2022-12-08 | Capital One Services, Llc | Data labeling for training artificial intelligence systems |
Also Published As
Publication number | Publication date |
---|---|
US20080103996A1 (en) | 2008-05-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7792353B2 (en) | Retraining a machine-learning classifier using re-labeled training samples | |
US11481456B2 (en) | Model and pattern structure online unital learning: mapsoul | |
US11232152B2 (en) | Efficient processing of neighborhood data | |
US11354590B2 (en) | Rule determination for black-box machine-learning models | |
US11770571B2 (en) | Matrix completion and recommendation provision with deep learning | |
US20190354810A1 (en) | Active learning to reduce noise in labels | |
US7801836B2 (en) | Automated predictive data mining model selection using a genetic algorithm | |
US8589317B2 (en) | Human-assisted training of automated classifiers | |
US9466023B1 (en) | Data mining technique with federated evolutionary coordination | |
US12298982B2 (en) | Diversity and explainability parameters for recommendation accuracy in machine learning recommendation systems | |
US12182713B2 (en) | Multi-task equidistant embedding | |
US7406452B2 (en) | Machine learning | |
US20230252418A1 (en) | Apparatus for classifying candidates to postings and a method for its use | |
US20220414470A1 (en) | Multi-Task Attention Based Recurrent Neural Networks for Efficient Representation Learning | |
US12314536B1 (en) | Method and system for generating a user-sensitive user interface | |
US20250217452A1 (en) | Apparatus and methods for determining a probability datum | |
US20250217199A1 (en) | Apparatus and methods for determining a resource distribution | |
US11544626B2 (en) | Methods and systems for classifying resources to niche models | |
CN108304568B (en) | Real estate public expectation big data processing method and system | |
US20250217013A1 (en) | Apparatus and methods for model selection between a first model and a second model using projector inferencing | |
WO2025059297A1 (en) | Identification of electronic books for audiobook publication | |
US20220335313A1 (en) | Impact Score Based Target Action Assignment | |
Bielik et al. | Big data, good data, and residential floor plans: feature selection for maximizing the information value and minimizing redundancy in residential floor plan data sets | |
US12346302B1 (en) | Apparatus and method for updating a user data structure | |
US12124352B1 (en) | Apparatus and method generating a path using classified distractions |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:FORMAN, GEORGE;SUERMONDT, HENRI JACQUES;REEL/FRAME:018993/0206;SIGNING DATES FROM 20061102 TO 20061106 Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:FORMAN, GEORGE;SUERMONDT, HENRI JACQUES;SIGNING DATES FROM 20061102 TO 20061106;REEL/FRAME:018993/0206 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: HEWLETT PACKARD ENTERPRISE DEVELOPMENT LP, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P.;REEL/FRAME:037079/0001 Effective date: 20151027 |
|
AS | Assignment |
Owner name: ENTIT SOFTWARE LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HEWLETT PACKARD ENTERPRISE DEVELOPMENT LP;REEL/FRAME:042746/0130 Effective date: 20170405 |
|
AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., DELAWARE Free format text: SECURITY INTEREST;ASSIGNORS:ENTIT SOFTWARE LLC;ARCSIGHT, LLC;REEL/FRAME:044183/0577 Effective date: 20170901 Owner name: JPMORGAN CHASE BANK, N.A., DELAWARE Free format text: SECURITY INTEREST;ASSIGNORS:ATTACHMATE CORPORATION;BORLAND SOFTWARE CORPORATION;NETIQ CORPORATION;AND OTHERS;REEL/FRAME:044183/0718 Effective date: 20170901 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552) Year of fee payment: 8 |
|
AS | Assignment |
Owner name: MICRO FOCUS LLC, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:ENTIT SOFTWARE LLC;REEL/FRAME:050004/0001 Effective date: 20190523 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20220907 |
|
AS | Assignment |
Owner name: MICRO FOCUS LLC (F/K/A ENTIT SOFTWARE LLC), CALIFORNIA Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0577;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:063560/0001 Effective date: 20230131 Owner name: NETIQ CORPORATION, WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: MICRO FOCUS SOFTWARE INC. (F/K/A NOVELL, INC.), WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: ATTACHMATE CORPORATION, WASHINGTON Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: SERENA SOFTWARE, INC, CALIFORNIA Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: MICRO FOCUS (US), INC., MARYLAND Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: BORLAND SOFTWARE CORPORATION, MARYLAND Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 Owner name: MICRO FOCUS LLC (F/K/A ENTIT SOFTWARE LLC), CALIFORNIA Free format text: RELEASE OF SECURITY INTEREST REEL/FRAME 044183/0718;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:062746/0399 Effective date: 20230131 |