[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US20200034737A1 - Architectures for natural language processing - Google Patents

Architectures for natural language processing Download PDF

Info

Publication number
US20200034737A1
US20200034737A1 US16/289,481 US201916289481A US2020034737A1 US 20200034737 A1 US20200034737 A1 US 20200034737A1 US 201916289481 A US201916289481 A US 201916289481A US 2020034737 A1 US2020034737 A1 US 2020034737A1
Authority
US
United States
Prior art keywords
natural language
documents
language model
annotation
topical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/289,481
Inventor
Robert J. Munro
Schuyler D. Erle
Christopher Walker
Sarah K. Luger
Jason Brenier
Gary C. King
Paul A. Tepper
Ross Mechanic
Andrew Gilchrist-Scott
Jessica D. Long
James B. Robinson
Brendan D. Callahan
Michelle Casbon
Ujjwal Sarin
Aneesh Nair
Veena Basavaraj
Tripti Saxena
Edgar Nunez
Martha G. Hinrichs
Haley Most
Tyler J. Schnoebelen
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
AI IP Investments Ltd
Original Assignee
Aiparc Holdings Pte Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aiparc Holdings Pte Ltd filed Critical Aiparc Holdings Pte Ltd
Priority to US16/289,481 priority Critical patent/US20200034737A1/en
Publication of US20200034737A1 publication Critical patent/US20200034737A1/en
Priority to US17/166,493 priority patent/US20210232762A1/en
Assigned to Idibon, Inc. reassignment Idibon, Inc. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BASAVARAJ, VEENA, CALLAHAN, BRENDAN D., CASBON, MICHELLE, GILCHRIST-SCOTT, ANDREW, LUGER, SARAH K., MECHANIC, ROSS, WALKER, CHRISTOPHER, BRENIER, JASON, MUNRO, ROBERT J., KING, GARY C., ERLE, SCHUYLER D., SCHNOEBELEN, TYLER J., NAIR, ANEESH, SARIN, UJJWAL, HINRICHS, MARTHA G., LONG, JESSICA D., MOST, HALEY, NUNEZ, EDGAR, ROBINSON, JAMES B., SAXENA, TRIPTI, TEPPER, PAUL A.
Assigned to IDIBON (ASSIGNMENT FOR THE BENEFIT OF CREDITORS), LLC reassignment IDIBON (ASSIGNMENT FOR THE BENEFIT OF CREDITORS), LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Idibon, Inc.
Assigned to HEALY, TREVOR reassignment HEALY, TREVOR ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: IDIBON (ASSIGNMENT FOR THE BENEFIT OF CREDITORS), LLC
Assigned to AIPARC HOLDINGS PTE. LTD. reassignment AIPARC HOLDINGS PTE. LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HEALY, TREVOR
Assigned to AI IP INVESTMENTS LTD reassignment AI IP INVESTMENTS LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AIPARC HOLDINGS PTE. LTD.
Assigned to 100.CO, LLC reassignment 100.CO, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AI IP INVESTMENTS LTD.
Assigned to 100.CO TECHNOLOGIES, INC. reassignment 100.CO TECHNOLOGIES, INC. NUNC PRO TUNC ASSIGNMENT (SEE DOCUMENT FOR DETAILS). Assignors: 100.CO, LLC
Assigned to DAASH INTELLIGENCE, INC. reassignment DAASH INTELLIGENCE, INC. CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: 100.CO TECHNOLOGIES, INC.
Assigned to 100.CO GLOBAL HOLDINGS, LLC reassignment 100.CO GLOBAL HOLDINGS, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DAASH INTELLIGENCE, INC.
Assigned to AI IP INVESTMENTS LTD. reassignment AI IP INVESTMENTS LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: 100.CO GLOBAL HOLDINGS, LLC
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/169Annotation, e.g. comment data or footnotes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/242Query formulation
    • G06F16/243Natural language query formulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2453Query optimisation
    • G06F16/24532Query optimisation of parallel queries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/285Clustering or classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/28Databases characterised by their database models, e.g. relational or object models
    • G06F16/284Relational databases
    • G06F16/288Entity relationship models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/93Document management systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/951Indexing; Web crawling techniques
    • G06F17/28
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/137Hierarchical processing, e.g. outlines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/221Parsing markup language streams
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/42Data-driven translation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/01Social networking

Definitions

  • 1402805.00014_IDB014 titled “METHODS AND SYSTEMS FOR IMPROVING MACHINE LEARNING PERFORMANCE,” (Attorney Docket No. 1402805.000015_IDB015), titled “METHODS AND SYSTEMS FOR MODELING COMPLEX TAXONOMIES WITH NATURAL LANGUAGE UNDERSTANDING,” (Attorney Docket No. 1402805.00016_IDB016), titled “AN INTELLIGENT SYSTEM THAT DYNAMICALLY IMPROVES ITS KNOWLEDGE AND CODE-BASE FOR NATURAL LANGUAGE UNDERSTANDING,” (Attorney Docket No.
  • the subject matter disclosed herein generally relates to processing data.
  • the present disclosures relate to systems for generating natural language models.
  • a system for generating a natural language model may include: at least one memory and at least one processor communicatively coupled to the at least one memory; and a database module, an application program interface (API) module, a background processing module, and an applications module, each stored on the at least one memory and executable by the at least one processor; the API module configured to ingest training data representative of documents to be analyzed by the natural language model and to store the training data in the database module; the background processing module configured to: generate a hierarchical data structure, the hierarchical data structure comprising at least two topical nodes, wherein the at least two topical nodes represent partitions organized by two or more topical themes among the topical content of the training data within which the training data is to be subdivided into; select among the training data a plurality of documents to be annotated; generate at least one annotation prompt for each document among the plurality of documents to be annotated, said annotation prompt configured to elicit an annotation about said document indicating which node
  • the background processing module is further configured to test performance of the natural language model using a subset of the documents among the training data that received annotations. In some embodiments, the background processing module is further configured to: compute a performance metric of the natural language model, based on results of the testing; and determine whether the natural language model satisfies at least one performance criterion based on the computed performance metric. In some embodiments, the background processing module is further configured to: performing one or more optimization techniques configured to improve performance of the natural language platform, in response to determining that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric.
  • the one or more optimization techniques comprises at least one of: a feature selection process, a padding and rebalancing process of the natural language model, a pruning process of the natural language model, a feature discovery process, a smoothing process of the natural language model, or a model interpolation process.
  • the background processing module is further configured to: determine that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric; in response to said determining: identify a topical node among the two or more topical nodes of the hierarchal data structure that the natural language model fails to accurately categorize documents into; select a second plurality of documents to be annotated, the second plurality comprising documents associated with said topical node that the natural language model failed to accurately categorize documents into; and generate a second set of at least one annotation prompt for each document among the second plurality of documents to be annotated, said annotation prompt among the second set configured to elicit an annotation about said document to improve the natural language model in accurately categorizing documents into said topical node; wherein the applications module is further configured to cause display of the second set of the at least one annotation prompt for each document among the second plurality of documents to be annotated; wherein the API module is further configured to receive for each document among the second plurality of documents to be annotated
  • generating the hierarchical data structure comprises: performing a topic modeling process configured to identify two or more topics among the content of the training data that is configured to define the two or more topical nodes of the hierarchical data structure.
  • the background processing module is further configured to access one or more rules configured to instruct the natural language model how to categorize one or more documents into the two or more nodes of the hierarchical data structure.
  • generating the hierarchical data structure comprises: conducting a rules generation process configured to evaluate logical consistency among the one or more rules.
  • generating the hierarchical data structure comprises: generating at least one annotation prompt for each topical node among the two or more topical nodes in the hierarchical data structure, said annotation prompt configured to elicit an annotation about said topical node indicating a level of accuracy of placement of the node within the hierarchical data structure; causing display of the at least one annotation prompt for each topical node; and receiving for each topical node, the annotation in response to the displayed annotation prompt.
  • the background processing module is further configured to evaluate performance of the hierarchical data structure based on the annotations.
  • the background processing module is further configured to determine that the hierarchical data structure fails to satisfy at least one performance criterion in response to the evaluating; and modify a logical relationship among the two or more topical nodes based on the annotations and in response to determining that the data structure fails to satisfy the at least one performance criterion.
  • the API module is further configured to receive a training guideline based on the annotations to the nodes, the training guideline configured to provide instructions to an annotator for answering one or more annotation prompts for each document among the plurality of documents to be annotated.
  • the hierarchical data structure comprises at least a third topical node and a fourth topical node, wherein the third and fourth topical nodes both represent sub-partitions within the topical theme of the first node and organized by a third and fourth topical theme, respectively, among the topical content of the training data within which the training data is to be subdivided into.
  • a method for generating a natural language platform system configured to generate a natural language model. The method includes: deriving from an analogous natural language platform system, parameters configured to optimize performance of said analogous system to generate an analogous natural language model configured to analyze similar but not identical documents as the natural language model; interpolating said parameters to be optimized for documents to be analyzed by the natural language model; and implementing the interpolated parameters in the natural language platform system such that the interpolated parameters are configured to generate the natural language model.
  • this system may include: a full service natural language platform geographically located at a remote host location and configured to: receive training data through a network connection; train a natural language model based on the received training data; and generate predictions about untested data using the natural language model; and a connector module geographically located at a local client host location and communicatively coupled to the full service natural language platform through the network connection and configured to: access the training data stored in a client data store at the local client host location; format the training data in a uniform manner; transmit the training data to the full service natural language platform through the network connection; receive the predictions about the untested data; and store the predictions about the untested data in a memory at the local client host location.
  • system further includes a text extraction module communicatively coupled to the connector module and configured to attach to the client data store and extract textual data for use as the training data.
  • the full service natural language platform is further configured to: generate a hierarchical data structure, the hierarchical data structure comprising at least two topical nodes, wherein the at least two topical nodes represent partitions organized by two or more topical themes among the topical content of the training data within which the training data is to be subdivided into; select among the training data a plurality of documents to be annotated; and generate at least one annotation prompt for each document among the plurality of documents to be annotated, said annotation prompt configured to elicit an annotation about said document indicating which node among the at least two topical nodes of the hierarchal data structure said document is to be classified into.
  • the full service natural language platform is further configured to: cause display of the at least one annotation prompt for each document among the plurality of documents to be annotated; receive for each document among the plurality of documents to be annotated, the annotation in response to the displayed annotation prompt; and generate the natural language model using an adaptive machine learning process configured to determine, among the received annotations, patterns for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
  • the full service natural language platform is further configured to: test performance of the natural language model using a subset of the documents among the training data that received annotations; compute a performance metric of the natural language model, based on results of the testing; determine that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric; in response to said determining: identify a topical node among the two or more topical nodes of the hierarchal data structure that the natural language model fails to accurately categorize documents into; select a second plurality of documents to be annotated, the second plurality comprising documents associated with said topical node that the natural language model failed to accurately categorize documents into; generate a second set of at least one annotation prompt for each document among the second plurality of documents to be annotated, said annotation prompt among the second set configured to elicit an annotation about said document to improve the natural language model in accurately categorizing documents into said topical node; cause display of the second set of the at least one annotation prompt for each document among the
  • FIG. 1A is a network diagram illustrating an example network environment suitable for aspects of the present disclosure, according to some example embodiments.
  • FIG. 1B is diagram providing additional examples of networked environments for generating natural language models, according to some embodiments.
  • FIG. 2A is a diagram showing an example system architecture for performing aspects of the present disclosure, according to some example embodiments.
  • FIG. 2B is a diagram of example flowchart for how the system architecture according to the diagram in FIG. 2A may be utilized to generate other natural language platforms that are more tailored to a client's specific needs, according to some embodiments.
  • FIG. 3 is a high level diagram showing various examples of types of human communications and what the objectives may be for a natural language model to accomplish, according to some embodiments.
  • FIG. 4 is a diagram showing an example flowchart for how different data structures within the system architecture may be related to one another, according to some example embodiments.
  • FIG. 5 is a flowchart of a high-level process for generating a natural language model, according to some embodiments.
  • FIG. 6 is a diagram of a more detailed view of the block 505 of FIG. 5 , describing the ingest data phase, according to some embodiments.
  • FIG. 7 shows a simple example of an ontology.
  • FIG. 8 shows a more complex example of an ontology.
  • FIGS. 9A and 9B show an even more complex example of an ontology.
  • FIGS. 10A-10C provide additional example details for generating the ontology within the block 510 of FIG. 5 , according to some embodiments.
  • FIGS. 11A and 11B show example displays for annotating as part of the adaptive machine learning process, according to some embodiments.
  • FIG. 12 provides a more detailed process flow of the block 515 of FIG. 5 , for conducting the adaptive machine learning process using the annotations to generate the natural language model, according to some embodiments.
  • FIG. 13 provides further details of different examples of optimization techniques that may be applied in block 1240 of FIG. 12 for performing optimization techniques to improve the natural language model, according to some embodiments.
  • FIG. 14 illustrates a modified process for generating natural language models, including iterating processes between some processes within the create ontology block and the conduct adaptive machine learning block of FIG. 5 , according to some embodiments.
  • FIG. 15 provides an additional variant for generating natural language models, this time including a maintenance process for updating currently operating natural language models, according to some embodiments.
  • FIG. 16 is a block diagram illustrating components of a machine, according to some example embodiments, able to read instructions from a machine-readable medium and perform any one or more of the methodologies discussed herein.
  • Example methods, apparatuses, and systems are presented for generating natural language models.
  • natural language processing may involve humans declaring a large number of rules for the natural language model to follow. For example, when trying to determine how many communications convey negative sentiment about a politician, a rule may be generated to search for “horrible” and “Bill Clinton” in the same document. Similarly, a rule looking for the terms “wonderful” and “Bill Clinton” may attempt to find all the communications conveying positive sentiment about a politician. Conventionally, the natural language model may then parse a given number of documents, follow the rules, and attempt to categorize the documents by the rules.
  • natural language processing may involve applying machine-learning techniques to a set of example documents that have been categorized by human annotators, generating statistical inferences for the natural language model to follow.
  • a collection of documents written about “Bill Clinton” may include a subset of those documents understood and categorized by a human as conveying negative sentiment, and a second subset of documents understood and categorized as conveying positive sentiment.
  • the natural language models generated through machine learning approaches are able to recognize many more characteristics of sentiment-conveying text than can be reasonably implemented through rules-only approaches. Accordingly, the accuracy of such models at categorizing human communications may be substantially higher than rules-based natural language processing approaches.
  • aspects of the present disclosure are presented for generating natural language models (also referred to herein as prediction engines) through a combination of processing human annotations of documents and adaptive machine learning techniques that determine patterns in the documents, based in part on the human annotations applied to the documents.
  • the natural language platform used to generate the models such as a software system stored in memory of one or more servers communicatively coupled in parallel, may be configured to dynamically generate human readable annotation prompts that may be expressed in varying levels of detail and granularity so as to efficiently obtain specific annotation information and maximize use of a human annotator's time.
  • An iterative or cyclic process involving cycling between supplying prompts for human annotations to documents and obtaining the annotations, then generating the natural language model using machine learning of the documents with the annotations, then evaluating the model and iterating back to obtaining more human annotations to refine the model, and so forth, is also presented.
  • the human annotations may assist the machine learning techniques to resolve inevitable ambiguities in the human communications, as well as provide intelligence or meaning to communications that the machine does not accurately comprehend a priori.
  • the human annotations can then enable a natural language platform configured to generate these natural language models to provide better natural language results of the human communications, which can then in turn be better refined with the assistance of more human annotations as necessary.
  • a human annotator may provide an annotation about a Tweet reading “Can you believe this guy?!” as expressing negative sentiment of the person being referenced, while a rule may fail to flag this Tweet as having negative sentiment because no typical negative words are contained in the Tweet.
  • the natural language platform used to generate the natural language models may be configured to generate human readable annotation prompts at varying degrees of detail, so as to elicit responses from human annotators that match more specifically a level of ambiguity the natural language platform is trying to resolve.
  • a first level of annotation prompts may include determining what subject category (or categories) a document may be categorized into, among a plurality of first categories. This first level may therefore generate a prompt to the human annotator that includes a multiple choice question including three or more categories that the human annotator may choose from, where the human annotator may be prompted to select among any or all of the categories.
  • a second level of annotation prompts may include determining what subject category a document may be best categorized into, among a plurality of second categories. This second level may therefore generate a prompt displayable to the human annotator that includes a multiple-choice question including three or more categories that the human annotator may choose from, where the human annotator may be prompted to select only one category that best describes the document.
  • a third level of annotation prompts may include supplying a binary decision to the human annotator, such as asking whether it fits into a category or not (e.g., a yes or no or a true or false question).
  • the natural language platform may dynamically adjust the type of annotation prompts to supply to human annotators. For example, if the natural language model possesses little information about a document or a category of documents, the natural language platform may be configured to supply a more open ended question to human annotators, such as an annotation prompt at the first level. On the other hand, if the natural language model possesses precise information about a document or a category of documents, the natural language platform may be configured to supply a more precise question to the human annotators, such as an annotation prompt at the third level. In this way, information to improve the natural language model may be more efficiently obtained and may maximize the use of human annotators' time.
  • aspects of the present disclosure may construct natural language models based on this iterative process that can be specifically tailored to a customer's unique needs or subject matter area.
  • the words important to categorizing communications in biotechnology may be different than the words important to categorizing communications in the automobile industry.
  • the biotechnology user may desire to tailor a natural language model to better understand articles related to biotechnology, while the automobile industry user may desire to tailor the same or a different natural language model to better understand customer feedback emails.
  • the language, grammar, and idioms used in social media may vary drastically from communications in professional writings, e.g., legal or medical journals.
  • a user focusing on Twitter® communications may desire to tailor the natural language model to better determine when tweets of adolescent teens convey positive sentiment or negative sentiment, while a user focusing on legal documents may desire to tailor a natural language model to better understand whether a legal decision is favorable or unfavorable, and to see how many scholarly articles are written about the legal decision and what percentage of them are critical of the legal decision. Aspects of the present disclosure therefore are robust enough to generate a natural language model to interpret documents in any language (or languages) and for any number of topics or subject areas contained within the documents.
  • the natural language generation platform may be configured to process the human annotations to any of these areas in order to better inform the natural language model how to understand these complex communications.
  • the natural language model can be trained through the iterative process utilizing human annotations to more easily determine how to categorize any of these diverse areas of human communications.
  • the various categories or topics, sometimes referred to herein as labels, that the documents are intended to be grouped into may be organized by a hierarchical data structure, referred to herein as an ontology.
  • the hierarchical structure design of the ontology may define a plurality of subcategories, or sub labels, that thematically or topically fit within a larger umbrella label.
  • a set of documents about customer service emails of a telecommunications company may be categorized into an ontology where a first label of the ontology may be simply “customer service.” Under this first label, multiple sub labels may be included that thematically fall within the broader label of customer service, such as “Internet,” “phone,” and “customer service experience.” There may be further sub labels, such as “Check the router power,” which could be specific responses to questions in the customer service exchange, representing a single intention type of the person emailing the customer service representative: in this example, the intention to get solutions for a broken internet connection.
  • an ontology may include multiple levels of sub labels under higher levels of sub labels, and so forth.
  • an ontology may include multiple labels which define a plurality of common subcategories, that thematically or topically fit within each larger umbrella. In its totality, the ontology can represent the multiple label, information extracted, and sequence of responses to questions, for an entire business unit and organization.
  • these labels of the ontology may define and provide structure for the categorization of millions of documents by the natural language model. That is, developing the natural language model will include a goal of training the natural language model to accurately classify each of the documents in a set of documents into one or more of the labels in the hierarchical data structure.
  • a process for generating the natural language model includes a subprocess for first generating this ontology. Once generated, the ontology may be used to guide the aforementioned iterative process of obtaining human annotations and applying them in adaptive machine learning processes to generate a natural language model.
  • the natural language model can then be applied to untested human communications, and may generate an output that summarizes which human communications and how many are categorized into the various labels in the ontology structure (e.g., positive sentiment of a politician, negative sentiment of a politician).
  • the process conducted through the natural language model for classifying the thousands or even millions of documents into the specified topics may be referred to herein as a document-scope task.
  • a natural language model of the present disclosures may also be configured to extract specific subsets of text from documents (e.g., extract every phrase mentioning an anticipated cost or savings in a document discussing tax reform, or extract all words naming the crew in one or more documents relating to the latest blockbuster movie).
  • This process conducted through the natural language model for extracting specific types of text from one or more documents may be referred to herein as a span-scope task.
  • a span may refer to a subset of words or characters in a document, such as a paragraph, multiple paragraphs, sentences, or a plurality of words.
  • Performing a span scope task may also include first generating an ontology, and then conducting the iterative human annotation and adaptive machine learning process, in some embodiments.
  • a natural language platform of the present disclosures may also be configured to simply identify or discover what categories or subject areas (e.g., labels) the documents may be thematically organized into. For example, it may not be known what are the top twenty subject areas being most discussed in all tweets within the first week of September 2015. Before a natural language model can be generated to summarize what sentiment is being conveyed about these top twenty most discussed subject areas, it should be known what kinds of subject matters these tweets generally discuss in the first place. The process for discovering what topics or subject areas the documents may be thematically organized into may be referred to as topic modeling.
  • a natural language platform of the present disclosures may also be configured to utilize rules in combination with the iterative human annotation/machine learning process.
  • the rules may be imported from an existing body specified by a user or customer.
  • the natural language model may be generated with the aid of any pre-existing conditions specified by the user, such as specified topics or rules, and embodiments are not so limited.
  • a natural language model may be generated using previously generated natural language models that were tailored to different but similar subject areas.
  • the natural language model may be refined or updated through a maintenance process so that the model may be better suited to handle new data with an evolving body of words. For example, a natural language model for processing tweets may be refined over time to account for new slang or idioms that were not previously used in the common vernacular.
  • the example network environment 100 includes a server machine 110 , a database 115 , a first device 120 for a first user 122 , and a second device 130 for a second user 132 , all communicatively coupled to each other via a network 190 .
  • the server machine 110 may form all or part of a network-based system 105 (e.g., a cloud-based server system configured to provide one or more services to the first and second devices 120 and 130 ).
  • the server machine 110 , the first device 120 , and the second device 130 may each be implemented in a computer system, in whole or in part, as described below with respect to FIG. 16 .
  • the network-based system 105 may be an example of a natural language platform configured to generate natural language models as described herein.
  • the server machine 110 and the database 115 may be components of the natural language platform configured to perform these functions. While the server machine 110 is represented as just a single machine and the database 115 where is represented as just a single database, in some embodiments, multiple server machines and multiple databases communicatively coupled in parallel or in serial may be utilized, and embodiments are not so limited.
  • first and second users 122 and 132 may be a human user, a machine user (e.g., a computer configured by a software program to interact with the first device 120 ), or any suitable combination thereof (e.g., a human assisted by a machine or a machine supervised by a human).
  • the first user 122 may be associated with the first device 120 and may be a user of the first device 120 .
  • the first device 120 may be a desktop computer, a vehicle computer, a tablet computer, a navigational device, a portable media device, a smartphone, or a wearable device (e.g., a smart watch or smart glasses) belonging to the first user 122 .
  • the second user 132 may be associated with the second device 130 .
  • the second device 130 may be a desktop computer, a vehicle computer, a tablet computer, a navigational device, a portable media device, a, smartphone, or a wearable device (e.g., a smart watch or smart glasses) belonging to the second user 132 .
  • the first user 122 and a second user 132 may be examples of users or customers interfacing with the network-based system 105 to utilize a natural language model according to their specific needs.
  • the users 122 and 132 may be examples of annotators who are supplying annotations to documents to be used for training purposes when developing a natural language model.
  • the users 122 and 132 may be examples of analysts who are providing inputs to the natural language platform to more efficiently train the natural language model.
  • the users 122 and 132 may interface with the network-based system 105 through the devices 120 and 130 , respectively.
  • Any of the machines, databases 115 , or first or second devices 120 or 130 shown in FIG. 1A may be implemented in a general-purpose computer modified (e.g., configured or programmed) by software (e.g., one or more software modules) to be a special-purpose computer to perform one or more of the functions described herein for that machine, database 115 , or first or second device 120 or 130 .
  • software e.g., one or more software modules
  • a computer system able to implement any one or more of the methodologies described herein is discussed below with respect to FIG. 16 .
  • a “database” may refer to a data storage resource and may store data structured as a text file, a table, a spreadsheet, a relational database (e.g., an object-relational database), a triple store, a hierarchical data store, any other suitable means for organizing and storing data or any suitable combination thereof.
  • a relational database e.g., an object-relational database
  • a triple store e.g., an object-relational database
  • a hierarchical data store any other suitable means for organizing and storing data or any suitable combination thereof.
  • any two or more of the machines, databases, or devices illustrated in FIG. 1A may be combined into a single machine, and the functions described herein for any single machine, database, or device may be subdivided among multiple machines, databases, or devices.
  • the network 190 may be any network that enables communication between or among machines, databases 115 , and devices (e.g., the server machine 110 and the first device 120 ). Accordingly, the network 190 may be a wired network, a wireless network (e.g., a mobile or cellular network), or any suitable combination thereof. The network 190 may include one or more portions that constitute a private network, a public network (e.g., the Internet), or any suitable combination thereof.
  • the network 190 may include, for example, one or more portions that incorporate a local area network (LAN), a wide area network (WAN), the Internet, a mobile telephone network (e.g., a cellular network), a wired telephone network (e.g., a plain old telephone system (POTS) network), a wireless data network (e.g., WiFi network or WiMax network), or any suitable combination thereof. Any one or more portions of the network 190 may communicate information via a transmission medium.
  • LAN local area network
  • WAN wide area network
  • the Internet a mobile telephone network
  • POTS plain old telephone system
  • WiFi network e.g., WiFi network or WiMax network
  • transmission medium may refer to any intangible (e.g., transitory) medium that is capable of communicating (e.g., transmitting) instructions for execution by a machine (e.g., by one or more processors of such a machine), and can include digital or analog communication signals or other intangible media to facilitate communication of such software.
  • diagram 150 provides additional examples of networked environments for generating natural language models, according to some embodiments.
  • Diagram 150 provides a block diagram view of various configurations of a natural language platform configured to generate natural language models may interact with various data structures of a client, such as a user 120 or 130 .
  • a flow of data transfer for generating natural language models begins at a client data store 155 .
  • the client data store 155 may contain a repository of data that is used to train and ultimately generate the natural language model.
  • a text extraction module 157 may be installed or communicatively coupled to the client data store 155 and may be configured to extract textual data.
  • the client data store 155 may contain scanned copies of articles or printed text.
  • the text extraction module 157 may be configured to convert the scanned copies into computer readable text.
  • the textual data may be transmitted to a connector 160 , which may include software and hardware configured to act as a bridge that interfaces between the client data store 155 and one or more natural language platforms.
  • the textual data may be transmitted to a full-service natural language platform 170 through the network 190 .
  • the full-service natural language platform 170 may be located geographically distinct from the client, such as in a cloud environment.
  • the full-service natural language platform 170 may be configured to provide all available services for generating natural language models described herein, the various services of which will be described more fully below.
  • the full-service natural language platform 170 may be configured to generate a natural language model for the client by training the model on the textual data derived from the client data store 155 . As shown, this process may occur in the cloud environment.
  • the full-service natural language platform 170 may also be configured to apply the trained natural language model on untested, “live” data, referred to herein as providing predictions on the untrained and untested data.
  • the untrained or untested data may also be obtained from the client, such as from the client data store 155 . As shown, this process also may occur in the cloud environment.
  • the results of these predictions may be transmitted back to the client, through the network 190 , bridged by the connector 160 , and ultimately stored in a results repository 185 .
  • the textual data may be transmitted to a client-controlled natural language platform 175 .
  • the client control platform 175 may be stored locally and within the exclusive control of the client, according to some embodiments.
  • the client-controlled platform 175 may be configured to deploy a pre-generated natural language model, meaning the natural language model was already trained with previous data.
  • the arrow from the connector 160 to the client-controlled platform 175 lists only “prediction,” meaning that the ability to train a natural language model is not available in this option.
  • the client is able to utilize the pre-generated natural language model to make predictions on the untrained and untested data.
  • These results are supplied to the connector 160 , and ultimately stored in the results repository 185 .
  • the client may have exclusive control of the natural language model, and is not required to release its data to an off-site platform in order to utilize any natural language model.
  • Option #2 allows for greater privacy for the client.
  • the textual data may be transmitted to a client controlled natural language platform 180 and that provides full services.
  • the full-service natural language platform 170 may be duplicated locally at the client site. This option may be more costly to the client, but may be more comprehensive.
  • the services provided through the cloud 190 at the full-service natural language platform 170 may be substantially identical at the local full-service natural language platform 180 , according to some embodiments.
  • a diagram 200 is presented showing an example system architecture for performing aspects of the present disclosure, according to some example embodiments.
  • the example system architecture according to diagram 200 represents various data structures and their interrelationships that may comprise a natural language platform, such as the natural language platform 170 , or the network-based system 105 .
  • a natural language platform such as the natural language platform 170
  • the network-based system 105 may be implemented through a combination of hardware and software, the details of which may be apparent to those with skill in the art based on the descriptions of the various data structures described herein.
  • an API module 205 includes one or more API processors, where multiple API processors may be connected in parallel.
  • the repeating boxes in the diagram 200 represent identical servers or machines, to signify that the system architecture in diagram 200 may be scalable to an arbitrary degree.
  • the API module 205 may represent a point of contact for multiple other modules, includes a database module 210 , a cache module 215 , background processes module 220 , applications module 225 , and even an interface for users 235 in some example embodiments.
  • the API module 205 may be configured to receive or access data from database module 210 .
  • the data may include digital forms of thousands or millions of human communications.
  • the cache module 215 may store in more accessible memory various information from the database module 210 or from users 235 or other subscribers.
  • the background module 220 may be configured to perform a number of background processes for aiding natural language processing functionality.
  • Various examples of the background processes include a model training module, a cross validation module, an intelligent queuing module, a model prediction module, a topic modeling module, an annotation aggregation module, an annotation validation module.
  • the API module 205 may also be configured to support display and functionality of one or more applications in applications module 225 .
  • the users 235 may access the API module 205 , in some cases enabling the users 235 to create their own applications using the system architecture of diagram 200 .
  • the users 235 may be other examples of the users 120 or 130 , and may also include project managers and analysts.
  • Project managers may utilize the natural language platform to direct the overall construction of one or more natural language models.
  • Analysts may utilize the natural language platform to provide expert analysis and annotations to more efficiently train a natural language model.
  • annotators 230 may have access to applications already created in applications module 225 . Further discussion about the annotators is found more below.
  • system architecture according to diagram 200 may be scalable and reproducible at various client sites, in some cases consistent with the descriptions of FIG. 1B .
  • the database modules 210 and the cache module 215 may be implemented specifically for each client, such that each client does not share memory capacity with another client to ensure better privacy.
  • diagram 250 provides an example flowchart for how the system architecture according to diagram 200 may be utilized to generate other natural language platforms that are more tailored to a client's specific needs, according to some embodiments.
  • a pre-existing natural language platform according to the diagram 200 may have already been constructed at the client site.
  • the various API modules 205 , database modules 210 , cache modules 215 , background modules 220 , and display modules 225 of the pre-existing natural language platform may have been constructed and tailored to specific client needs and/or specific subject matter areas.
  • the pre-existing natural language platform may have been constructed to generate one or more natural language models related to comments on Twitter about baseball.
  • a new natural language platform for similar subject matter, it may be desirable to start with the pre-existing natural language platform. For example, it may be desirable to base a new natural language platform for a client that desires to process text from Twitter related to all professional sports on the pre-existing natural language platform configured to generate natural language models about baseball.
  • an interpolation operation 255 may be applied to the pre-existing natural language platform.
  • the interpolation operation 255 may include copying various modules, data and labels from the pre-existing natural language platform over to the new natural language platform.
  • the various background modules 220 may be calibrated to more efficiently handle text from Twitter, knowing the general size of tweets and having various libraries for understanding contemporary vernacular in tweets.
  • the cache sizes and other various settings in the cache modules 215 may also be attuned to efficiently handle tweets, as another example.
  • the same settings may be applied to the new natural language platform as a starting point.
  • the new natural language platform may be generated, and may include various adjustments from the pre-existing natural language platform.
  • the server computation resources assigned to the various background modules 220 may be adjusted in relation to the API modules 205 , so that more computation resources are available for intelligent queuing.
  • the natural language platform may be implemented by partitioning each of the API modules 205 , application modules 225 , background processing modules 220 , cache modules 215 and database modules 210 to be implemented within individual containers.
  • platform interpolation may be implemented by adjusting the proportion of processing resources (e.g., memory, network, storage and computation) assigned to each such container.
  • processing resources e.g., memory, network, storage and computation
  • Other components that may be interpolated from pre-existing and analogous platforms may include computational resources (CPU performance, memory) reserved for the background processing modules relative to the API module. If the interpolated system will be used primarily for prediction, rather than for topic modeling or re-training, then a greater share of CPU resources may be allocated to the API module.
  • Another example includes selecting an optimized default text encoding for the language(s) used in the interpolated systems. For example, if the interpolated system will only process Western European characters (no emoji, Cyrillic, or East Asian languages, et al.), then using ISO-8859-1 instead of UTF-8 can improve performance and reduce storage cost.
  • FIG. 3 a high level diagram 300 is presented showing various examples of types of human communications and what the objectives may be for a natural language model to accomplish. That is, while FIGS. 1A-2B have described the various structures of a natural language platform configured to generate one or more natural language models, FIG. 3 and the following figures will now discuss the various example functional objectives to be accomplished and various example technical processes used to achieve these functional objectives.
  • various sources of data sometimes referred to as a collection of documents 305 , may be obtained and stored in, for example database 115 , client data store 155 , or database modules 210 , and may represent different types of human communications, all capable of being analyzed by a natural language model.
  • Examples of the types of documents 305 include, but are not limited to, posts in social media, emails or other writings for customer feedback, pieces of or whole journalistic articles, commands spoken or written to electronic devices, and pieces of or whole scholarly texts.
  • Other types include transcribed call center recordings, electronic (instant) messages, corporate communications (e.g., SEC 10-k, 10-q), confidential documents and communications stored on internal collaboration systems (e.g., SharePoint, Notes), and questions and answers, either in isolation or as part of longer dialogues.
  • a user 130 in telecommunications may supply thousands of customer service emails related to services provided by a telecommunications company.
  • the user 130 may desire to have a natural language model generated that classifies the emails into predetermined categories, such as negative sentiment about their Internet service, positive sentiment about their Internet service, negative sentiment about their cable service, and positive sentiment about their cable service.
  • these various categories for which a natural language model may classify the emails into e.g.
  • labels “negative” sentiment about “Internet service,” “positive” sentiment about “Internet service,” “negative” sentiment about “cable service,” etc., may be referred to as “labels.” Based on these objectives, at block 315 , a natural language model may be generated that is tailored to classify these types of emails into these types of labels.
  • a span-scope task in reference to the fact that this function focuses on a subset within each document (as previously mentioned, referred to herein as a “span”).
  • a user 130 may desire to identify all instances of a keyword, key phrase, or general subject matter within a novel.
  • this span scope task may be applied to multiple novels or other documents.
  • a natural language model may be generated that is tailored to perform this function for a specified number of documents.
  • the user 130 may utilize the natural language platform only to perform topic modeling and to discover what topics are most discussed in a specified collection of documents 305 .
  • the natural language platform may be configured to conduct topic modeling analysis at block 330 . Topic modeling is discussed in more detail below, as well as in applications (Attorney Docket Nos. 1402805.00012_IDB012, 1402805.00013_IDB013, 1402805.00015_IDB015, and 1402805.00019_IDB019), each of which are again incorporated herein by reference in their entireties.
  • the natural language model may also be generated at block 315 .
  • the collections data structure 410 represents a set of documents 435 that in some cases may generally be homogenous.
  • a document 435 represents a human communication expressed in a single discrete package, such as a single tweet, a webpage, a chapter of a book, a command to a device, or a journal article, or any part thereof.
  • Each collection 410 may have one or more tasks 430 associated with it.
  • a task 430 may be thought of as a classification scheme. For example, a collection 410 of tweets may be classified by its sentiment, e.g.
  • a label 445 refers to a specific prediction about a specific classification.
  • a label 445 may be the “positive sentiment” of a human communication, or the “negative sentiment” of a human communication.
  • labels 445 can be applied to merely portions of documents 435 , such as paragraphs in an article or particular names or places mentioned in a document 435 .
  • a label 445 may be a “positive opinion” expressed about a product mentioned in a human communication, or a “negative opinion” expressed about a product mentioned in a human communication.
  • a task may be a sub-task of another task, allowing for a hierarchy or complex network of tasks. For example, if a task has a label of “positive opinion,” there might be sub-tasks for types of “positives opinions,” like “intention to purchase the product,” “positive review,” “recommendation to friend,” and so on, and there may be subtasks that capture other relevant information, such as “positive features.”
  • Annotations 440 refer to classifications imputed onto a collection 410 or a document 435 by an external source, such as human input, interpolation from an existing natural language model, or application of an external transformation process on the metadata included with the document (e.g., customer value, geographic location, etc.).
  • an annotation 440 applies a label 445 manually to a document 435 .
  • annotations 440 are provided by users 235 from pre-existing data.
  • annotations 440 may be derived from human critiques of one or more documents 435 , where the computer determines what annotation 440 should be placed on a document 435 (or collection 410 ) based on the human critique.
  • annotations 440 of a collection 410 can be derived from one or more patterns of pre-existing annotations found in the collection 410 or a similar collection 410 .
  • features 450 refer to a library or collection of certain key words or groups of words that may be used to determine whether a task 430 should be associated with a collection 410 or document 435 .
  • each task 430 has associated with it one or more features 450 that help define the task 430 .
  • features 450 can also include a length of words or other linguistic descriptions about the language structure of a document 435 , in order to define the task 430 .
  • classifying a document 435 as being a legal document may be based on determining if the document 435 contains a threshold number of words with particularly long lengths, words belonging to a pre-defined dictionary of legal-terms, or words that are related through syntactic structures and semantic relationships.
  • features 450 are defined by code, while in other cases features 450 are discovered by statistical methods. In some example embodiments, features 450 are treated independently, while in other cases features 450 are networked combinations of simpler features that are used in combination utilizing techniques like “deep-learning” or “topic modeling.” In some example embodiments, combinations of the methods described herein may be used to define the features 450 , and embodiments are not so limited.
  • One or more processors may be used to identify in a document 435 the words found in features data structure 450 to determine what task should be associated with the document 435 .
  • a work unit's data structure 455 specifies when humans should be tasked to further examine a document 425 .
  • human annotations may be applied to a document 435 after one or more work units 455 is applied to the document 435 .
  • the work units 455 may specify how many human annotators should examine the document 435 and in what order of documents should document 435 be examined.
  • work units 455 may also determine what annotations should be reviewed in a particular document 435 and what the optimal user interface should be for review.
  • the data structures 405 , 415 , 420 and 425 represent data groupings related to user authentication and user access to data in system architecture.
  • the subscribers block 405 may represent users and associated identification information about the users.
  • the subscribers 405 may have associated API keys 415 , which may represent one or more authentication data structures used to authenticate subscribers and provide access to the collections 410 .
  • Groups 420 may represent a grouping of subscribers based on one or more common traits, such as subscribers 405 belonging to the same company. Individual users 425 capable of accessing the collections 410 may also result from one or more groups 420 .
  • each group 420 , user 425 , or subscriber 405 may have associated with it a more personalized or customized set of collections 510 , documents 435 , annotations 440 , tasks, 430 , features 450 , and labels 445 , based on the specific needs of the customer.
  • flowchart 500 provides a high-level process for generating a natural language model, according to some embodiments.
  • the high-level process described in FIG. 5 may be implemented by a natural language platform, various examples of which have been described herein.
  • a natural language platform may be configured to ingest training data representative of documents that will ultimately be analyzed by the natural language model once the natural language model is fully trained. Examples of training data may include exactly the types of documents that are intended to be analyzed by the natural language model, including any of the example collections of documents 305 .
  • subsets of documents may be used for training data, such as paragraphs, excerpts, chapters, or articles of larger bodies of text.
  • the training data may be obtained from a number of different sources, including for example, from databases controlled by the client like the client data store 155 . In other cases, the training data may be obtained from public sources or from data sets used to derive previously trained natural language models. Example processes describing further details of how the natural language platform may be configured to ingest the training data will be described in more detail below.
  • the natural language platform may then be configured to create a hierarchical data structure used to organize the collection of documents in question, according to some embodiments.
  • this hierarchical data structure may be referred to as an ontology.
  • the ontology may include one or more labels that the documents are to be categorized into.
  • the one or more labels may be organized into a hierarchical structure, based on a logical relationship between the labels. For example, if one label may be thought of as having a logical subset of one or more additional labels, then the one or more additional labels may be represented as a sub node under the one label. Examples of this hierarchical tree structure will be described in more detail below.
  • the ontology may be created based on labels generated or discovered through the topic modeling process mentioned above. For example, if a client desires to generate a natural language model for performing the task of classification analysis of a random collection of tweets generated within the last two weeks of a calendar year, and the client wants to classify the collection of tweets into the top 10 most discussed topics, topic modeling may first be performed on the collection of tweets to discover precisely what these top 10 topics are. These top 10 topics may then be organized in a hierarchical structure of an ontology, and may include some of the topics as sub nodes under other topics, based on a logical or thematic relationship.
  • Christmas may be a very often discussed topic within the last two weeks of the calendar year, and within the general topic of “Christmas,” there may be discussed “presents” and “Santa Claus.”
  • the topics (also called labels) of “presents” and “Santa Claus” may be organized in the ontology as sub nodes of the “Christmas” label.
  • the ontology may also be created based on direct inputs from the client, such as manual inputs for labels the client knows or intends should be included in the ontology, as well as previously generated ontology's to be imported by the client.
  • the ontology may be created through an iterative verification process involving receiving annotations from expert annotators or other analysts. Additional details of these variants will be described more below.
  • the natural language platform may be configured to conduct an adaptive machine learning process to generate the natural language model.
  • This adaptive machine learning process may train the natural language model on how to categorize the training data into the various labels as structured in the ontology.
  • the natural language platform may be configured to conduct an iterative process including generating and receiving human annotations about the documents and then training the model based on these annotations.
  • the iterative process may be repeated based on a determination of whether the trained model meets various performance criteria.
  • the natural language platform may also be configured to conduct one or more optimization techniques before repeating the process of generating and receiving new human annotations and then training the model based on these annotations. Further details for conducting this adaptive machine learning process are described in more detail below.
  • the natural language model may be utilized by the natural language platform to analyze untested data as specified by the user or client.
  • this phase of the process may be referred to as predicting or generating predictions, and the utilization of the natural language model at this stage may be consistent with the discussion of predictions as described in FIG. 1B .
  • An example output for analyzing the untested data using the natural language model may include organizing the documents into the various labels and sub labels as specified in the ontology structure, or tagging individual documents with all relevant labels and sub-labels from the ontology.
  • various graphics in one or more graphical user interfaces may also be included and may being configured to graphically depict how the documents are organized within the ontology structure. Examples of these displays are described in application (Attorney Docket No. 1402805.00013_IDB013), which is again incorporated herein by reference in its entirety.
  • diagram 600 provides a more detailed view of the block 505 of FIG. 5 , describing the ingest data phase, according to some embodiments.
  • the natural language platform may be configured to ingest data related to one or more pre-existing ontologies.
  • the data of the pre-existing ontologies may be used to help build the ontology for the present natural language model, while in other cases one of the pre-existing ontologies may be used exactly as the intended ontology for the present natural language model.
  • These pre-existing ontologies may be obtained from the user or the client, for example, when the client has already performed some work in building the natural language model previously.
  • the natural language platform may also be configured to ingest existing classification rules for use in training the natural language model, according to some embodiments.
  • existing classification rules may be obtained from the user or the client, for example, when the client has already performed some work in building the natural language model previously.
  • the included rules may represent absolute conditions, such as always classifying a document when the document contains a specific word or phrase.
  • the included rules may represent probabilistic conditions that may be weighted against other probabilistic factors. For example, the rule may assign a weight to the document to be classified as a certain label, where said weight may be factored into an overall composite score including other weights based on other factors.
  • the natural language platform may also be configured to ingest test data in a specified data format.
  • An example of ingesting the test data may include the example processes for obtaining data from the client data store 155 in FIG. 1B .
  • Examples of a specified data format may include the test data being converted into all text, or the test data being converted into another uniform format.
  • the test data may be first ingested in its various formats, and then within the natural language platform, additional processes may convert the data into the uniform format.
  • the test data may be first converted into a uniform format, for example, at the client level, which is then ingested by the natural language platform.
  • the natural language platform may also be configured to perform various normalization and filtering procedures on the test data.
  • the documents converted into all text may then be broken up into individual units of language, referred to as tokens.
  • tokens can include single words, character sequences, punctuation marks, and spaces.
  • related and semantically-equivalent tokens may be converted into canonical forms.
  • canonical normalizations include decomposing character sequencing containing diacriticals (e.g., decomposing “A” to “A” and a combining grave accent) and jamo, expanding abbreviations into the original word (e.g., expanding “Assn.” to “Association”), and replacing alternate variations of emoticons with a primary form (e.g., replacing “;-)” or “ ⁇ _ ⁇ circumflex over ( ) ⁇ ” with “;)”).
  • Another example includes deduplication of identical, near-identical (retweet/quoted reply), and thematically similar items
  • each of the individual blocks 605 , 610 , 615 , and 620 may be performed in different orders, and in some cases not all of the example blocks 605 , 610 , 615 , and 620 need be performed.
  • the totality of the ingested data may be processed for use in the natural language platform for generating one or more ontologies and for conducting the adaptive machine learning process to generate one or more natural language models.
  • FIGS. 7-10 describe further details of block 510 , related to describing example details of various ontologies and an example detailed process for creating the ontology, according to some embodiments.
  • FIG. 7 a simple example of an ontology is shown through the various charts 700 , 710 , and 720 .
  • Chart 700 shows a spreadsheet of two tasks, with each task including multiple labels.
  • a “sentiment” task includes three labels, “positive,” “negative,” and “neutral.”
  • a user may desire to utilize a natural language model to perform a task of classifying documents by the sentiment each document conveys.
  • the various types of sentiment that the documents may be categorized into are enumerated by the labels under this sentiment task, i.e., “positive,” “negative,” and “neutral.”
  • the set of documents may be a collection of tweets about a basketball game, for example.
  • the views expressed in the tweets may therefore be classified as either having positive sentiment, negative sentiment, or neutral sentiment, and the user or client may desire to have a natural language model generated to evaluate these tweets and classify them into one of these various labels.
  • a “relevant” task includes two labels, “relevant,” and “irrelevant.”
  • the user may desire to utilize a natural language model to perform a task of classifying documents according to their relevance, where the determination of relevance may be specified by the user or associated analysts or annotators. For example, the determination of relevance may be based on whether a tweet among the collection of tweets pertains to the basketball game. The collection of tweets may have been obtained over the approximate period of time that the basketball games played, but since other events may be occurring at that time, it may not be a guarantee that all of the tweets necessarily discuss the basketball game.
  • the natural language model may be configured to evaluate these tweets and classify them as either relevant or irrelevant, according to the criteria used to train the model.
  • a column for “subtasks” is shown, which may represent connections between one task to another. Examples of subtasks will be discussed in the later figures. All told, these relationships according to the chart 700 comprise an example of a simple ontology. That is, a natural language model may perform the sentiment task and/or the relevant task to classify a collection of documents (or spans). The natural language model may perform these tasks by evaluating each document among the set of documents and classifying them into one or more of the labels under each task.
  • the text of the chart 700 is expressed in a textual format.
  • This example format may represent how the natural language model is configured to ingest the ontology so as to perform the proper analysis.
  • the example ontology formed by the words in chart 700 are shown here being organized in a hierarchical tree structure. That is, the overarching tasks, i.e., “relevant,” and “sentiment,” may represent the main nodes of the hierarchical structure, as shown.
  • the labels associated with the respective tasks i.e., “positive,” “negative,” and “neutral” for the “sentiment” task, and “irrelevant” and “relevant” for the “relevant” task, are placed under their respective tasks as sub nodes.
  • chart 800 a more complex example of an ontology is described.
  • chart 800 the same tasks and associated labels for each task as in FIG. 7 are shown.
  • a subtask of “sentiment” is listed and is associated with the label “relevant” under the task “relevant.” What this means, according to some embodiments, is that the task of “sentiment” and associated labels are to be embedded as sub nodes under the label “relevant.”
  • the text of the chart 800 is expressed in a textual format.
  • This example format may represent may be how the natural language model is configured to ingest the ontology so as to perform the proper analysis. Notice how the third word “sentiment” is placed after the label “relevant” and a delimiter (e.g., a comma).
  • the natural language platform may be configured to understand that the third word or phrase (e.g., words or phrases after a second delimiter the same line) represent a subtask, based on the organization defined in the first line of chart 810 .
  • the example ontology formed by the words in chart 800 is expressed in a hierarchical tree structure according to the organization defined by the headers in chart 800 . That is, because the task “sentiment” is listed as a subtask associated with the label “relevant,” the “sentiment” task and associated labels are placed as sub nodes under the “relevant” label, which in itself is a sub node of the “relevant” task, as shown. In this way, when the natural language model analyzes various documents (or spans of documents) according to the structure of this ontology, the natural language model will perform the sentiment analysis task only if the document has first been classified as being “relevant” when performing the “relevant” analysis task.
  • the natural language model analyzes a document (or a span) and determines that the document or span is to be classified as “irrelevant,” then the natural language model will discontinue any further analysis of the document or span, since the “irrelevant” label has no further sub nodes to travel down.
  • this example illustrates that the tree structure of the ontology can be used to guide the varying levels of analysis that a natural language model is to perform when analyzing documents.
  • FIGS. 9A and 9B an even more complex example of an ontology is shown.
  • an additional task called “genre” is included in the ontology.
  • the genre task is listed as a subtask under the “relevant” label.
  • the genre task includes four labels, “mystery,” “comedy,” “action,” and “horror.” Also, the mystery, comedy, and action labels have “sentiment” as a subtask, while the horror label has no subtask listed.
  • This example shows that a single task may be listed as a subtask under multiple labels, and that tasks can be listed as sub tasks in a cascading or hierarchical fashion (e.g., “sentiment” is a subtask of multiple labels in the “genre” task, which is itself a subtask of the “relevant” label of the “relevant” task).
  • the text of the chart 900 is expressed in a textual format.
  • This example format may represent may be how the natural language model is configured to ingest the ontology so as to perform the proper analysis.
  • the example ontology formed by the words in chart 900 is expressed in a hierarchical tree structure according to the organization defined by the headers in chart 900 .
  • the sentiment task is listed as a subtask for three of the labels in the genre task
  • the sentiment task and its associated labels are embedded as sub nodes under each of the three labels under the genre tasks, i.e., the “action,” “comedy,” and “mystery” labels.
  • the entire genre task and associated labels (and subsequent subtasks) is embedded as a sub node under the relevant label, consistent with the organization described in chart 900 .
  • a natural language model may conduct analysis of documents or spans of documents according to the hierarchical structure as shown. Following the hierarchical structure of the ontology, a natural language model may start by conducting “relevant” analysis of the document or span to determine if the document is “relevant” or “irrelevant.” If it is determined that the document is irrelevant, then the analysis ends. However, if the document is determined to be relevant, then the natural language model may proceed by conducting “genre” analysis, to determine whether the document should be classified as fitting into the “action” genre, the “comedy” genre, the “horror” genre, or the “mystery” genre. If classified as fitting into the horror genre, the analysis ends. However, if classified as any of the other three labels, the natural language model may continue by performing “sentiment” analysis, and classifying the document by its sentiment according to the sentiment labels “negative,” “neutral,” or “positive.”
  • the example ontology described herein is shown in a logical hierarchy format.
  • Various tasks in the ontology are shown in rectangular boxes, while the labels are shown in ovals.
  • the arrows show the progressive relationship for how a natural language model may proceed through the ontology when analyzing a document.
  • the first level analysis performed is to analyze relevance. If deemed relevant, the next level of analysis performed is genre analysis. If the document is deemed to be categorized into one of the three labels other than “horror,” then sentiment analysis is performed last.
  • This logical structure also illustrates how some ontologies may be structured in more complicated ways that a simple tree structure, as the analysis paths after “Mystery,” “Comedy” and “Action” all converge to continue performing sentiment analysis.
  • flowchart 1000 provides additional example details for generating the ontology, according to some embodiments.
  • the example processes of flowchart 1000 may represent further details in block 510 of the high level flow chart of FIG. 5 for generating natural language models. These example sub processes may begin at block 1005 , which refers to ingesting training data and other types of data.
  • Block 1005 may be consistent with block 505 and the related details described in flowchart 600 of FIG. 6 , where the example outputs at block 625 may serve as the example inputs into flowchart 1000 .
  • this data may include a collection of documents or spans of documents, one or more rules, and/or one or more pre-generated ontologies as defined by the client or user.
  • the natural language platform may be configured to first perform a topic modeling process, at block 1010 .
  • This topic modeling process may be consistent with those descriptions of topic modeling described above, as well as examples for performing topic modeling as described in applications (Attorney Docket Nos. 1402805.00012_IDB012, 1402805.00013_IDB013, 1402805.00014_IDB014, 1402805.00016_IDB016, 1402805.00017_IDB017, and 1402805.00019_IDB019), again incorporated herein by reference.
  • the topic modeling process may be used to determine one or more labels for the natural language model to categorize the collection of documents into, and of which may be organized into a hierarchical structure according to the following processes described below.
  • topic modeling may not be performed, and instead the natural language platform may access pre-generated or predetermined sets of topics as defined by the client or user or previously obtained in the ingest data phase.
  • the natural language platform may be configured to conduct a rules generation process to confirm and refine rules that may help define how the ontology is to be structured. For example, the natural language platform may access a number of rules, where some may be obtained from the previously ingested data, while other rules may be programmed or otherwise specified by human analysts with expert knowledge of the subject matter for which the natural language model is to be generated for. The natural language platform may receive these inputs about new rules or modifying pre-existing rules from the analysts through one or more graphical user interfaces, such as through the one or more application modules 225 as described in FIG. 2A . The rules generation process at block 1015 may then include verifying the enumerated rules for logical consistencies.
  • the natural language platform may be configured to evaluate the totality of rules to see if any of the rules lead to infinite circular loops, or lead to definitions of labels or tasks that lack proper node connections sufficient to fit in the ontology.
  • the rules may be evaluated to determine if two or more of the rules conflict with each other, or if two or more of the rules are redundant.
  • the natural language platform may be configured to generate a first version of the ontology based on the topics discovered or previously obtained and the verified rules in the previous steps.
  • the topics may define the content of the nodes for an ontology
  • the rules define the relationships among the various topics.
  • a rule may specify to perform a genre analysis task of a document after determining whether the document is determined to be relevant. Thus, this rule may help define that the genre task should be a sub node under the relevant task.
  • An example process for how rules contribute to generating an ontology includes the following steps:
  • an analyst may:
  • the natural language platform may be configured to test the effectiveness of the ontology, by first generating one or more annotation questions designed to prompt one or more annotators to evaluate the ontology. For example, an annotator might be asked “is there positive sentiment expressed about the car in this sentence?” or “is this statement an acceptable answer to this question?” These annotation prompts may be displayed in the one or more graphical user interfaces, such as through the application modules 225 . In some embodiments, analysts may interact with these annotation prompts at the client site, while in other cases the analysts may be stationed locally near the natural language platform.
  • illustration 1027 shows an example annotation interface to collect input in the form of document annotations/classifications from human analysts and other experts, according to some embodiments. These annotations can be used to evaluate the effectiveness of the ontology and to train the natural language processing models.
  • the panel on the left includes a sample document, while the panel on the right includes an annotation prompt and a series of labels in the ontology that the human analyst may select to specify under which label(s) the document belongs to.
  • the natural language platform may be configured to receive the various annotations entered by the human analysts, such as those inputs provided at the example interface in FIG. 10B . These annotations may provide a form of feedback to rate the quality of the ontology structure.
  • the natural language platform may be configured to evaluate the performance or effectiveness of the ontology based on the received annotations.
  • Evaluation metrics may include the label distribution (how frequently does the annotator select each label) and the proportion of annotated documents are not assigned any label.
  • illustration 1037 shows a sample annotation interface to collect input in the form of document annotations/classifications from human experts, according to some embodiments. These annotations can be used to evaluate the effectiveness of the ontology and to train the natural language processing models.
  • the ontology generation process within block 510 may involve refining the ontology through a series of iterations. For example, it may be determined at block 1035 that the ontology in its current form does not satisfy certain performance criteria. Again, these evaluations may be based on the feedback supplied by the annotations. To improve the ontology content and structure, at step 1040 , it may be determined that new rules should be generated or that existing rules may need to be modified. Thus, the ontology creation process may cycle back to block 1015 . From here, the natural language platform may facilitate generation of new rules or modification of existing rules through the graphical user interface, so as to receive inputs from analysts to modify the rules.
  • the rules generation process may again verify the veracity of the totality of rules to make sure that there is logical consistency. The process may then proceed to block 1020 , then to block 1025 , and so on to generate a refined ontology in a second iteration.
  • the natural language platform may cycle back to block 1020 and may refine the ontology by incorporating the feedback from the annotations. Similar to step 1040 , the ontology creation process may then proceed to block 1025 , then to block 1030 , and so on in a second or later iteration.
  • the iterative process for continually refining the ontology between steps 1040 and steps 1045 may continue a number of times until the evaluation of the ontology at block 1035 satisfies one or more performance criteria like inter-annotator agreement, using a metric derived from Krippendorf's Alpha. If it is determined that the performance criteria is satisfied at step 1050 , then at block 1055 , the ontology may be considered complete enough and may be used in the next step of the natural language model generation process, which is the adaptive machine learning process at block 515 .
  • analysts may evaluate the data and determined that certain clarifications may be needed to distinguish among various labels or tasks within the ontology. That is, part of the ontology refinement process may include trying to disambiguate two or more topics from each other, because the topics may be very closely related or potentially confusing. For example, when trying to classify news articles about a financial crisis, the label of “legal” articles may be potentially confusing or overlapping in some ways with the label of “security fraud.” While there may be a number of ways to distinguish between these two labels, the analysts may provide and the natural language platform may be configured to receive manual inputs to specify particularly how it is desired for these two topics to be distinguished.
  • training guidelines may be used for instructing annotators in the adaptive machine learning process phase.
  • These training guidelines may be expressed as human readable instructions that help define what the different labels mean, and/or how to choose among two or more labels when it is possible for a document to be classified into the two or more labels. For example, to distinguish between “legal” and “security fraud,” a training guideline may instruct the annotator to always classify the document as “security fraud” if it appears the document may fall under both categories. In other cases, the training guideline may attempt to provide definitions to help determine what constitutes “security fraud” compared to “legal” generally.
  • these training guidelines may be displayed to the annotators in one or more of the graphical user interfaces. An example of these training guidelines as described more in application (Attorney Docket No. 1402805.00013_IDB013), which again is incorporated herein by reference.
  • FIGS. 11A-13 will focus on more detailed examples of generating the natural language model through the adaptive machine learning process (see block 515 of FIG. 5 ), according to some embodiments.
  • illustration 1100 shows an example display for annotating as part of the adaptive machine learning process, according to some embodiments.
  • the adaptive machine learning process according to aspects of the present disclosure include an iterative interplay between annotations of documents and machine learning to generate the natural language model based on the annotations.
  • Illustration 1100 shows an example of one type of annotation applied to a document.
  • the display interface for allowing a human annotator to supply an annotation is called a work unit.
  • each work unit represents a single task for supplying a single annotation to one document (or span).
  • the natural language platform is configured to efficiently generate work units such that human annotators' time is efficiently utilized.
  • examples of a work unit may include the text from a document or at least a subset of text.
  • the document on the left side of the work unit is a tweet from Top World News.
  • the text states, among other things, “J.P. Morgan faces scrutiny over Asia hiring practices: J.P. Morgan Chase's hiring of the son of a Chinese . . . ”
  • This document also includes a photo and a title and introductory words to a news article that is referenced in the Tweet.
  • an annotation prompt On the right side of the work units is an annotation prompt, that includes a subject header and a human readable question.
  • the annotation prompt including the human readable question, may be generated automatically by the natural language platform.
  • the natural language platform may determine the form of the question so as to most efficiently retrieve the desirable annotation from the annotator.
  • the natural language platform has generated a human readable question that prompts the annotator to select as many topics (labels) that the annotator believes best applies to the document on the left.
  • This type of question is a rather open ended question, and in other cases, the natural language platform may be configured to generate more pointed questions when it has determined that it needs more specific information.
  • the annotator has selected two labels that he or she believes is most applicable: “legal & compliance,” and “corporate news.” Once submitted, this particular document would then have this particular annotation applied to it. Additional annotations may be applied to this document based on inputs from other annotators, when the other annotators see this particular work unit or other work units that include this particular document.
  • illustration 1150 represents another example of a work unit used to annotate a document, according to some embodiments.
  • the text of the document to be annotated is shown on the left.
  • the annotation prompt which includes a header and a human readable question for the annotator to answer.
  • the question of “What is the best label for this document?” allows for the annotator to select only one of three options: “highly relevant,” “semi-relevant,” and “totally irrelevant.”
  • the annotator may have been provided training guidelines to help define for the annotator what it means for something to be relevant, semi-relevant, or irrelevant.
  • a training guideline may have instructed the annotator to determine the relevance of the document to the topic of “corporate finance.”
  • the training guideline may have provided definitions to distinguish between the document being highly relevant versus merely semi-relevant.
  • a button or icon for the training guideline may be placed in the display of the work unit that displays the training guideline when the button or icon is pressed.
  • a previous display screen may have provided instructions based on the training guidelines to the annotator to prepare the annotator for how to respond to the annotation prompts.
  • the type of annotation prompt shown here is a more narrow scope than in illustration 1100 .
  • the annotator is restricted to selecting only one of the three options.
  • the annotator in illustration 1100 is allowed to select as many options as deemed relevant.
  • these differences in questions and their intended scopes are intentionally and automatically produced by the natural language platform and may be based on what type of information the natural language platform desires to gain from the annotators.
  • the types of annotation prompts may be more open ended, and may allow for annotators to answer broad questions, such as questions that may have multiple answers (e.g., see illustration 1100 ).
  • the natural language platform may desire specific information to determine differences between two specific labels.
  • the natural language platform may be configured to generate an annotation prompt that directs the annotator to distinguish the two specific labels using a pertinent document.
  • an example annotation prompt may include the question, “Is the best label for this document label ‘A’ or label ‘B’?” Then, the natural language platform allows the annotator to select between only label A or label B.
  • the natural language platform may be configured to generate annotation prompts with varying degrees of specificity, depending on the type of information desired to better refine the natural language model.
  • Other examples can include true/false questions, yes/no questions, and other types of questions that direct and annotator to varying levels of specificity apparent to those with skill in the art, and embodiments not so limited.
  • These varying degrees of specificity also help to efficiently designate a human's time. As there are thousands, millions, or even billions of documents that could be annotated, a human's time can be valuable and scarce, such that the techniques presented herein should determine what documents are most appropriate to be sent to human annotators, and what types of information about the documents should be obtained.
  • flowchart 1200 provides a more detailed process flow for conducting the adaptive machine learning process using the annotations to generate the natural language model, according to some embodiments.
  • the example details herein may represent further details in block 515 of FIG. 5 .
  • This sub process may start with block 1205 , involving natural language platform ingesting the data and inputting or applying an ontology for use in the adaptive machine learning process. Examples of these processes are described in previous figures, such as FIG. 6 and FIG. 10 .
  • the natural language platform may begin the adaptive machine learning process to generate a natural language model by selecting particular documents to be annotated. The process for selecting which documents are to be annotated may be referred to herein as “intelligent queuing.”
  • annotations to documents may be useful in aiding a computer to find patterns in the text of the collection of documents in order to determine how the document should be classified.
  • obtaining annotations for tens of thousands or even millions of documents in a collection may be impractical, and in some cases, a substantially fewer number of documents that are carefully selected may help produce a natural language model with substantially the same performance.
  • obtaining annotations for certain documents that are representative of the various topics or labels in the ontology as opposed to exhaustively obtaining annotations for all documents in a single topic, may be a more efficient use of a human annotator's time.
  • the process of performing intelligent queuing includes these processes of carefully selecting which documents among an entire collection of documents should be slated to receive annotations, with one objective being to efficiently utilize the human annotator's time.
  • the intelligent queuing process may include obtaining a seed set of documents that is representative of the various tasks and labels as outlined in the ontology. Other factors may be included in the seed set of documents, and further details of intelligent queuing are described in application (Attorney Docket No. 1402805.00012_IDB012), which again is incorporated herein by reference in its entirety.
  • the process of intelligent queuing at block 1210 may also include generating the annotation questions to be presented to an annotator (or multiple annotators) in order to elicit an annotation response.
  • annotation questions include the examples in FIGS. 11A and 11B .
  • the wording of the question including the level of specificity desired to be elicited from the annotator may be automatically generated by the natural language platform, based on the type of information desired when performing intelligent queuing.
  • block 1210 may also include causing display of the annotation prompt in a work unit that is accessible to an annotator. This may be displayed in one of the application modules 225 , for example. Other example displays are provided in application (Attorney Docket No. 1402805.00013_IDB013), which again is incorporated herein by reference.
  • providing annotation questions in the intelligent queuing process may also include presenting the annotation questions of the same document to multiple annotators.
  • a single annotator may not be relied on to provide the annotation to a single document.
  • a single annotator may make mistakes, or in worst cases a single annotator may intentionally provide wrong answers.
  • obtaining annotations from annotators may involve a crowdsourcing process, and it may not be known the reliability of any single annotator.
  • the natural language platform may be configured to receive annotations in response to the annotation questions generated in block 1210 .
  • one or more annotators may interact with a graphical user interface to analyze one or more documents and provide annotations according to the annotation prompts supplied by the natural language platform. For example, an annotator may answer a question similar to one of the questions in FIG. 11 A or 11 B. These responses may be stored in a database repository, such as in the database modules 210 .
  • the natural language platform may be configured to conduct an annotation aggregation process that analyzes all of the annotations made to a single document to determine whether said document may be reliably annotated for use in training the natural language model. For example, if six annotators provided annotations to the same document, and three annotators classified the document as belonging to a first label, while the other three annotators classified the document as belonging to a second label, then the annotation aggregation process of block 1220 may determine that this documents is too ambiguous to be used in the adaptive machine learning training process. As another example, if five of the six annotators provided a consistent annotation to the document, while a sixth annotator provided a different annotation, it may be determined that this document has sufficiently consistent annotations so as to be used in the adaptive machine learning process.
  • the annotation aggregation process of block 1220 may include computing an annotation aggregation score based on the total number of annotations and reflecting a level of agreement between the annotations. Based on this computed score, in some embodiments, the annotation aggregation process may determine whether the document may be used to train the natural language model, e.g., if the score satisfies a first threshold criterion. In addition, in some embodiments, based on the score, it may be determined that additional annotations from other annotators should be elicited before the document may be used in the model training process, e.g., if the score satisfies a second threshold criterion.
  • based on the score may be determined that the document should not be used for the model training process, e.g., if the score satisfies a third threshold criterion. This process may be applied for each document to determine if said document is valid for training the natural language model. Further details about the annotation aggregation process are described in application (Attorney Docket No. 1402805.00019_IDB019), which again is incorporated herein by reference in its entirety.
  • the natural language platform may be configured to generate the natural language model using the aggregated annotated data (e.g. documents with consistent enough aggregation score) in an adaptive machine learning training process.
  • the adaptive machine learning training process may find patterns in the training data (e.g., aggregated annotated data) that may be used to make predictions about untested data.
  • the process for adaptive machine learning includes: (a) elicit annotations from humans (b) build initial version of model (c) evaluate the effectiveness of the model and determine if stopping criteria have been met (d) identify specific areas of improvement and annotations required for improvement in those areas (e) queue additional documents for human annotations, and serve those documents to the optimal annotator with the optimal interface for the combination of annotator and task (f) update the model using the new annotations (g) repeat the whole process until stopping criteria are met. From this, a first version of the natural language model may be generated, and, if desired, continually updated new versions of the model as well.
  • the natural language platform first seeks to obtain annotations for particular documents before conducting any machine learning techniques.
  • Receiving annotations before conducting any machine learning techniques may allow for the adaptive machine learning process to be more efficient, in that the adaptive machine learning begins building the natural language model from documents whose “true values” are already known, e.g., the documents already have reliable annotations.
  • the natural language platform may be configured to test the performance of newly created natural language model, according to some embodiments. Testing the performance of the natural language model may allow the natural language platform (and/or a user) to determine later on whether the natural language model should be refined and improved before being finalized.
  • testing the performance of the natural language model at block 1230 includes performing a cross validation process.
  • the cross validation process may involve creating a first “test” natural language model, using the same adaptive machine learning training process to generate the original natural language model, but based on only a subset of the training data, e.g., 90% of the documents in the training data (e.g., supposing there are 1000 documents in the training data, then documents 1 - 900 ).
  • This first “test” natural language model is tested by applying the first “test” natural language model on the remaining 10% of the documents not used to create the first “test” natural language model, but still already having annotations.
  • the cross validation process may also involve creating a second “test” natural language model, using the same adaptive machine learning processes to generate the original natural language model, but based on a different subset of the training data, e.g., a different 90% of the training data (e.g., documents 1 - 800 and 901 - 1000 ).
  • This second test natural language model may be tested in the same way as the first test natural language model, and therefore may assess the veracity of the second test natural language model against the second 10% subset of the training data (e.g. documents 801 - 900 ).
  • This cross validation process may continue each for a different 90% of the training data (for a total of 10 different iterations), so that the veracity of the machine learning process used to the create original natural language model (and each of the test natural language models) may be verified against different subsets of the training data, and ultimately against all of the training data divided into separate parts, where the “true answers” are known. Because all of the training data is being tested against, and because the test natural language models are generated using the same machine learning process as the one used to generate the original natural language model in question, the aggregate performances of the test natural language models is used as a reliable approximation to the original natural language model. In general, subdividing the training data into 10% increments is merely one example, and other partitions may be used.
  • the natural language platform may be configured to generate a performance metric of the natural language model, based on the tests run against it. For example, if the performance of the natural language model was determined using the after mentioned cross validation techniques, one or more metrics quantifying a degree of accuracy may be supplied. In other cases, one accuracy measurement for each of the “test” models in the cross validation process may be supplied, and the natural language platform may be configured to provide an aggregate score at block 1235 . In some embodiments, these performance metrics may be displayed, such as in application modules 225 . Other examples of these performance metrics are described in application (Attorney Docket No.
  • Example performance metrics include: model precision, recall, and f1-measure, derived through cross-validation; area under the receiver operating characteristic curve (AUC); rate of improvement in precision, recall, and/or f1-measure as a function of the number of annotations; and percentage of the feature vocabulary in the total collection that is represented in the annotated data set.
  • the natural language platform may determine if the performance of the natural language model satisfies one or more performance criteria.
  • the one or more performance criteria may include a pre-determined threshold, or in other cases may be specified by a user.
  • the process may proceed to block 1245 . Otherwise, the process may proceed to block 1240 , where the natural language model is to be refined through one or more optimization techniques according to some embodiments.
  • the natural language platform may be configured to perform various optimization techniques to improve the performance of the natural language model, according to some embodiments. Numerous examples of optimization techniques are described in FIG. 13 below. Some or all of the techniques may be performed, and embodiments are not so limited.
  • one of the optimization techniques includes conducting another iteration of annotations on the same or different documents among the collection of documents.
  • the process in flowchart 1200 may cycle back to block 1210 , consistent with the iterative adaptive machine learning process described above.
  • the natural language platform may therefore be configured to conduct a second iteration of intelligent queuing that aims to obtain annotations from documents to help resolve labels that the original natural language model showed unsatisfactory performance in.
  • the adaptive machine learning process may continue through blocks 1215 through 1235 , consistent with the descriptions above but with new annotations and possibly other optimization techniques performed to generate the revised natural language model.
  • the model generation process may proceed straight to block 1225 to generate an updated natural language model using the updated parameters from the one or more optimization techniques and the existing aggregated annotations.
  • it may be determined that the natural language model satisfies the one or more performance criteria, at which point the process may proceed to block 1245 .
  • the finalized natural language model may be transmitted to make predictions on untested data.
  • the finalized natural language model may be transmitted to a system exclusively controlled by a client, such as in some of the examples shown in FIG. 2A , or in other cases may be transmitted to specific devices, such as devices 120 or 130 .
  • the transmitted natural language model may be stored in a format that is universally portable.
  • the natural language model may be stored in memory efficiently such that the same model may be executable in mobile devices as well as conventional computers. Additional details of the universal portability aspect according to some embodiments are described further in application (Attorney Docket No. 1402805.00018_IDB018), which is incorporated herein by reference in its entirety.
  • the natural language model may be used to analyze untested data.
  • the untested data may be based on data obtained from the client not used as part of the training data, while in other cases the untested data may include live or real-time data, such as live streams of tweets or customer service emails.
  • the natural language model may analyze untested data by classifying each unit of the untested data into one or more of the tasks or labels as organized in the ontology.
  • the natural language model may be used on millions or even billions of units of data.
  • illustration 1300 provides further details of different examples of optimization techniques that may be applied in block 1240 of FIG. 12 for performing optimization techniques to improve the natural language model, according to some embodiments.
  • the following are some descriptions of each of the different examples listed, although some embodiments may use only some of these examples, and embodiments are not so limited.
  • one example for performing optimization techniques to improve the natural language model may include using the results from cross validation or other performance evaluation techniques to guide the selection of further documents to be annotated when performing a next iteration of intelligent queuing.
  • additional annotations to documents may be sought after to help resolve some ambiguities in the natural language model or to help improve areas that the natural language model has shown difficulty in classifying.
  • the intelligent queuing process may base these decisions on the performance results generated by performing the cross validation or other performance evaluation techniques from block 1230 of FIG. 12 .
  • Block 1305 may also include a process for what type of annotation questions should be presented to an annotator to achieve the proper level of detail for improving the natural language model.
  • another example technique includes performing the intelligent queuing process itself, which is to say that the process proceeds back to block 1210 , which has already been discussed.
  • Feature selection may include applying one or more transformations to each annotated document to generate a set of machine learning features for that document, applying one or more evaluation metrics to each feature to estimate the relative importance of each feature to the natural language model, and finally selecting an appropriate subset of features for inclusion in the natural language model. For example, every token in each document may be designated as a feature, and each feature may be ordered based on the frequency it appears in the total set. Other examples may use statistical co-occurrence criteria (such as chi square and mutual information) or information theoretic criteria (such as information gain) to perform feature evaluation.
  • the feature transformation process is described in more detail in application (Attorney Docket No. 1402805.00017_IDB017), again incorporated herein by reference.
  • another example includes performing padding and rebalancing of the data in a natural language model.
  • This may include duplicating or deleting specific annotations to create an artificial mix of training items to improve model performance in specific areas.
  • Padding and rebalancing may include identifying individual items in the data that are representative of rare but significant labels in the ontology and artificially duplicating these items so that they make up a larger proportion of the training data used to develop the natural language models.
  • This technique can also be referred to as “upsampling.” For example, if positive items are rare relative to negative items in a set of training data used to develop a sentiment classification model, the positive items identified during annotation can be copied one or more times and included in the training data for the model.
  • the number of times that the positive items are copied can be determined through a number of optimization techniques in machine learning, including cross validation.
  • Cross validation can identify the number of positive and negative items to include in training that will achieve the best model performance (as measured by accuracy, precision, recall or f-score) by cyclically adjusting the numbers of items in each category, retraining the model, and testing the performance of the resulting model.
  • individual items in the data that cause the natural language model to make erroneous classifications may be removed from the training data to improve model performance. This technique can also be referred to as “downsampling.”
  • Pruning techniques may include processes to reduce the size and complexity of certain types of statistical models such as decision trees that are used to develop natural language models. Pruning has the benefit of reducing the tendency of a model to represent noise in the data (overfitting) by removing parts of the model that are useful in classifying only a small number of items in the training data. For example, a natural language model that is trained to distinguish five different topical categories of news articles using decision trees and individual word features, may perform well on training data, but poorly on unseen data. In order to improve the generalizability of the model and to make performance more consistent, certain nodes or features used by the model that only help in classifying less than 1% of the data may be trimmed away or removed to simplify the model and improve performance across different data sets.
  • another example includes performing ensemble or voting methods. This may include combining multiple pre-trained natural language processing models to generate multiple predictions for each unlabeled item in the data set, identifying the items in the dataset that meet specific prediction matching criteria (e.g., unanimous, majority, at least n), and using these items to train a new natural language processing model.
  • This approach may sometimes be referred to as semi-supervised machine learning.
  • the pre-trained models have been trained using a subset of training items from the target data set. In other embodiments, the pre-trained models have been trained using training items from external data sets. A variety of matching criteria can be applied to the resulting predictions to identify items that meet the minimum requirements for being included as training items in a new natural language model.
  • a majority vote of the models will determine the label that is assigned to an item in the dataset. For example, in an ensemble of six different sentiment models, if four models predict that an unlabeled tweet is positive, and two models predict that that same tweet is negative, the item will be assigned a positive label and subsequently used for training a new sentiment model. On the other hand, if three of the models predict that an unlabeled tweet is positive, and three models predict that it is negative, the item will not be assigned a label and will not be used for training a new sentiment model. In other embodiments, an unlabeled item will receive a label only if that label corresponds to at least n matching predictions from the ensemble of classifiers. In other embodiments, an unlabeled item will receive a label only if that label corresponds to a unanimous set of predictions from the ensemble of classifiers.
  • another example includes performing feature discovery. This may include identifying the set of feature transformations and feature selection criteria that provide the highest model performance for the provided data.
  • the set of selected feature transformations and feature selection criteria can be represented as a string of independent binary values. For example, each supported feature transformation or selection criteria may be assigned a position within the string, where a 1 value corresponds to the feature or criteria being enabled and a 0 corresponds to disabled.
  • a genetic search process may then be used to iteratively define numerous configurations, training and evaluating each in sequence according to the training and testing processes 1225 , 1230 and 1235 in FIG. 12 , refining and improving the configuration each iteration.
  • the final configuration of feature transformations and selection criteria may be saved to a database so that later applications of the feature selection process 1315 use the optimized configuration.
  • another example includes performing model interpolation. This may include combining model parameters from at least one previous model with a provided set of model parameters for the desired model to create an interpolated model which has greater performance than would be achieved from the provided set of model parameters.
  • the parameters incorporated from the at least one previous model may include the annotations and documents used to train each previous model, and the provided parameters for the desired model may include a set of label names, documents, annotations, and rules.
  • the interpolated model is generated by selecting the parameters from the at least one previous model which do not conflict with the provided parameters, and proceeding with the model training and optimization processes using the combined data set.
  • a previous model may include a large set of English social media posts from Twitter® discussing a variety of topics of general interest (movie releases, consumer electronics, recent personal experiences, and current events), and annotations imputed to each document according to the sentiment (“positive” or “negative”) conveyed by that document.
  • an interpolated model may be generated by providing a small amount of parameters (for example, a rule defining “overweight” as conveying “positive” sentiment), combined with the documents and annotations from the previous model parameters that do not conflict with this. In this way, highly accurate domain-specific models may be interpolated from a large amount of previous parameters and a comparatively small amount of provided parameters.
  • another example includes developing new rules for use in defining the ontology or additional annotations. Further descriptions for how the new rules may be implemented are described in FIG. 14 , below.
  • model hyper-parameters include regularization constraints applied to the entire model training process (such as a constraint on the L 1 or L 2 norms), e.g., at block 1355 , or adding a smoothing term to each feature individually during the feature selection process (such as a Laplacian smoothing term), e.g., at block 1360 .
  • regularization constraints applied to the entire model training process such as a constraint on the L 1 or L 2 norms
  • adding a smoothing term to each feature individually during the feature selection process such as a Laplacian smoothing term
  • Conventional machine learning techniques produce models which attempt to closely match their training data.
  • a conventional model training process may learn to interpret “Kenneth Starr” as strongly conveying “negative” sentiment, although this learned interpretation is an artifact of limited training examples rather than an actual conveyance of “negative” sentiment; such a case is known to those with skill in the art as an example of “over-fitting.”
  • introducing hyper-parameter constraints can be used to penalize rare terms, such that the model training process will not treat such terms as strong conveyances of any label.
  • each hyper-parameter may assume an infinite number of possible values (for example, a scale factor or additive term that may be assigned any numerical value).
  • selecting appropriate hyper-parameter values involves training and evaluating a large number of models with different hyper-parameter configurations.
  • Monte Carlo techniques e.g., block 1365 , such as the Metropolis-Hastings algorithm may be used to iteratively select and refine hyper-parameter values until a suitably accurate model is created or a maximum number of iterations have been performed.
  • the iterative process as described in flowchart 1200 may efficiently develop a natural language model by identifying and focusing on problem areas of the natural language model through multiple annotation and test phases. In this way, whereas conventional processes for developing a natural language model may take up to six months, achieving the same performance through aspects of the present disclosure may take only three weeks. In general, the processes described herein for producing a natural language model drastically reduce time to convergence and efficiently utilize time spent by humans to provide input for the machine learning components.
  • flowchart 1400 illustrates a modified process for generating natural language models, including iterating processes between some processes within the create ontology block and the conduct adaptive machine learning block of FIG. 5 , according to some embodiments.
  • iterative processes were described to occur within the create ontology block 510 and within the conduct adaptive machine learning block 515 .
  • a third relationship of iterations may also be allowed, such as processes within the adaptive machine learning block 515 iterating back to certain blocks within the create ontology block 510 .
  • blocks 1405 , 1410 , 1415 , and 1425 may include descriptions that are consistent with a first iteration of processes described in previous figures, such as in FIGS. 5, 10, and 12 , additional iterative loops may also be possible from the adaptive machine learning block 1425 .
  • the natural language platform may be configured to proceed back to block 1420 to allow for the ontology to be modified.
  • the natural language model may access inputs provided by analysts to modify one or more of the labels or tasks in the ontology.
  • the natural language platform may be configured to proceed back to block 1415 , which in some cases is a sub process within the overall process for creating the ontology.
  • the natural language platform may be configured to apply a weight to each of the rules, where the weight represents a degree to which the rules should be heeded in comparison to other rules.
  • modifying the rules may also include adjusting any weights to the rules. Modifying the rules may also involve creating new rules or modifying the logical parameters to which the rules should be applied. After any rules are modified at block 1415 , the process may again proceed to block 1420 , and continue down the process in another iteration.
  • the natural language platform may be configured to proceed back to block 1410 to conduct topic modeling again, which in some cases is a sub process within the overall process for creating the ontology.
  • the natural language platform may be configured to accept various parameters for discovering new topics. For example, new documents may be accessed, or certain documents may be given more weight.
  • the natural language platform may receive inputs to merge certain topics together into a single composite task or label within the ontology.
  • the natural language platform may be configured to receive inputs to divide one or more topics into multiple topics.
  • Analysts may be offered displays of various metrics that express levels of agreement or levels of performance of the natural language model when the topics are organized in one configuration versus another configuration. The analysts may then be able to compare how the model may improve if various topics are modified.
  • Example displays for these various interfaces for modifying the topics, the rules, or the ontology may be described further in application (Attorney Docket No. 1402805.00013_IDB013), which again is incorporated herein by reference.
  • the process may proceed directly to block 1420 to modify the ontology in light of the newly modified topics. In other cases, and may also be desirable to modify any rules, so the modification iteration may move from block 1410 to block 1415 before proceeding to block 1420 .
  • these iterative processes may cease once performance of the natural language model is determined to satisfy some threshold criterion. At that point, the natural language model may be available for use to analyze untested data at block 1430 .
  • flowchart 1500 provides an additional variant for generating natural language models, this time including a maintenance process for updating currently operating natural language models, according to some embodiments.
  • the various functional blocks in flowchart 1400 remain the same, but in addition, after the natural language model is in use to analyze untested data at block 1430 , an additional block 1510 is provided to modify the currently running natural language model with updated data.
  • the model may need to be updated periodically in order to adjust to changing language that may occur in certain collections of documents. For example, if the model is analyzing communications on Twitter, new idioms and slang may inevitably be communicated in tweets over time. A model trained six months ago may therefore lack proper training for how to handle any new changes in language.
  • the existing natural language model may simply be retrained with updated data.
  • the retraining may include revising or discovering new topics, modifying or creating new rules, or modifying the ontology structure to account for the updated data.
  • the natural language platform may be configured to retrain the model with updated data by proceeding to either block 1410 , 1415 , or 1420 , depending on the need.
  • modifications may include simply conducting one or more rounds of adaptive machine learning at block 1425 , first by performing intelligent queuing to generate new annotations for the updated data.
  • generating natural language processing models using the descriptions provided herein allows for: production of the models in a substantially shorter time period than conventional methods (e.g., highly accurate and fast performing models generated in a period of weeks rather than months), highly adaptive models that can evolve with changing input data, maintaining of consistent levels of model performance over time, the ability to leverage existing models and training data to develop domain-specific models with fewer new annotations, and the ability to add new labels and categorizations as requirements change.
  • the block diagram illustrates components of a machine 1600 , according to some example embodiments, able to read instructions 1624 from a machine-readable medium 1622 (e.g., a non-transitory machine-readable medium, a machine-readable storage medium, a computer-readable storage medium, or any suitable combination thereof) and perform any one or more of the methodologies discussed herein, in whole or in part.
  • a machine-readable medium 1622 e.g., a non-transitory machine-readable medium, a machine-readable storage medium, a computer-readable storage medium, or any suitable combination thereof
  • 16 shows the machine 1600 in the example form of a computer system (e.g., a computer) within which the instructions 1624 (e.g., software, a program, an application, an applet, an app, or other executable code) for causing the machine 1600 to perform any one or more of the methodologies discussed herein may be executed, in whole or in part.
  • the instructions 1624 e.g., software, a program, an application, an applet, an app, or other executable code
  • the machine 1600 operates as a standalone device or may be connected (e.g., networked) to other machines.
  • the machine 1600 may operate in the capacity of a server machine 110 or a client machine in a server-client network environment, or as a peer machine in a distributed (e.g., peer-to-peer) network environment.
  • the machine 1600 may include hardware, software, or combinations thereof, and may, as example, be a server computer, a client computer, a personal computer (PC), a tablet computer, a laptop computer, a netbook, a cellular telephone, a smartphone, a set-top box (STB), a personal digital assistant (PDA), a web appliance, a network router, a network switch, a network bridge, or any machine capable of executing the instructions 1624 , sequentially or otherwise, that specify actions to be taken by that machine.
  • PC personal computer
  • PDA personal digital assistant
  • the machine 1600 includes a processor 1602 (e.g., a central processing unit (CPU), a graphics processing unit (GPU), a digital signal processor (DSP), an application specific integrated circuit (ASIC), a radio-frequency integrated circuit (RFIC), or any suitable combination thereof), a main memory 1604 , and a static memory 1606 , which are configured to communicate with each other via a bus 1608 .
  • the processor 1602 may contain microcircuits that are configurable, temporarily or permanently, by some or all of the instructions 1624 such that the processor 1602 is configurable to perform any one or more of the methodologies described herein, in whole or in part.
  • a set of one or more microcircuits of the processor 1602 may be configurable to execute one or more modules (e.g., software modules) described herein.
  • the machine 1600 may further include a video display 1610 (e.g., a plasma display panel (PDP), a light emitting diode (LED) display, a liquid crystal display (LCD), a projector, a cathode ray tube (CRT), or any other display capable of displaying graphics or video).
  • a video display 1610 e.g., a plasma display panel (PDP), a light emitting diode (LED) display, a liquid crystal display (LCD), a projector, a cathode ray tube (CRT), or any other display capable of displaying graphics or video).
  • PDP plasma display panel
  • LED light emitting diode
  • LCD liquid crystal display
  • CRT cathode ray tube
  • the machine 1600 may also include an alphanumeric input device 1612 (e.g., a keyboard or keypad), a cursor control device 1614 (e.g., a mouse, a touchpad, a trackball, a joystick, a motion sensor, an eye tracking device, or other pointing instrument), a storage unit 1616 , a signal generation device 1618 (e.g., a sound card, an amplifier, a speaker, a headphone jack, or any suitable combination thereof), and a network interface device 1620 .
  • an alphanumeric input device 1612 e.g., a keyboard or keypad
  • a cursor control device 1614 e.g., a mouse, a touchpad, a trackball, a joystick, a motion sensor, an eye tracking device, or other pointing instrument
  • a storage unit 1616 e.g., a storage unit 1616 , a signal generation device 1618 (e.g., a sound card, an amplifier, a speaker,
  • the storage unit 1616 includes the machine-readable medium 1622 (e.g., a tangible and non-transitory machine-readable storage medium) on which are stored the instructions 1624 embodying any one or more of the methodologies or functions described herein, including, for example, any of the descriptions of FIGS. 1A-15 .
  • the instructions 1624 may also reside, completely or at least partially, within the main memory 1604 , within the processor 1602 (e.g., within the processor's cache memory), or both, before or during execution thereof by the machine 1600 .
  • the instructions 1624 may also reside in the static memory 1606 .
  • the main memory 1604 and the processor 1602 may be considered machine-readable media 1622 (e.g., tangible and non-transitory machine-readable media).
  • the instructions 1624 may be transmitted or received over a network 1626 via the network interface device 1620 .
  • the network interface device 1620 may communicate the instructions 1624 using any one or more transfer protocols (e.g., HTTP).
  • the machine 1600 may also represent example means for performing any of the functions described herein, including the processes described in FIGS. 1A-15 .
  • the machine 1600 may be a portable computing device, such as a smart phone or tablet computer, and have one or more additional input components (e.g., sensors or gauges) (not shown).
  • additional input components e.g., sensors or gauges
  • input components include an image input component (e.g., one or more cameras), an audio input component (e.g., a microphone), a direction input component (e.g., a compass), a location input component (e.g., a GPS receiver), an orientation component (e.g., a gyroscope), a motion detection component (e.g., one or more accelerometers), an altitude detection component (e.g., an altimeter), and a gas detection component (e.g., a gas sensor).
  • Inputs harvested by any one or more of these input components may be accessible and available for use by any of the modules described herein.
  • the term “memory” refers to a machine-readable medium 1622 able to store data temporarily or permanently and may be taken to include, but not be limited to, random-access memory (RAM), read-only memory (ROM), buffer memory, flash memory, and cache memory. While the machine-readable medium 1622 is shown in an example embodiment to be a single medium, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database 115 , or associated caches and servers) able to store instructions 1624 .
  • machine-readable medium shall also be taken to include any medium, or combination of multiple media, that is capable of storing the instructions 1624 for execution by the machine 1600 , such that the instructions 1624 , when executed by one or more processors of the machine 1600 (e.g., processor 1602 ), cause the machine 1600 to perform any one or more of the methodologies described herein, in whole or in part.
  • a “machine-readable medium” refers to a single storage apparatus or device 120 or 130 , as well as cloud-based storage systems or storage networks that include multiple storage apparatus or devices 120 or 130 .
  • machine-readable medium shall accordingly be taken to include, but not be limited to, one or more tangible (e.g., non-transitory) data repositories in the form of a solid-state memory, an optical medium, a magnetic medium, or any suitable combination thereof.
  • the machine-readable medium 1622 is non-transitory in that it does not embody a propagating signal. However, labeling the tangible machine-readable medium 1622 as “non-transitory” should not be construed to mean that the medium is incapable of movement; the medium should be considered as being transportable from one physical location to another. Additionally, since the machine-readable medium 1622 is tangible, the medium may be considered to be a machine-readable device.
  • Modules may constitute software modules (e.g., code stored or otherwise embodied on a machine-readable medium 1622 or in a transmission medium), hardware modules, or any suitable combination thereof.
  • a “hardware module” is a tangible (e.g., non-transitory) unit capable of performing certain operations and may be configured or arranged in a certain physical manner.
  • one or more computer systems e.g., a standalone computer system, a client computer system, or a server computer system
  • one or more hardware modules of a computer system e.g., a processor 1602 or a group of processors 1602
  • software e.g., an application or application portion
  • a hardware module may be implemented mechanically, electronically, or any suitable combination thereof.
  • a hardware module may include dedicated circuitry or logic that is permanently configured to perform certain operations.
  • a hardware module may be a special-purpose processor, such as a field programmable gate array (FPGA) or an ASIC.
  • a hardware module may also include programmable logic or circuitry that is temporarily configured by software to perform certain operations.
  • a hardware module may include software encompassed within a general-purpose processor 1602 or other programmable processor 1602 . It will be appreciated that the decision to implement a hardware module mechanically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) may be driven by cost and time considerations.
  • Hardware modules can provide information to, and receive information from, other hardware modules. Accordingly, the described hardware modules may be regarded as being communicatively coupled. Where multiple hardware modules exist contemporaneously, communications may be achieved through signal transmission (e.g., over appropriate circuits and buses 1608 ) between or among two or more of the hardware modules. In embodiments in which multiple hardware modules are configured or instantiated at different times, communications between such hardware modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple hardware modules have access. For example, one hardware module may perform an operation and store the output of that operation in a memory device to which it is communicatively coupled. A further hardware module may then, at a later time, access the memory device to retrieve and process the stored output. Hardware modules may also initiate communications with input or output devices, and can operate on a resource (e.g., a collection of information).
  • a resource e.g., a collection of information
  • processors 1602 may be temporarily configured (e.g., by software) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors 1602 may constitute processor-implemented modules that operate to perform one or more operations or functions described herein. As used herein, “processor-implemented module” refers to a hardware module implemented using one or more processors 1602 .
  • processors 1602 being an example of hardware.
  • processor-implemented module refers to a hardware module in which the hardware includes one or more processors 1602 .
  • processors 1602 may also operate to support performance of the relevant operations in a “cloud computing” environment or as a “software as a service” (SaaS).
  • At least some of the operations may be performed by a group of computers (as examples of machines 1600 including processors 1602 ), with these operations being accessible via a network 1626 (e.g., the Internet) and via one or more appropriate interfaces (e.g., an API).
  • a network 1626 e.g., the Internet
  • one or more appropriate interfaces e.g., an API
  • the performance of certain operations may be distributed among the one or more processors 1602 , not only residing within a single machine 1600 , but deployed across a number of machines 1600 .
  • the one or more processors 1602 or processor-implemented modules may be located in a single geographic location (e.g., within a home environment, an office environment, or a server farm). In other example embodiments, the one or more processors 1602 or processor-implemented modules may be distributed across a number of geographic locations.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Mathematical Physics (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Human Computer Interaction (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Tourism & Hospitality (AREA)
  • Economics (AREA)
  • Human Resources & Organizations (AREA)
  • Marketing (AREA)
  • Primary Health Care (AREA)
  • Strategic Management (AREA)
  • Animal Behavior & Ethology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Machine Translation (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Systems are presented for generating a natural language model. The system may comprise a database module, an application program interface (API) module, a background processing module, and an applications module, each stored on the at least one memory and executable by the at least one processor. The system may be configured to generate the natural language model by: ingesting training data, generating a hierarchical data structure, selecting a plurality of documents among the training data to be annotated, generating an annotation prompt for each document configured to elicit an annotation about said document, receiving the annotation based on the annotation prompt, and generating the natural language model using an adaptive machine learning process configured to determine patterns among the annotations for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.

Description

    CROSS REFERENCES TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 14/964,518, filed Dec. 9, 2015, and titled “ARCHITECTURES FOR NATURAL LANGUAGE PROCESSING,” which claims the benefits of U.S. Provisional Application 62/089,736, filed Dec. 9, 2014, and titled, “METHODS AND SYSTEMS FOR ANNOTATING NATURAL LANGUAGE PROCESSING,” U.S. Provisional Application 62/089,742, filed Dec. 9, 2014, and titled, “METHODS AND SYSTEMS FOR IMPROVING MACHINE PERFORMANCE IN NATURAL LANGUAGE PROCESSING,” U.S. Provisional Application 62/089,745, filed Dec. 9, 2014, and titled, “METHODS AND SYSTEMS FOR IMPROVING FUNCTIONALITY IN NATURAL LANGUAGE PROCESSING,” and U.S. Provisional Application 62/089,747, filed Dec. 9, 2014, and titled, “METHODS AND SYSTEMS FOR SUPPORTING NATURAL LANGUAGE PROCESSING,” the disclosures of which are incorporated herein in their entireties and for all purposes.
  • This application is also related to US non provisional applications (Attorney Docket No. 1402805.00006_IDB006), titled “METHODS FOR GENERATING NATURAL LANGUAGE PROCESSING SYSTEMS,” (Attorney Docket No. 1402805.00012_IDB012), titled “OPTIMIZATION TECHNIQUES FOR ARTIFICIAL INTELLIGENCE,” (Attorney Docket No. 1402805.00013_IDB013), titled “GRAPHICAL SYSTEMS AND METHODS FOR HUMAN-IN-THE-LOOP MACHINE INTELLIGENCE,” (Attorney Docket No. 1402805.00014_IDB014), titled “METHODS AND SYSTEMS FOR IMPROVING MACHINE LEARNING PERFORMANCE,” (Attorney Docket No. 1402805.000015_IDB015), titled “METHODS AND SYSTEMS FOR MODELING COMPLEX TAXONOMIES WITH NATURAL LANGUAGE UNDERSTANDING,” (Attorney Docket No. 1402805.00016_IDB016), titled “AN INTELLIGENT SYSTEM THAT DYNAMICALLY IMPROVES ITS KNOWLEDGE AND CODE-BASE FOR NATURAL LANGUAGE UNDERSTANDING,” (Attorney Docket No. 1402805.00017_IDB017), titled “METHODS AND SYSTEMS FOR LANGUAGE-AGNOSTIC MACHINE LEARNING IN NATURAL LANGUAGE PROCESSING USING FEATURE EXTRACTION,” (Attorney Docket No. 1402805.00018_IDB018), titled “METHODS AND SYSTEMS FOR PROVIDING UNIVERSAL PORTABILITY IN MACHINE LEARNING,” and (Attorney Docket No. 1402805.00019_IDB019), titled “TECHNIQUES FOR COMBINING HUMAN AND MACHINE LEARNING IN NATURAL LANGUAGE PROCESSING,” each of which were filed concurrently with U.S. patent application Ser. No. 14/964,518, and the entire contents and substance of all of which are hereby incorporated in total by reference in their entireties and for all purposes.
  • TECHNICAL FIELD
  • The subject matter disclosed herein generally relates to processing data. In some example embodiments, the present disclosures relate to systems for generating natural language models.
  • BACKGROUND
  • It has long been a goal to program machines to process human readable language, sometimes in part as an effort to generate artificial intelligence. However, programming computers to process human readable language has proven to be far more difficult than imagined, particularly as languages continue to change and evolve, and the meaning of words and phrases are more ambiguous and nuanced than assumed. A number of techniques are available for processing natural language by computers, but the methods for generating these models either are inaccurate and imprecise or require months of refinement and programming to accurately model specific subject areas of language. It is desirable therefore to develop improved methods for generating natural language models that are accurate and quick while also reducing human time spent generating the models.
  • BRIEF SUMMARY
  • In some embodiments, a system for generating a natural language model is presented. The system may include: at least one memory and at least one processor communicatively coupled to the at least one memory; and a database module, an application program interface (API) module, a background processing module, and an applications module, each stored on the at least one memory and executable by the at least one processor; the API module configured to ingest training data representative of documents to be analyzed by the natural language model and to store the training data in the database module; the background processing module configured to: generate a hierarchical data structure, the hierarchical data structure comprising at least two topical nodes, wherein the at least two topical nodes represent partitions organized by two or more topical themes among the topical content of the training data within which the training data is to be subdivided into; select among the training data a plurality of documents to be annotated; generate at least one annotation prompt for each document among the plurality of documents to be annotated, said annotation prompt configured to elicit an annotation about said document indicating which node among the at least two topical nodes of the hierarchal data structure said document is to be classified into; the application module configured to cause display of the at least one annotation prompt for each document among the plurality of documents to be annotated; the API module further configured to receive for each document among the plurality of documents to be annotated, the annotation in response to the displayed annotation prompt; and the background processing module further configured to generate the natural language model using an adaptive machine learning process configured to determine, among the received annotations, patterns for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
  • In some embodiments, the background processing module is further configured to test performance of the natural language model using a subset of the documents among the training data that received annotations. In some embodiments, the background processing module is further configured to: compute a performance metric of the natural language model, based on results of the testing; and determine whether the natural language model satisfies at least one performance criterion based on the computed performance metric. In some embodiments, the background processing module is further configured to: performing one or more optimization techniques configured to improve performance of the natural language platform, in response to determining that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric. In some embodiments, the one or more optimization techniques comprises at least one of: a feature selection process, a padding and rebalancing process of the natural language model, a pruning process of the natural language model, a feature discovery process, a smoothing process of the natural language model, or a model interpolation process.
  • In some embodiments, the background processing module is further configured to: determine that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric; in response to said determining: identify a topical node among the two or more topical nodes of the hierarchal data structure that the natural language model fails to accurately categorize documents into; select a second plurality of documents to be annotated, the second plurality comprising documents associated with said topical node that the natural language model failed to accurately categorize documents into; and generate a second set of at least one annotation prompt for each document among the second plurality of documents to be annotated, said annotation prompt among the second set configured to elicit an annotation about said document to improve the natural language model in accurately categorizing documents into said topical node; wherein the applications module is further configured to cause display of the second set of the at least one annotation prompt for each document among the second plurality of documents to be annotated; wherein the API module is further configured to receive for each document among the second plurality of documents to be annotated, a second set of annotations in response to the second set of displayed annotation prompts; and wherein the background processing module is further configured to generate a refined natural language model using the adaptive machine learning process and based on the hierarchical data structure, the training data and the second set of annotations.
  • In some embodiments, generating the hierarchical data structure comprises: performing a topic modeling process configured to identify two or more topics among the content of the training data that is configured to define the two or more topical nodes of the hierarchical data structure.
  • In some embodiments, the background processing module is further configured to access one or more rules configured to instruct the natural language model how to categorize one or more documents into the two or more nodes of the hierarchical data structure. In some embodiments, generating the hierarchical data structure comprises: conducting a rules generation process configured to evaluate logical consistency among the one or more rules.
  • In some embodiments, generating the hierarchical data structure comprises: generating at least one annotation prompt for each topical node among the two or more topical nodes in the hierarchical data structure, said annotation prompt configured to elicit an annotation about said topical node indicating a level of accuracy of placement of the node within the hierarchical data structure; causing display of the at least one annotation prompt for each topical node; and receiving for each topical node, the annotation in response to the displayed annotation prompt. In some embodiments, the background processing module is further configured to evaluate performance of the hierarchical data structure based on the annotations. In some embodiments, the background processing module is further configured to determine that the hierarchical data structure fails to satisfy at least one performance criterion in response to the evaluating; and modify a logical relationship among the two or more topical nodes based on the annotations and in response to determining that the data structure fails to satisfy the at least one performance criterion.
  • In some embodiments, the API module is further configured to receive a training guideline based on the annotations to the nodes, the training guideline configured to provide instructions to an annotator for answering one or more annotation prompts for each document among the plurality of documents to be annotated.
  • In some embodiments, the hierarchical data structure comprises at least a third topical node and a fourth topical node, wherein the third and fourth topical nodes both represent sub-partitions within the topical theme of the first node and organized by a third and fourth topical theme, respectively, among the topical content of the training data within which the training data is to be subdivided into.
  • In some embodiments, a method for generating a natural language platform system configured to generate a natural language model is presented. The method includes: deriving from an analogous natural language platform system, parameters configured to optimize performance of said analogous system to generate an analogous natural language model configured to analyze similar but not identical documents as the natural language model; interpolating said parameters to be optimized for documents to be analyzed by the natural language model; and implementing the interpolated parameters in the natural language platform system such that the interpolated parameters are configured to generate the natural language model.
  • In some embodiments, another system for generating natural language models is presented. This system may include: a full service natural language platform geographically located at a remote host location and configured to: receive training data through a network connection; train a natural language model based on the received training data; and generate predictions about untested data using the natural language model; and a connector module geographically located at a local client host location and communicatively coupled to the full service natural language platform through the network connection and configured to: access the training data stored in a client data store at the local client host location; format the training data in a uniform manner; transmit the training data to the full service natural language platform through the network connection; receive the predictions about the untested data; and store the predictions about the untested data in a memory at the local client host location.
  • In some embodiments, the system further includes a text extraction module communicatively coupled to the connector module and configured to attach to the client data store and extract textual data for use as the training data.
  • In some embodiments, the full service natural language platform is further configured to: generate a hierarchical data structure, the hierarchical data structure comprising at least two topical nodes, wherein the at least two topical nodes represent partitions organized by two or more topical themes among the topical content of the training data within which the training data is to be subdivided into; select among the training data a plurality of documents to be annotated; and generate at least one annotation prompt for each document among the plurality of documents to be annotated, said annotation prompt configured to elicit an annotation about said document indicating which node among the at least two topical nodes of the hierarchal data structure said document is to be classified into. In some embodiments, the full service natural language platform is further configured to: cause display of the at least one annotation prompt for each document among the plurality of documents to be annotated; receive for each document among the plurality of documents to be annotated, the annotation in response to the displayed annotation prompt; and generate the natural language model using an adaptive machine learning process configured to determine, among the received annotations, patterns for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
  • In some embodiments, the full service natural language platform is further configured to: test performance of the natural language model using a subset of the documents among the training data that received annotations; compute a performance metric of the natural language model, based on results of the testing; determine that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric; in response to said determining: identify a topical node among the two or more topical nodes of the hierarchal data structure that the natural language model fails to accurately categorize documents into; select a second plurality of documents to be annotated, the second plurality comprising documents associated with said topical node that the natural language model failed to accurately categorize documents into; generate a second set of at least one annotation prompt for each document among the second plurality of documents to be annotated, said annotation prompt among the second set configured to elicit an annotation about said document to improve the natural language model in accurately categorizing documents into said topical node; cause display of the second set of the at least one annotation prompt for each document among the second plurality of documents to be annotated; receive for each document among the second plurality of documents to be annotated, a second set of annotations in response to the second set of displayed annotation prompts; and generate a refined natural language model using the adaptive machine learning process and based on the hierarchical data structure, the training data and the second set of annotations.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Some embodiments are illustrated by way of example and not limitation in the figures of the accompanying drawings.
  • FIG. 1A is a network diagram illustrating an example network environment suitable for aspects of the present disclosure, according to some example embodiments.
  • FIG. 1B is diagram providing additional examples of networked environments for generating natural language models, according to some embodiments.
  • FIG. 2A is a diagram showing an example system architecture for performing aspects of the present disclosure, according to some example embodiments.
  • FIG. 2B is a diagram of example flowchart for how the system architecture according to the diagram in FIG. 2A may be utilized to generate other natural language platforms that are more tailored to a client's specific needs, according to some embodiments.
  • FIG. 3 is a high level diagram showing various examples of types of human communications and what the objectives may be for a natural language model to accomplish, according to some embodiments.
  • FIG. 4 is a diagram showing an example flowchart for how different data structures within the system architecture may be related to one another, according to some example embodiments.
  • FIG. 5 is a flowchart of a high-level process for generating a natural language model, according to some embodiments.
  • FIG. 6 is a diagram of a more detailed view of the block 505 of FIG. 5, describing the ingest data phase, according to some embodiments.
  • FIG. 7 shows a simple example of an ontology.
  • FIG. 8 shows a more complex example of an ontology.
  • FIGS. 9A and 9B show an even more complex example of an ontology.
  • FIGS. 10A-10C provide additional example details for generating the ontology within the block 510 of FIG. 5, according to some embodiments.
  • FIGS. 11A and 11B show example displays for annotating as part of the adaptive machine learning process, according to some embodiments.
  • FIG. 12 provides a more detailed process flow of the block 515 of FIG. 5, for conducting the adaptive machine learning process using the annotations to generate the natural language model, according to some embodiments.
  • FIG. 13 provides further details of different examples of optimization techniques that may be applied in block 1240 of FIG. 12 for performing optimization techniques to improve the natural language model, according to some embodiments.
  • FIG. 14 illustrates a modified process for generating natural language models, including iterating processes between some processes within the create ontology block and the conduct adaptive machine learning block of FIG. 5, according to some embodiments.
  • FIG. 15 provides an additional variant for generating natural language models, this time including a maintenance process for updating currently operating natural language models, according to some embodiments.
  • FIG. 16 is a block diagram illustrating components of a machine, according to some example embodiments, able to read instructions from a machine-readable medium and perform any one or more of the methodologies discussed herein.
  • DETAILED DESCRIPTION
  • Example methods, apparatuses, and systems (e.g., machines) are presented for generating natural language models.
  • The modes of human communications brought upon by digital technologies have created a deluge of information that can be difficult for human readers to handle alone. Companies and research groups may want to determine trends in the human communications to determine what people generally care about for any particular topic, whether it be what car features are being most expressed on Twitter®, what political topics are being most expressed on Facebook®, what people are saying about the customer's latest product in their customer feedback page, what are the key categories written about in a large body of legal documents, and so forth. In some cases, the companies or the research groups may want to determine what are the general topics being talked about, to begin with. It may be desirable for companies to aggregate and then synthesize the thousands or even millions of human communications from the many different modes available in the digital age (e.g., Twitter®, transcribed speech, email, OCR-ed text, etc.) to determine these objectives. Processing all this information by humans alone can be overwhelming and cost-inefficient. Methods today may therefore rely on computers to apply natural language processing in order to analyze the many human communications available in order to filter, organize, interpret, categorize, ask questions, respond to questions, and extract information from the many human communications into digestible patterns of communication.
  • Conventionally, natural language processing may involve humans declaring a large number of rules for the natural language model to follow. For example, when trying to determine how many communications convey negative sentiment about a politician, a rule may be generated to search for “horrible” and “Bill Clinton” in the same document. Similarly, a rule looking for the terms “wonderful” and “Bill Clinton” may attempt to find all the communications conveying positive sentiment about a politician. Conventionally, the natural language model may then parse a given number of documents, follow the rules, and attempt to categorize the documents by the rules. However, it can become clear that the number of rules necessary to capture all forms of negative and positive sentiment may be immeasurable, as there alone are simply dozens of synonyms for the words “bad” and “good,” not to mention all the slang terms, and sentiment conveyed by a combination of words that may not be captured by any singular rule. Generally, performing natural language processing through a rules-only approach oftentimes fails to adequately and reliably categorize human communications in a meaningful way.
  • Alternately, natural language processing may involve applying machine-learning techniques to a set of example documents that have been categorized by human annotators, generating statistical inferences for the natural language model to follow. For example, a collection of documents written about “Bill Clinton” may include a subset of those documents understood and categorized by a human as conveying negative sentiment, and a second subset of documents understood and categorized as conveying positive sentiment. In this way, the natural language models generated through machine learning approaches are able to recognize many more characteristics of sentiment-conveying text than can be reasonably implemented through rules-only approaches. Accordingly, the accuracy of such models at categorizing human communications may be substantially higher than rules-based natural language processing approaches. However, conventionally, the use of machine learning for performing natural language processing requires human annotators to label hundreds or even thousands of example documents before the natural language models achieve a satisfactory level of accuracy. Therefore, it is desirable to use more effective natural language processing techniques to rapidly develop natural language models that efficiently and accurately summarize the thousands or millions of human communications.
  • Aspects of the present disclosure are presented for generating natural language models (also referred to herein as prediction engines) through a combination of processing human annotations of documents and adaptive machine learning techniques that determine patterns in the documents, based in part on the human annotations applied to the documents. The natural language platform used to generate the models, such as a software system stored in memory of one or more servers communicatively coupled in parallel, may be configured to dynamically generate human readable annotation prompts that may be expressed in varying levels of detail and granularity so as to efficiently obtain specific annotation information and maximize use of a human annotator's time. An iterative or cyclic process, involving cycling between supplying prompts for human annotations to documents and obtaining the annotations, then generating the natural language model using machine learning of the documents with the annotations, then evaluating the model and iterating back to obtaining more human annotations to refine the model, and so forth, is also presented.
  • As an example, the human annotations may assist the machine learning techniques to resolve inevitable ambiguities in the human communications, as well as provide intelligence or meaning to communications that the machine does not accurately comprehend a priori. The human annotations can then enable a natural language platform configured to generate these natural language models to provide better natural language results of the human communications, which can then in turn be better refined with the assistance of more human annotations as necessary. For example, a human annotator may provide an annotation about a Tweet reading “Can you believe this guy?!” as expressing negative sentiment of the person being referenced, while a rule may fail to flag this Tweet as having negative sentiment because no typical negative words are contained in the Tweet.
  • In some embodiments, the natural language platform used to generate the natural language models may be configured to generate human readable annotation prompts at varying degrees of detail, so as to elicit responses from human annotators that match more specifically a level of ambiguity the natural language platform is trying to resolve. For example, a first level of annotation prompts may include determining what subject category (or categories) a document may be categorized into, among a plurality of first categories. This first level may therefore generate a prompt to the human annotator that includes a multiple choice question including three or more categories that the human annotator may choose from, where the human annotator may be prompted to select among any or all of the categories. As another example, a second level of annotation prompts may include determining what subject category a document may be best categorized into, among a plurality of second categories. This second level may therefore generate a prompt displayable to the human annotator that includes a multiple-choice question including three or more categories that the human annotator may choose from, where the human annotator may be prompted to select only one category that best describes the document. As another example, a third level of annotation prompts may include supplying a binary decision to the human annotator, such as asking whether it fits into a category or not (e.g., a yes or no or a true or false question).
  • Depending on the amount and precision of information that the natural language model has already obtained, the natural language platform may dynamically adjust the type of annotation prompts to supply to human annotators. For example, if the natural language model possesses little information about a document or a category of documents, the natural language platform may be configured to supply a more open ended question to human annotators, such as an annotation prompt at the first level. On the other hand, if the natural language model possesses precise information about a document or a category of documents, the natural language platform may be configured to supply a more precise question to the human annotators, such as an annotation prompt at the third level. In this way, information to improve the natural language model may be more efficiently obtained and may maximize the use of human annotators' time.
  • In addition, aspects of the present disclosure may construct natural language models based on this iterative process that can be specifically tailored to a customer's unique needs or subject matter area. For example, the words important to categorizing communications in biotechnology may be different than the words important to categorizing communications in the automobile industry. The biotechnology user may desire to tailor a natural language model to better understand articles related to biotechnology, while the automobile industry user may desire to tailor the same or a different natural language model to better understand customer feedback emails. As another example, the language, grammar, and idioms used in social media may vary drastically from communications in professional writings, e.g., legal or medical journals. A user focusing on Twitter® communications may desire to tailor the natural language model to better determine when tweets of adolescent teens convey positive sentiment or negative sentiment, while a user focusing on legal documents may desire to tailor a natural language model to better understand whether a legal decision is favorable or unfavorable, and to see how many scholarly articles are written about the legal decision and what percentage of them are critical of the legal decision. Aspects of the present disclosure therefore are robust enough to generate a natural language model to interpret documents in any language (or languages) and for any number of topics or subject areas contained within the documents.
  • The natural language generation platform may be configured to process the human annotations to any of these areas in order to better inform the natural language model how to understand these complex communications. The natural language model can be trained through the iterative process utilizing human annotations to more easily determine how to categorize any of these diverse areas of human communications.
  • In some embodiments, the various categories or topics, sometimes referred to herein as labels, that the documents are intended to be grouped into may be organized by a hierarchical data structure, referred to herein as an ontology. The hierarchical structure design of the ontology may define a plurality of subcategories, or sub labels, that thematically or topically fit within a larger umbrella label. For example, a set of documents about customer service emails of a telecommunications company may be categorized into an ontology where a first label of the ontology may be simply “customer service.” Under this first label, multiple sub labels may be included that thematically fall within the broader label of customer service, such as “Internet,” “phone,” and “customer service experience.” There may be further sub labels, such as “Check the router power,” which could be specific responses to questions in the customer service exchange, representing a single intention type of the person emailing the customer service representative: in this example, the intention to get solutions for a broken internet connection. In general, an ontology may include multiple levels of sub labels under higher levels of sub labels, and so forth. Similarly, an ontology may include multiple labels which define a plurality of common subcategories, that thematically or topically fit within each larger umbrella. In its totality, the ontology can represent the multiple label, information extracted, and sequence of responses to questions, for an entire business unit and organization.
  • In some embodiments, these labels of the ontology may define and provide structure for the categorization of millions of documents by the natural language model. That is, developing the natural language model will include a goal of training the natural language model to accurately classify each of the documents in a set of documents into one or more of the labels in the hierarchical data structure. In some embodiments, a process for generating the natural language model includes a subprocess for first generating this ontology. Once generated, the ontology may be used to guide the aforementioned iterative process of obtaining human annotations and applying them in adaptive machine learning processes to generate a natural language model.
  • Once tuned to a user's specific needs through the aforementioned iterative process, the natural language model can then be applied to untested human communications, and may generate an output that summarizes which human communications and how many are categorized into the various labels in the ontology structure (e.g., positive sentiment of a politician, negative sentiment of a politician). The process conducted through the natural language model for classifying the thousands or even millions of documents into the specified topics (also referred to herein as labels) may be referred to herein as a document-scope task.
  • In some embodiments, a natural language model of the present disclosures may also be configured to extract specific subsets of text from documents (e.g., extract every phrase mentioning an anticipated cost or savings in a document discussing tax reform, or extract all words naming the crew in one or more documents relating to the latest blockbuster movie). This process conducted through the natural language model for extracting specific types of text from one or more documents may be referred to herein as a span-scope task. In general, as used herein, a span may refer to a subset of words or characters in a document, such as a paragraph, multiple paragraphs, sentences, or a plurality of words. Performing a span scope task may also include first generating an ontology, and then conducting the iterative human annotation and adaptive machine learning process, in some embodiments.
  • In some embodiments, a natural language platform of the present disclosures may also be configured to simply identify or discover what categories or subject areas (e.g., labels) the documents may be thematically organized into. For example, it may not be known what are the top twenty subject areas being most discussed in all tweets within the first week of September 2015. Before a natural language model can be generated to summarize what sentiment is being conveyed about these top twenty most discussed subject areas, it should be known what kinds of subject matters these tweets generally discuss in the first place. The process for discovering what topics or subject areas the documents may be thematically organized into may be referred to as topic modeling.
  • In some embodiments, a natural language platform of the present disclosures may also be configured to utilize rules in combination with the iterative human annotation/machine learning process. In some embodiments, the rules may be imported from an existing body specified by a user or customer. In addition, the natural language model may be generated with the aid of any pre-existing conditions specified by the user, such as specified topics or rules, and embodiments are not so limited.
  • In some embodiments, a natural language model may be generated using previously generated natural language models that were tailored to different but similar subject areas. In addition, the natural language model may be refined or updated through a maintenance process so that the model may be better suited to handle new data with an evolving body of words. For example, a natural language model for processing tweets may be refined over time to account for new slang or idioms that were not previously used in the common vernacular.
  • Examples merely demonstrate possible variations. Unless explicitly stated otherwise, components and functions are optional and may be combined or subdivided, and operations may vary in sequence or be combined or subdivided. In the following description, for purposes of explanation, numerous specific details are set forth to provide a thorough understanding of example embodiments. It will be evident to one skilled in the art, however, that the present subject matter may be practiced without these specific details.
  • Referring to FIG. 1A, a network diagram illustrating an example network environment 100 suitable for performing aspects of the present disclosure is shown, according to some example embodiments. The example network environment 100 includes a server machine 110, a database 115, a first device 120 for a first user 122, and a second device 130 for a second user 132, all communicatively coupled to each other via a network 190. The server machine 110 may form all or part of a network-based system 105 (e.g., a cloud-based server system configured to provide one or more services to the first and second devices 120 and 130). The server machine 110, the first device 120, and the second device 130 may each be implemented in a computer system, in whole or in part, as described below with respect to FIG. 16. The network-based system 105 may be an example of a natural language platform configured to generate natural language models as described herein. The server machine 110 and the database 115 may be components of the natural language platform configured to perform these functions. While the server machine 110 is represented as just a single machine and the database 115 where is represented as just a single database, in some embodiments, multiple server machines and multiple databases communicatively coupled in parallel or in serial may be utilized, and embodiments are not so limited.
  • Also shown in FIG. 1A are a first user 122 and a second user 132. One or both of the first and second users 122 and 132 may be a human user, a machine user (e.g., a computer configured by a software program to interact with the first device 120), or any suitable combination thereof (e.g., a human assisted by a machine or a machine supervised by a human). The first user 122 may be associated with the first device 120 and may be a user of the first device 120. For example, the first device 120 may be a desktop computer, a vehicle computer, a tablet computer, a navigational device, a portable media device, a smartphone, or a wearable device (e.g., a smart watch or smart glasses) belonging to the first user 122. Likewise, the second user 132 may be associated with the second device 130. As an example, the second device 130 may be a desktop computer, a vehicle computer, a tablet computer, a navigational device, a portable media device, a, smartphone, or a wearable device (e.g., a smart watch or smart glasses) belonging to the second user 132. The first user 122 and a second user 132 may be examples of users or customers interfacing with the network-based system 105 to utilize a natural language model according to their specific needs. In other cases, the users 122 and 132 may be examples of annotators who are supplying annotations to documents to be used for training purposes when developing a natural language model. In other cases, the users 122 and 132 may be examples of analysts who are providing inputs to the natural language platform to more efficiently train the natural language model. The users 122 and 132 may interface with the network-based system 105 through the devices 120 and 130, respectively.
  • Any of the machines, databases 115, or first or second devices 120 or 130 shown in FIG. 1A may be implemented in a general-purpose computer modified (e.g., configured or programmed) by software (e.g., one or more software modules) to be a special-purpose computer to perform one or more of the functions described herein for that machine, database 115, or first or second device 120 or 130. For example, a computer system able to implement any one or more of the methodologies described herein is discussed below with respect to FIG. 16. As used herein, a “database” may refer to a data storage resource and may store data structured as a text file, a table, a spreadsheet, a relational database (e.g., an object-relational database), a triple store, a hierarchical data store, any other suitable means for organizing and storing data or any suitable combination thereof. Moreover, any two or more of the machines, databases, or devices illustrated in FIG. 1A may be combined into a single machine, and the functions described herein for any single machine, database, or device may be subdivided among multiple machines, databases, or devices.
  • The network 190 may be any network that enables communication between or among machines, databases 115, and devices (e.g., the server machine 110 and the first device 120). Accordingly, the network 190 may be a wired network, a wireless network (e.g., a mobile or cellular network), or any suitable combination thereof. The network 190 may include one or more portions that constitute a private network, a public network (e.g., the Internet), or any suitable combination thereof. Accordingly, the network 190 may include, for example, one or more portions that incorporate a local area network (LAN), a wide area network (WAN), the Internet, a mobile telephone network (e.g., a cellular network), a wired telephone network (e.g., a plain old telephone system (POTS) network), a wireless data network (e.g., WiFi network or WiMax network), or any suitable combination thereof. Any one or more portions of the network 190 may communicate information via a transmission medium. As used herein, “transmission medium” may refer to any intangible (e.g., transitory) medium that is capable of communicating (e.g., transmitting) instructions for execution by a machine (e.g., by one or more processors of such a machine), and can include digital or analog communication signals or other intangible media to facilitate communication of such software.
  • Referring to FIG. 1B, diagram 150 provides additional examples of networked environments for generating natural language models, according to some embodiments. Diagram 150 provides a block diagram view of various configurations of a natural language platform configured to generate natural language models may interact with various data structures of a client, such as a user 120 or 130. In diagram 150, three options are illustrated. According to any of the three options, a flow of data transfer for generating natural language models begins at a client data store 155. The client data store 155 may contain a repository of data that is used to train and ultimately generate the natural language model. In some embodiments, a text extraction module 157 may be installed or communicatively coupled to the client data store 155 and may be configured to extract textual data. For example, the client data store 155 may contain scanned copies of articles or printed text. The text extraction module 157 may be configured to convert the scanned copies into computer readable text. The textual data may be transmitted to a connector 160, which may include software and hardware configured to act as a bridge that interfaces between the client data store 155 and one or more natural language platforms.
  • According to Option #1, from the connector 160, the textual data may be transmitted to a full-service natural language platform 170 through the network 190. The full-service natural language platform 170 may be located geographically distinct from the client, such as in a cloud environment. The full-service natural language platform 170 may be configured to provide all available services for generating natural language models described herein, the various services of which will be described more fully below. At the least, the full-service natural language platform 170 may be configured to generate a natural language model for the client by training the model on the textual data derived from the client data store 155. As shown, this process may occur in the cloud environment. In addition, the full-service natural language platform 170 may also be configured to apply the trained natural language model on untested, “live” data, referred to herein as providing predictions on the untrained and untested data. The untrained or untested data may also be obtained from the client, such as from the client data store 155. As shown, this process also may occur in the cloud environment. The results of these predictions may be transmitted back to the client, through the network 190, bridged by the connector 160, and ultimately stored in a results repository 185.
  • According to Option #2, from the connector 160, the textual data may be transmitted to a client-controlled natural language platform 175. The client control platform 175 may be stored locally and within the exclusive control of the client, according to some embodiments. In this option, the client-controlled platform 175 may be configured to deploy a pre-generated natural language model, meaning the natural language model was already trained with previous data. The arrow from the connector 160 to the client-controlled platform 175 lists only “prediction,” meaning that the ability to train a natural language model is not available in this option. However, as shown, the client is able to utilize the pre-generated natural language model to make predictions on the untrained and untested data. These results are supplied to the connector 160, and ultimately stored in the results repository 185. According to this option, the client may have exclusive control of the natural language model, and is not required to release its data to an off-site platform in order to utilize any natural language model. Thus, Option #2 allows for greater privacy for the client.
  • According to Option #3, from the controller 160, the textual data may be transmitted to a client controlled natural language platform 180 and that provides full services. In this option, the full-service natural language platform 170 may be duplicated locally at the client site. This option may be more costly to the client, but may be more comprehensive. The services provided through the cloud 190 at the full-service natural language platform 170 may be substantially identical at the local full-service natural language platform 180, according to some embodiments.
  • Referring to FIG. 2A, a diagram 200 is presented showing an example system architecture for performing aspects of the present disclosure, according to some example embodiments. The example system architecture according to diagram 200 represents various data structures and their interrelationships that may comprise a natural language platform, such as the natural language platform 170, or the network-based system 105. These various data structures may be implemented through a combination of hardware and software, the details of which may be apparent to those with skill in the art based on the descriptions of the various data structures described herein. For example, an API module 205 includes one or more API processors, where multiple API processors may be connected in parallel. In some example embodiments, the repeating boxes in the diagram 200 represent identical servers or machines, to signify that the system architecture in diagram 200 may be scalable to an arbitrary degree. The API module 205 may represent a point of contact for multiple other modules, includes a database module 210, a cache module 215, background processes module 220, applications module 225, and even an interface for users 235 in some example embodiments. The API module 205 may be configured to receive or access data from database module 210. The data may include digital forms of thousands or millions of human communications. The cache module 215 may store in more accessible memory various information from the database module 210 or from users 235 or other subscribers. Because the database module 210 and cache module 215 show accessibility through API module 205, the API module 205 can also support authentication and authorization of the data in these modules. The background module 220 may be configured to perform a number of background processes for aiding natural language processing functionality. Various examples of the background processes include a model training module, a cross validation module, an intelligent queuing module, a model prediction module, a topic modeling module, an annotation aggregation module, an annotation validation module. These various modules are described in more detail below as well as in non-provisional applications (Attorney Docket Nos. 1402805.00012_IDB012, 1402805.00013_IDB013, 1402805.00014_IDB014, 1402805.00016_IDB016, 1402805.00017_IDB017, and 1402805.00019_IDB019), each of which are again incorporated by reference in their entireties. The API module 205 may also be configured to support display and functionality of one or more applications in applications module 225.
  • In some example embodiments, the users 235 may access the API module 205, in some cases enabling the users 235 to create their own applications using the system architecture of diagram 200. The users 235 may be other examples of the users 120 or 130, and may also include project managers and analysts. Project managers may utilize the natural language platform to direct the overall construction of one or more natural language models. Analysts may utilize the natural language platform to provide expert analysis and annotations to more efficiently train a natural language model. Also, annotators 230 may have access to applications already created in applications module 225. Further discussion about the annotators is found more below.
  • In some embodiments, the system architecture according to diagram 200 may be scalable and reproducible at various client sites, in some cases consistent with the descriptions of FIG. 1B. Thus, the database modules 210 and the cache module 215 may be implemented specifically for each client, such that each client does not share memory capacity with another client to ensure better privacy.
  • Referring to FIG. 2B, diagram 250 provides an example flowchart for how the system architecture according to diagram 200 may be utilized to generate other natural language platforms that are more tailored to a client's specific needs, according to some embodiments. For example, a pre-existing natural language platform according to the diagram 200 may have already been constructed at the client site. The various API modules 205, database modules 210, cache modules 215, background modules 220, and display modules 225 of the pre-existing natural language platform may have been constructed and tailored to specific client needs and/or specific subject matter areas. For example, the pre-existing natural language platform may have been constructed to generate one or more natural language models related to comments on Twitter about baseball. When generating a new natural language platform for similar subject matter, it may be desirable to start with the pre-existing natural language platform. For example, it may be desirable to base a new natural language platform for a client that desires to process text from Twitter related to all professional sports on the pre-existing natural language platform configured to generate natural language models about baseball.
  • In this respect, still referring to diagram 250, in some embodiments, an interpolation operation 255 may be applied to the pre-existing natural language platform. The interpolation operation 255 may include copying various modules, data and labels from the pre-existing natural language platform over to the new natural language platform. For example, the various background modules 220 may be calibrated to more efficiently handle text from Twitter, knowing the general size of tweets and having various libraries for understanding contemporary vernacular in tweets. The cache sizes and other various settings in the cache modules 215 may also be attuned to efficiently handle tweets, as another example. The same settings may be applied to the new natural language platform as a starting point. At block 260, the new natural language platform may be generated, and may include various adjustments from the pre-existing natural language platform. As yet another example, the server computation resources assigned to the various background modules 220 may be adjusted in relation to the API modules 205, so that more computation resources are available for intelligent queuing.
  • In some embodiments, the natural language platform may be implemented by partitioning each of the API modules 205, application modules 225, background processing modules 220, cache modules 215 and database modules 210 to be implemented within individual containers. In such embodiments, platform interpolation may be implemented by adjusting the proportion of processing resources (e.g., memory, network, storage and computation) assigned to each such container. Other components that may be interpolated from pre-existing and analogous platforms may include computational resources (CPU performance, memory) reserved for the background processing modules relative to the API module. If the interpolated system will be used primarily for prediction, rather than for topic modeling or re-training, then a greater share of CPU resources may be allocated to the API module. Another example includes selecting an optimized default text encoding for the language(s) used in the interpolated systems. For example, if the interpolated system will only process Western European characters (no emoji, Cyrillic, or East Asian languages, et al.), then using ISO-8859-1 instead of UTF-8 can improve performance and reduce storage cost.
  • Referring to FIG. 3, a high level diagram 300 is presented showing various examples of types of human communications and what the objectives may be for a natural language model to accomplish. That is, while FIGS. 1A-2B have described the various structures of a natural language platform configured to generate one or more natural language models, FIG. 3 and the following figures will now discuss the various example functional objectives to be accomplished and various example technical processes used to achieve these functional objectives. Here, various sources of data, sometimes referred to as a collection of documents 305, may be obtained and stored in, for example database 115, client data store 155, or database modules 210, and may represent different types of human communications, all capable of being analyzed by a natural language model. Examples of the types of documents 305 include, but are not limited to, posts in social media, emails or other writings for customer feedback, pieces of or whole journalistic articles, commands spoken or written to electronic devices, and pieces of or whole scholarly texts. Other types include transcribed call center recordings, electronic (instant) messages, corporate communications (e.g., SEC 10-k, 10-q), confidential documents and communications stored on internal collaboration systems (e.g., SharePoint, Notes), and questions and answers, either in isolation or as part of longer dialogues.
  • In some embodiments, at block 310, it may be desired to classify any of the documents 305 into a number of enumerated categories or topics, consistent with some of the descriptions mentioned above. This may be referred to as performing a document-scope task. For example, a user 130 in telecommunications may supply thousands of customer service emails related to services provided by a telecommunications company. The user 130 may desire to have a natural language model generated that classifies the emails into predetermined categories, such as negative sentiment about their Internet service, positive sentiment about their Internet service, negative sentiment about their cable service, and positive sentiment about their cable service. As previously mentioned, these various categories for which a natural language model may classify the emails into, e.g. “negative” sentiment about “Internet service,” “positive” sentiment about “Internet service,” “negative” sentiment about “cable service,” etc., may be referred to as “labels.” Based on these objectives, at block 315, a natural language model may be generated that is tailored to classify these types of emails into these types of labels.
  • As another example, in some embodiments, at block 320, it may be desired to extract specific subsets of text from documents, consistent with some of the descriptions mentioned above. This may be another example of performing a span-scope task, in reference to the fact that this function focuses on a subset within each document (as previously mentioned, referred to herein as a “span”). For example, a user 130 may desire to identify all instances of a keyword, key phrase, or general subject matter within a novel. Certainly, this span scope task may be applied to multiple novels or other documents. Here too, based on this objective, at block 315, a natural language model may be generated that is tailored to perform this function for a specified number of documents.
  • As another example, in some embodiments, at block 325, it may be desired to discover what categories the documents may be thematically or topically organized into in the first place, consistent with descriptions above about topic modeling. In some cases, the user 130 may utilize the natural language platform only to perform topic modeling and to discover what topics are most discussed in a specified collection of documents 305. To this end, the natural language platform may be configured to conduct topic modeling analysis at block 330. Topic modeling is discussed in more detail below, as well as in applications (Attorney Docket Nos. 1402805.00012_IDB012, 1402805.00013_IDB013, 1402805.00015_IDB015, and 1402805.00019_IDB019), each of which are again incorporated herein by reference in their entireties. In some cases, it may be desired to then generate a natural language model that categorizes the documents 305 into these newfound topics. Thus, after performing the topic modeling analysis 230, in some embodiments, the natural language model may also be generated at block 315.
  • Referring to FIG. 4, a diagram 400 is presented showing an example flowchart for how different data structures within the system architecture may be related to one another, according to some example embodiments. Here, the collections data structure 410 represents a set of documents 435 that in some cases may generally be homogenous. A document 435 represents a human communication expressed in a single discrete package, such as a single tweet, a webpage, a chapter of a book, a command to a device, or a journal article, or any part thereof. Each collection 410 may have one or more tasks 430 associated with it. A task 430 may be thought of as a classification scheme. For example, a collection 410 of tweets may be classified by its sentiment, e.g. a positive sentiment or a negative sentiment, where each classification constitutes a task 430 about a collection 410. A label 445 refers to a specific prediction about a specific classification. For example, a label 445 may be the “positive sentiment” of a human communication, or the “negative sentiment” of a human communication. In some cases, labels 445 can be applied to merely portions of documents 435, such as paragraphs in an article or particular names or places mentioned in a document 435. For example, a label 445 may be a “positive opinion” expressed about a product mentioned in a human communication, or a “negative opinion” expressed about a product mentioned in a human communication. In some example embodiments, a task may be a sub-task of another task, allowing for a hierarchy or complex network of tasks. For example, if a task has a label of “positive opinion,” there might be sub-tasks for types of “positives opinions,” like “intention to purchase the product,” “positive review,” “recommendation to friend,” and so on, and there may be subtasks that capture other relevant information, such as “positive features.”
  • Annotations 440 refer to classifications imputed onto a collection 410 or a document 435 by an external source, such as human input, interpolation from an existing natural language model, or application of an external transformation process on the metadata included with the document (e.g., customer value, geographic location, etc.). As an example, an annotation 440 applies a label 445 manually to a document 435. In other cases, annotations 440 are provided by users 235 from pre-existing data. In other cases, annotations 440 may be derived from human critiques of one or more documents 435, where the computer determines what annotation 440 should be placed on a document 435 (or collection 410) based on the human critique. In other cases, with enough data in a language model, annotations 440 of a collection 410 can be derived from one or more patterns of pre-existing annotations found in the collection 410 or a similar collection 410.
  • In some example embodiments, features 450 refer to a library or collection of certain key words or groups of words that may be used to determine whether a task 430 should be associated with a collection 410 or document 435. Thus, each task 430 has associated with it one or more features 450 that help define the task 430. In some example embodiments, features 450 can also include a length of words or other linguistic descriptions about the language structure of a document 435, in order to define the task 430. For example, classifying a document 435 as being a legal document may be based on determining if the document 435 contains a threshold number of words with particularly long lengths, words belonging to a pre-defined dictionary of legal-terms, or words that are related through syntactic structures and semantic relationships. In some example embodiments, features 450 are defined by code, while in other cases features 450 are discovered by statistical methods. In some example embodiments, features 450 are treated independently, while in other cases features 450 are networked combinations of simpler features that are used in combination utilizing techniques like “deep-learning” or “topic modeling.” In some example embodiments, combinations of the methods described herein may be used to define the features 450, and embodiments are not so limited. One or more processors may be used to identify in a document 435 the words found in features data structure 450 to determine what task should be associated with the document 435.
  • In some example embodiments, a work unit's data structure 455 specifies when humans should be tasked to further examine a document 425. Thus, human annotations may be applied to a document 435 after one or more work units 455 is applied to the document 435. The work units 455 may specify how many human annotators should examine the document 435 and in what order of documents should document 435 be examined. In some example embodiments, work units 455 may also determine what annotations should be reviewed in a particular document 435 and what the optimal user interface should be for review.
  • In some example embodiments, the data structures 405, 415, 420 and 425 represent data groupings related to user authentication and user access to data in system architecture. For example, the subscribers block 405 may represent users and associated identification information about the users. The subscribers 405 may have associated API keys 415, which may represent one or more authentication data structures used to authenticate subscribers and provide access to the collections 410. Groups 420 may represent a grouping of subscribers based on one or more common traits, such as subscribers 405 belonging to the same company. Individual users 425 capable of accessing the collections 410 may also result from one or more groups 420. In addition, in some cases, each group 420, user 425, or subscriber 405 may have associated with it a more personalized or customized set of collections 510, documents 435, annotations 440, tasks, 430, features 450, and labels 445, based on the specific needs of the customer.
  • Referring to FIG. 5, flowchart 500 provides a high-level process for generating a natural language model, according to some embodiments. The high-level process described in FIG. 5 may be implemented by a natural language platform, various examples of which have been described herein. Starting at block 505, a natural language platform may be configured to ingest training data representative of documents that will ultimately be analyzed by the natural language model once the natural language model is fully trained. Examples of training data may include exactly the types of documents that are intended to be analyzed by the natural language model, including any of the example collections of documents 305. In addition, subsets of documents may be used for training data, such as paragraphs, excerpts, chapters, or articles of larger bodies of text. The training data may be obtained from a number of different sources, including for example, from databases controlled by the client like the client data store 155. In other cases, the training data may be obtained from public sources or from data sets used to derive previously trained natural language models. Example processes describing further details of how the natural language platform may be configured to ingest the training data will be described in more detail below.
  • At block 510, the natural language platform may then be configured to create a hierarchical data structure used to organize the collection of documents in question, according to some embodiments. As previously mentioned, this hierarchical data structure may be referred to as an ontology. The ontology may include one or more labels that the documents are to be categorized into. The one or more labels may be organized into a hierarchical structure, based on a logical relationship between the labels. For example, if one label may be thought of as having a logical subset of one or more additional labels, then the one or more additional labels may be represented as a sub node under the one label. Examples of this hierarchical tree structure will be described in more detail below.
  • In some embodiments, the ontology may be created based on labels generated or discovered through the topic modeling process mentioned above. For example, if a client desires to generate a natural language model for performing the task of classification analysis of a random collection of tweets generated within the last two weeks of a calendar year, and the client wants to classify the collection of tweets into the top 10 most discussed topics, topic modeling may first be performed on the collection of tweets to discover precisely what these top 10 topics are. These top 10 topics may then be organized in a hierarchical structure of an ontology, and may include some of the topics as sub nodes under other topics, based on a logical or thematic relationship. For example, Christmas may be a very often discussed topic within the last two weeks of the calendar year, and within the general topic of “Christmas,” there may be discussed “presents” and “Santa Claus.” Thus, the topics (also called labels) of “presents” and “Santa Claus” may be organized in the ontology as sub nodes of the “Christmas” label.
  • In some embodiments, the ontology may also be created based on direct inputs from the client, such as manual inputs for labels the client knows or intends should be included in the ontology, as well as previously generated ontology's to be imported by the client. In addition, the ontology may be created through an iterative verification process involving receiving annotations from expert annotators or other analysts. Additional details of these variants will be described more below.
  • At block 515, after an ontology has been established, the natural language platform may be configured to conduct an adaptive machine learning process to generate the natural language model. This adaptive machine learning process may train the natural language model on how to categorize the training data into the various labels as structured in the ontology. In some embodiments, the natural language platform may be configured to conduct an iterative process including generating and receiving human annotations about the documents and then training the model based on these annotations. In some embodiments, the iterative process may be repeated based on a determination of whether the trained model meets various performance criteria. If the performance criteria are not satisfied, in some embodiments, the natural language platform may also be configured to conduct one or more optimization techniques before repeating the process of generating and receiving new human annotations and then training the model based on these annotations. Further details for conducting this adaptive machine learning process are described in more detail below.
  • At block 520, once the training process is complete and the natural language model has been sufficiently trained (e.g., the trained model satisfies one or more threshold performance criteria), then the natural language model may be utilized by the natural language platform to analyze untested data as specified by the user or client. As referred to in previous figures, such as FIG. 1B, this phase of the process may be referred to as predicting or generating predictions, and the utilization of the natural language model at this stage may be consistent with the discussion of predictions as described in FIG. 1B. An example output for analyzing the untested data using the natural language model may include organizing the documents into the various labels and sub labels as specified in the ontology structure, or tagging individual documents with all relevant labels and sub-labels from the ontology. In some embodiments, various graphics in one or more graphical user interfaces may also be included and may being configured to graphically depict how the documents are organized within the ontology structure. Examples of these displays are described in application (Attorney Docket No. 1402805.00013_IDB013), which is again incorporated herein by reference in its entirety.
  • Referring to FIG. 6, diagram 600 provides a more detailed view of the block 505 of FIG. 5, describing the ingest data phase, according to some embodiments. For example, at block 605, the natural language platform may be configured to ingest data related to one or more pre-existing ontologies. The data of the pre-existing ontologies may be used to help build the ontology for the present natural language model, while in other cases one of the pre-existing ontologies may be used exactly as the intended ontology for the present natural language model. These pre-existing ontologies may be obtained from the user or the client, for example, when the client has already performed some work in building the natural language model previously.
  • At block 610, the natural language platform may also be configured to ingest existing classification rules for use in training the natural language model, according to some embodiments. As previously mentioned, relying on numerous rules has been a classic way for training natural language models. While aspects of the present disclosure do not rely exclusively on utilizing rules, or even utilizing any rules at all, in some embodiments, the inclusion of at least some rules may still be helpful in training the natural language model. In some cases, the existing classification rules may be obtained from the user or the client, for example, when the client has already performed some work in building the natural language model previously. In some cases, the included rules may represent absolute conditions, such as always classifying a document when the document contains a specific word or phrase. In other cases, the included rules may represent probabilistic conditions that may be weighted against other probabilistic factors. For example, the rule may assign a weight to the document to be classified as a certain label, where said weight may be factored into an overall composite score including other weights based on other factors.
  • At block 615, the natural language platform may also be configured to ingest test data in a specified data format. An example of ingesting the test data may include the example processes for obtaining data from the client data store 155 in FIG. 1B. Examples of a specified data format may include the test data being converted into all text, or the test data being converted into another uniform format. In some embodiments, the test data may be first ingested in its various formats, and then within the natural language platform, additional processes may convert the data into the uniform format. In other cases, the test data may be first converted into a uniform format, for example, at the client level, which is then ingested by the natural language platform.
  • At block 620, the natural language platform may also be configured to perform various normalization and filtering procedures on the test data. For example, the documents converted into all text may then be broken up into individual units of language, referred to as tokens. Examples of tokens can include single words, character sequences, punctuation marks, and spaces. As another example, related and semantically-equivalent tokens may be converted into canonical forms. Examples of such canonical normalizations include decomposing character sequencing containing diacriticals (e.g., decomposing “A” to “A” and a combining grave accent) and jamo, expanding abbreviations into the original word (e.g., expanding “Assn.” to “Association”), and replacing alternate variations of emoticons with a primary form (e.g., replacing “;-)” or “˜_{circumflex over ( )}” with “;)”). Another example includes deduplication of identical, near-identical (retweet/quoted reply), and thematically similar items
  • While the diagram 600 shows arrows progressing from one block to another, in some cases each of the individual blocks 605, 610, 615, and 620 may be performed in different orders, and in some cases not all of the example blocks 605, 610, 615, and 620 need be performed. In general, it may be apparent to those with skill in the art other processes for ingesting data to be used in generating a natural language model, and embodiments are not so limited.
  • After the various types of data have been ingested at block 505, e.g., any or all of the example blocks 605, 610, 615, and 620 are performed, at block 625, the totality of the ingested data may be processed for use in the natural language platform for generating one or more ontologies and for conducting the adaptive machine learning process to generate one or more natural language models.
  • Having described example details of block 505 for ingesting data, the next FIGS. 7-10 describe further details of block 510, related to describing example details of various ontologies and an example detailed process for creating the ontology, according to some embodiments. Referring to FIG. 7, a simple example of an ontology is shown through the various charts 700, 710, and 720.
  • Chart 700 shows a spreadsheet of two tasks, with each task including multiple labels. For example, a “sentiment” task includes three labels, “positive,” “negative,” and “neutral.” As an example, a user may desire to utilize a natural language model to perform a task of classifying documents by the sentiment each document conveys. The various types of sentiment that the documents may be categorized into are enumerated by the labels under this sentiment task, i.e., “positive,” “negative,” and “neutral.” The set of documents may be a collection of tweets about a basketball game, for example. The views expressed in the tweets may therefore be classified as either having positive sentiment, negative sentiment, or neutral sentiment, and the user or client may desire to have a natural language model generated to evaluate these tweets and classify them into one of these various labels.
  • In another example, still referring to chart 700, a “relevant” task includes two labels, “relevant,” and “irrelevant.” Here, the user may desire to utilize a natural language model to perform a task of classifying documents according to their relevance, where the determination of relevance may be specified by the user or associated analysts or annotators. For example, the determination of relevance may be based on whether a tweet among the collection of tweets pertains to the basketball game. The collection of tweets may have been obtained over the approximate period of time that the basketball games played, but since other events may be occurring at that time, it may not be a guarantee that all of the tweets necessarily discuss the basketball game. Thus, the natural language model may be configured to evaluate these tweets and classify them as either relevant or irrelevant, according to the criteria used to train the model.
  • Still referring to chart 700, a column for “subtasks” is shown, which may represent connections between one task to another. Examples of subtasks will be discussed in the later figures. All told, these relationships according to the chart 700 comprise an example of a simple ontology. That is, a natural language model may perform the sentiment task and/or the relevant task to classify a collection of documents (or spans). The natural language model may perform these tasks by evaluating each document among the set of documents and classifying them into one or more of the labels under each task.
  • Referring to chart 710, the text of the chart 700 is expressed in a textual format. This example format may represent how the natural language model is configured to ingest the ontology so as to perform the proper analysis.
  • Referring to chart 720, the example ontology formed by the words in chart 700 are shown here being organized in a hierarchical tree structure. That is, the overarching tasks, i.e., “relevant,” and “sentiment,” may represent the main nodes of the hierarchical structure, as shown. The labels associated with the respective tasks, i.e., “positive,” “negative,” and “neutral” for the “sentiment” task, and “irrelevant” and “relevant” for the “relevant” task, are placed under their respective tasks as sub nodes.
  • Referring to FIG. 8, a more complex example of an ontology is described. In chart 800, the same tasks and associated labels for each task as in FIG. 7 are shown. In addition, a subtask of “sentiment” is listed and is associated with the label “relevant” under the task “relevant.” What this means, according to some embodiments, is that the task of “sentiment” and associated labels are to be embedded as sub nodes under the label “relevant.”
  • Referring to chart 810, the text of the chart 800 is expressed in a textual format. This example format may represent may be how the natural language model is configured to ingest the ontology so as to perform the proper analysis. Notice how the third word “sentiment” is placed after the label “relevant” and a delimiter (e.g., a comma). The natural language platform may be configured to understand that the third word or phrase (e.g., words or phrases after a second delimiter the same line) represent a subtask, based on the organization defined in the first line of chart 810.
  • Referring to chart 820, the example ontology formed by the words in chart 800 is expressed in a hierarchical tree structure according to the organization defined by the headers in chart 800. That is, because the task “sentiment” is listed as a subtask associated with the label “relevant,” the “sentiment” task and associated labels are placed as sub nodes under the “relevant” label, which in itself is a sub node of the “relevant” task, as shown. In this way, when the natural language model analyzes various documents (or spans of documents) according to the structure of this ontology, the natural language model will perform the sentiment analysis task only if the document has first been classified as being “relevant” when performing the “relevant” analysis task. Put another way, if the natural language model analyzes a document (or a span) and determines that the document or span is to be classified as “irrelevant,” then the natural language model will discontinue any further analysis of the document or span, since the “irrelevant” label has no further sub nodes to travel down. In general, this example illustrates that the tree structure of the ontology can be used to guide the varying levels of analysis that a natural language model is to perform when analyzing documents.
  • Referring to FIGS. 9A and 9B, an even more complex example of an ontology is shown. Referring to chart 900, an additional task called “genre” is included in the ontology. The genre task is listed as a subtask under the “relevant” label. The genre task includes four labels, “mystery,” “comedy,” “action,” and “horror.” Also, the mystery, comedy, and action labels have “sentiment” as a subtask, while the horror label has no subtask listed. This example shows that a single task may be listed as a subtask under multiple labels, and that tasks can be listed as sub tasks in a cascading or hierarchical fashion (e.g., “sentiment” is a subtask of multiple labels in the “genre” task, which is itself a subtask of the “relevant” label of the “relevant” task).
  • Referring to chart 910, the text of the chart 900 is expressed in a textual format. This example format may represent may be how the natural language model is configured to ingest the ontology so as to perform the proper analysis.
  • Referring to chart 920, the example ontology formed by the words in chart 900 is expressed in a hierarchical tree structure according to the organization defined by the headers in chart 900. As shown, because the sentiment task is listed as a subtask for three of the labels in the genre task, the sentiment task and its associated labels are embedded as sub nodes under each of the three labels under the genre tasks, i.e., the “action,” “comedy,” and “mystery” labels. In addition, it can be seen that the entire genre task and associated labels (and subsequent subtasks) is embedded as a sub node under the relevant label, consistent with the organization described in chart 900.
  • Still referring to chart 920, a natural language model may conduct analysis of documents or spans of documents according to the hierarchical structure as shown. Following the hierarchical structure of the ontology, a natural language model may start by conducting “relevant” analysis of the document or span to determine if the document is “relevant” or “irrelevant.” If it is determined that the document is irrelevant, then the analysis ends. However, if the document is determined to be relevant, then the natural language model may proceed by conducting “genre” analysis, to determine whether the document should be classified as fitting into the “action” genre, the “comedy” genre, the “horror” genre, or the “mystery” genre. If classified as fitting into the horror genre, the analysis ends. However, if classified as any of the other three labels, the natural language model may continue by performing “sentiment” analysis, and classifying the document by its sentiment according to the sentiment labels “negative,” “neutral,” or “positive.”
  • Referring to chart 930 in FIG. 9B, the example ontology described herein is shown in a logical hierarchy format. Various tasks in the ontology are shown in rectangular boxes, while the labels are shown in ovals. The arrows show the progressive relationship for how a natural language model may proceed through the ontology when analyzing a document. Thus, the first level analysis performed is to analyze relevance. If deemed relevant, the next level of analysis performed is genre analysis. If the document is deemed to be categorized into one of the three labels other than “horror,” then sentiment analysis is performed last. This logical structure also illustrates how some ontologies may be structured in more complicated ways that a simple tree structure, as the analysis paths after “Mystery,” “Comedy” and “Action” all converge to continue performing sentiment analysis.
  • Now that examples of ontologies and their applications have been described, referring to FIG. 10A, flowchart 1000 provides additional example details for generating the ontology, according to some embodiments. The example processes of flowchart 1000 may represent further details in block 510 of the high level flow chart of FIG. 5 for generating natural language models. These example sub processes may begin at block 1005, which refers to ingesting training data and other types of data. Block 1005 may be consistent with block 505 and the related details described in flowchart 600 of FIG. 6, where the example outputs at block 625 may serve as the example inputs into flowchart 1000. Again, this data may include a collection of documents or spans of documents, one or more rules, and/or one or more pre-generated ontologies as defined by the client or user.
  • Having at least obtained a collection of documents intended for a natural language model to analyze, in some embodiments, the natural language platform may be configured to first perform a topic modeling process, at block 1010. This topic modeling process may be consistent with those descriptions of topic modeling described above, as well as examples for performing topic modeling as described in applications (Attorney Docket Nos. 1402805.00012_IDB012, 1402805.00013_IDB013, 1402805.00014_IDB014, 1402805.00016_IDB016, 1402805.00017_IDB017, and 1402805.00019_IDB019), again incorporated herein by reference. The topic modeling process may be used to determine one or more labels for the natural language model to categorize the collection of documents into, and of which may be organized into a hierarchical structure according to the following processes described below. In some embodiments, topic modeling may not be performed, and instead the natural language platform may access pre-generated or predetermined sets of topics as defined by the client or user or previously obtained in the ingest data phase.
  • At block 1015, the natural language platform may be configured to conduct a rules generation process to confirm and refine rules that may help define how the ontology is to be structured. For example, the natural language platform may access a number of rules, where some may be obtained from the previously ingested data, while other rules may be programmed or otherwise specified by human analysts with expert knowledge of the subject matter for which the natural language model is to be generated for. The natural language platform may receive these inputs about new rules or modifying pre-existing rules from the analysts through one or more graphical user interfaces, such as through the one or more application modules 225 as described in FIG. 2A. The rules generation process at block 1015 may then include verifying the enumerated rules for logical consistencies. For example, the natural language platform may be configured to evaluate the totality of rules to see if any of the rules lead to infinite circular loops, or lead to definitions of labels or tasks that lack proper node connections sufficient to fit in the ontology. As another example, the rules may be evaluated to determine if two or more of the rules conflict with each other, or if two or more of the rules are redundant.
  • At block 1020, the natural language platform may be configured to generate a first version of the ontology based on the topics discovered or previously obtained and the verified rules in the previous steps. For example, the topics may define the content of the nodes for an ontology, while the rules define the relationships among the various topics. For example, a rule may specify to perform a genre analysis task of a document after determining whether the document is determined to be relevant. Thus, this rule may help define that the genre task should be a sub node under the relevant task. An example process for how rules contribute to generating an ontology includes the following steps:
  • Given a goal of performing Genre analysis, over a collection of documents discussing movies, an analyst may:
  • Use topic modeling to automatically identify relationships and clusters between documents;
  • Manually review the topics, recognizing that several of the identified topics are discussing Horror movies. There may be multiple topics related to Horror movies, e.g., a topic about the Halloween series of films, identified by keywords Halloween, Myers, Curtis, Knife, and a topic about the Nightmare on Elm Street series, identified by keywords Nightmare, Krueger, Craven, and Dreams;
  • Manually create a classification (label) for Horror within the Genre task; and
  • Manually define rules that cause documents which include the words “Halloween” or “Nightmare” to be classified as Horror with greater confidence.
  • In some embodiments, it is possible to create an ontology without topic modeling or rule-writing, by creating a classification scheme and annotating it. This can be sufficient if the classification problem is well-defined and annotations can be crowd-sourced effectively. Sentiment analysis is an example of this.
  • After applying all of the rules to the various topics to create a hierarchical structure for the ontology, at block 1025, the natural language platform may be configured to test the effectiveness of the ontology, by first generating one or more annotation questions designed to prompt one or more annotators to evaluate the ontology. For example, an annotator might be asked “is there positive sentiment expressed about the car in this sentence?” or “is this statement an acceptable answer to this question?” These annotation prompts may be displayed in the one or more graphical user interfaces, such as through the application modules 225. In some embodiments, analysts may interact with these annotation prompts at the client site, while in other cases the analysts may be stationed locally near the natural language platform.
  • Referring to FIG. 10B, illustration 1027 shows an example annotation interface to collect input in the form of document annotations/classifications from human analysts and other experts, according to some embodiments. These annotations can be used to evaluate the effectiveness of the ontology and to train the natural language processing models. The panel on the left includes a sample document, while the panel on the right includes an annotation prompt and a series of labels in the ontology that the human analyst may select to specify under which label(s) the document belongs to.
  • Referring back to FIG. 10A now, at block 1030, the natural language platform may be configured to receive the various annotations entered by the human analysts, such as those inputs provided at the example interface in FIG. 10B. These annotations may provide a form of feedback to rate the quality of the ontology structure.
  • At block 1035, the natural language platform may be configured to evaluate the performance or effectiveness of the ontology based on the received annotations. Evaluation metrics may include the label distribution (how frequently does the annotator select each label) and the proportion of annotated documents are not assigned any label.
  • Referring to FIG. 10C, illustration 1037 shows a sample annotation interface to collect input in the form of document annotations/classifications from human experts, according to some embodiments. These annotations can be used to evaluate the effectiveness of the ontology and to train the natural language processing models.
  • Referring back to FIG. 10A now, in some embodiments, the ontology generation process within block 510 may involve refining the ontology through a series of iterations. For example, it may be determined at block 1035 that the ontology in its current form does not satisfy certain performance criteria. Again, these evaluations may be based on the feedback supplied by the annotations. To improve the ontology content and structure, at step 1040, it may be determined that new rules should be generated or that existing rules may need to be modified. Thus, the ontology creation process may cycle back to block 1015. From here, the natural language platform may facilitate generation of new rules or modification of existing rules through the graphical user interface, so as to receive inputs from analysts to modify the rules. The rules generation process may again verify the veracity of the totality of rules to make sure that there is logical consistency. The process may then proceed to block 1020, then to block 1025, and so on to generate a refined ontology in a second iteration.
  • In other cases, it may be determined that modifying the rules may not be necessary as the accuracy is already at a level acceptable to the system user, but at the least, that the ontology may still be improved based on the annotations provided previously. Thus, at step 1045, the natural language platform may cycle back to block 1020 and may refine the ontology by incorporating the feedback from the annotations. Similar to step 1040, the ontology creation process may then proceed to block 1025, then to block 1030, and so on in a second or later iteration.
  • The iterative process for continually refining the ontology between steps 1040 and steps 1045 may continue a number of times until the evaluation of the ontology at block 1035 satisfies one or more performance criteria like inter-annotator agreement, using a metric derived from Krippendorf's Alpha. If it is determined that the performance criteria is satisfied at step 1050, then at block 1055, the ontology may be considered complete enough and may be used in the next step of the natural language model generation process, which is the adaptive machine learning process at block 515.
  • In some embodiments, at block 1060, during the evaluation phase of the ontology creation, analysts may evaluate the data and determined that certain clarifications may be needed to distinguish among various labels or tasks within the ontology. That is, part of the ontology refinement process may include trying to disambiguate two or more topics from each other, because the topics may be very closely related or potentially confusing. For example, when trying to classify news articles about a financial crisis, the label of “legal” articles may be potentially confusing or overlapping in some ways with the label of “security fraud.” While there may be a number of ways to distinguish between these two labels, the analysts may provide and the natural language platform may be configured to receive manual inputs to specify particularly how it is desired for these two topics to be distinguished. In order to do this, analysts may help generate training guidelines that may be used for instructing annotators in the adaptive machine learning process phase. These training guidelines may be expressed as human readable instructions that help define what the different labels mean, and/or how to choose among two or more labels when it is possible for a document to be classified into the two or more labels. For example, to distinguish between “legal” and “security fraud,” a training guideline may instruct the annotator to always classify the document as “security fraud” if it appears the document may fall under both categories. In other cases, the training guideline may attempt to provide definitions to help determine what constitutes “security fraud” compared to “legal” generally. In some embodiments, these training guidelines may be displayed to the annotators in one or more of the graphical user interfaces. An example of these training guidelines as described more in application (Attorney Docket No. 1402805.00013_IDB013), which again is incorporated herein by reference.
  • Having now described an example of a more detailed process for creating the ontology (see block 510 of FIG. 5), FIGS. 11A-13 will focus on more detailed examples of generating the natural language model through the adaptive machine learning process (see block 515 of FIG. 5), according to some embodiments.
  • Referring to FIG. 11A, illustration 1100 shows an example display for annotating as part of the adaptive machine learning process, according to some embodiments. As previously mentioned, the adaptive machine learning process according to aspects of the present disclosure include an iterative interplay between annotations of documents and machine learning to generate the natural language model based on the annotations. Illustration 1100 shows an example of one type of annotation applied to a document. In some embodiments and as previously alluded to, the display interface for allowing a human annotator to supply an annotation is called a work unit. In some embodiments, each work unit represents a single task for supplying a single annotation to one document (or span). In some embodiments, the natural language platform is configured to efficiently generate work units such that human annotators' time is efficiently utilized.
  • As shown in illustration 1100, examples of a work unit may include the text from a document or at least a subset of text. As shown, for example, the document on the left side of the work unit is a tweet from Top World News. The text states, among other things, “J.P. Morgan faces scrutiny over Asia hiring practices: J.P. Morgan Chase's hiring of the son of a Chinese . . . ” This document also includes a photo and a title and introductory words to a news article that is referenced in the Tweet. On the right side of the work units is an annotation prompt, that includes a subject header and a human readable question. The annotation prompt, including the human readable question, may be generated automatically by the natural language platform. The natural language platform may determine the form of the question so as to most efficiently retrieve the desirable annotation from the annotator. In this case, the natural language platform has generated a human readable question that prompts the annotator to select as many topics (labels) that the annotator believes best applies to the document on the left. This type of question is a rather open ended question, and in other cases, the natural language platform may be configured to generate more pointed questions when it has determined that it needs more specific information. In this case, the annotator has selected two labels that he or she believes is most applicable: “legal & compliance,” and “corporate news.” Once submitted, this particular document would then have this particular annotation applied to it. Additional annotations may be applied to this document based on inputs from other annotators, when the other annotators see this particular work unit or other work units that include this particular document.
  • Referring to FIG. 11B, illustration 1150 represents another example of a work unit used to annotate a document, according to some embodiments. Like before, the text of the document to be annotated is shown on the left. On the right side of the work unit is the annotation prompt, which includes a header and a human readable question for the annotator to answer. In this case, the question of “What is the best label for this document?” allows for the annotator to select only one of three options: “highly relevant,” “semi-relevant,” and “totally irrelevant.” In some embodiments, the annotator may have been provided training guidelines to help define for the annotator what it means for something to be relevant, semi-relevant, or irrelevant. For example, a training guideline may have instructed the annotator to determine the relevance of the document to the topic of “corporate finance.” In addition, the training guideline may have provided definitions to distinguish between the document being highly relevant versus merely semi-relevant. In some embodiments, a button or icon for the training guideline may be placed in the display of the work unit that displays the training guideline when the button or icon is pressed. In other cases, a previous display screen may have provided instructions based on the training guidelines to the annotator to prepare the annotator for how to respond to the annotation prompts.
  • Still referring to illustration 1150, the type of annotation prompt shown here is a more narrow scope than in illustration 1100. Here, the annotator is restricted to selecting only one of the three options. In contrast, the annotator in illustration 1100 is allowed to select as many options as deemed relevant. In some embodiments, these differences in questions and their intended scopes are intentionally and automatically produced by the natural language platform and may be based on what type of information the natural language platform desires to gain from the annotators.
  • For example, if the natural language platform is in the early training phases for generating the natural language model, general information about the various types of documents may be desired. Therefore, the types of annotation prompts may be more open ended, and may allow for annotators to answer broad questions, such as questions that may have multiple answers (e.g., see illustration 1100).
  • As another example, if the natural language platform is in more mature training phases, more specific information to refine determination between labels may be desired. Therefore, the natural language platform may desire specific information to determine differences between two specific labels. The natural language platform may be configured to generate an annotation prompt that directs the annotator to distinguish the two specific labels using a pertinent document. In this case, an example annotation prompt may include the question, “Is the best label for this document label ‘A’ or label ‘B’?” Then, the natural language platform allows the annotator to select between only label A or label B.
  • In general, the natural language platform may be configured to generate annotation prompts with varying degrees of specificity, depending on the type of information desired to better refine the natural language model. Other examples can include true/false questions, yes/no questions, and other types of questions that direct and annotator to varying levels of specificity apparent to those with skill in the art, and embodiments not so limited. These varying degrees of specificity also help to efficiently designate a human's time. As there are thousands, millions, or even billions of documents that could be annotated, a human's time can be valuable and scarce, such that the techniques presented herein should determine what documents are most appropriate to be sent to human annotators, and what types of information about the documents should be obtained.
  • With various examples of annotations now described, referring to FIG. 12, flowchart 1200 provides a more detailed process flow for conducting the adaptive machine learning process using the annotations to generate the natural language model, according to some embodiments. The example details herein may represent further details in block 515 of FIG. 5.
  • This sub process may start with block 1205, involving natural language platform ingesting the data and inputting or applying an ontology for use in the adaptive machine learning process. Examples of these processes are described in previous figures, such as FIG. 6 and FIG. 10. Having accessed the proper data, at block 1210, the natural language platform may begin the adaptive machine learning process to generate a natural language model by selecting particular documents to be annotated. The process for selecting which documents are to be annotated may be referred to herein as “intelligent queuing.”
  • In general, annotations to documents may be useful in aiding a computer to find patterns in the text of the collection of documents in order to determine how the document should be classified. The higher number of documents that are annotated, the more reliable it should be for the computer to create an accurate natural language model for the overall collection of documents. However, obtaining annotations for tens of thousands or even millions of documents in a collection may be impractical, and in some cases, a substantially fewer number of documents that are carefully selected may help produce a natural language model with substantially the same performance. For example, obtaining annotations for certain documents that are representative of the various topics or labels in the ontology, as opposed to exhaustively obtaining annotations for all documents in a single topic, may be a more efficient use of a human annotator's time. As another example, it may be more useful to obtain an annotation for a document that is more complex or possesses content that may arguably be categorized into two or more labels, as opposed to a document that is unambiguously within just a single label, because the annotation may help the machine learn what to do in potentially ambiguous cases consistent with this document. The process of performing intelligent queuing includes these processes of carefully selecting which documents among an entire collection of documents should be slated to receive annotations, with one objective being to efficiently utilize the human annotator's time.
  • When performing intelligent queuing in a first iteration, i.e., conducting block 1210 for the first time, the intelligent queuing process may include obtaining a seed set of documents that is representative of the various tasks and labels as outlined in the ontology. Other factors may be included in the seed set of documents, and further details of intelligent queuing are described in application (Attorney Docket No. 1402805.00012_IDB012), which again is incorporated herein by reference in its entirety.
  • After selecting which documents are to be annotated, the process of intelligent queuing at block 1210 may also include generating the annotation questions to be presented to an annotator (or multiple annotators) in order to elicit an annotation response. Examples of various types of annotation questions include the examples in FIGS. 11A and 11B. As previously mentioned, the wording of the question including the level of specificity desired to be elicited from the annotator may be automatically generated by the natural language platform, based on the type of information desired when performing intelligent queuing. Thus, block 1210 may also include causing display of the annotation prompt in a work unit that is accessible to an annotator. This may be displayed in one of the application modules 225, for example. Other example displays are provided in application (Attorney Docket No. 1402805.00013_IDB013), which again is incorporated herein by reference.
  • In some embodiments, providing annotation questions in the intelligent queuing process may also include presenting the annotation questions of the same document to multiple annotators. In this way, a single annotator may not be relied on to provide the annotation to a single document. For example, a single annotator may make mistakes, or in worst cases a single annotator may intentionally provide wrong answers. For example, obtaining annotations from annotators may involve a crowdsourcing process, and it may not be known the reliability of any single annotator.
  • At block 1215, the natural language platform may be configured to receive annotations in response to the annotation questions generated in block 1210. In between blocks 1210 and 1215, one or more annotators may interact with a graphical user interface to analyze one or more documents and provide annotations according to the annotation prompts supplied by the natural language platform. For example, an annotator may answer a question similar to one of the questions in FIG. 11A or 11B. These responses may be stored in a database repository, such as in the database modules 210.
  • At block 1220, in the event that multiple annotators supply annotations to the same document, in some embodiments, the natural language platform may be configured to conduct an annotation aggregation process that analyzes all of the annotations made to a single document to determine whether said document may be reliably annotated for use in training the natural language model. For example, if six annotators provided annotations to the same document, and three annotators classified the document as belonging to a first label, while the other three annotators classified the document as belonging to a second label, then the annotation aggregation process of block 1220 may determine that this documents is too ambiguous to be used in the adaptive machine learning training process. As another example, if five of the six annotators provided a consistent annotation to the document, while a sixth annotator provided a different annotation, it may be determined that this document has sufficiently consistent annotations so as to be used in the adaptive machine learning process.
  • In general, in some embodiments, the annotation aggregation process of block 1220 may include computing an annotation aggregation score based on the total number of annotations and reflecting a level of agreement between the annotations. Based on this computed score, in some embodiments, the annotation aggregation process may determine whether the document may be used to train the natural language model, e.g., if the score satisfies a first threshold criterion. In addition, in some embodiments, based on the score, it may be determined that additional annotations from other annotators should be elicited before the document may be used in the model training process, e.g., if the score satisfies a second threshold criterion. In addition, in some embodiments, based on the score, and may be determined that the document should not be used for the model training process, e.g., if the score satisfies a third threshold criterion. This process may be applied for each document to determine if said document is valid for training the natural language model. Further details about the annotation aggregation process are described in application (Attorney Docket No. 1402805.00019_IDB019), which again is incorporated herein by reference in its entirety.
  • At block 1225, the natural language platform may be configured to generate the natural language model using the aggregated annotated data (e.g. documents with consistent enough aggregation score) in an adaptive machine learning training process. In general, the adaptive machine learning training process may find patterns in the training data (e.g., aggregated annotated data) that may be used to make predictions about untested data. The process for adaptive machine learning according to some embodiments includes: (a) elicit annotations from humans (b) build initial version of model (c) evaluate the effectiveness of the model and determine if stopping criteria have been met (d) identify specific areas of improvement and annotations required for improvement in those areas (e) queue additional documents for human annotations, and serve those documents to the optimal annotator with the optimal interface for the combination of annotator and task (f) update the model using the new annotations (g) repeat the whole process until stopping criteria are met. From this, a first version of the natural language model may be generated, and, if desired, continually updated new versions of the model as well.
  • In general, the natural language platform according to aspects of the present disclosure first seeks to obtain annotations for particular documents before conducting any machine learning techniques. Receiving annotations before conducting any machine learning techniques may allow for the adaptive machine learning process to be more efficient, in that the adaptive machine learning begins building the natural language model from documents whose “true values” are already known, e.g., the documents already have reliable annotations.
  • At block 1230, the natural language platform may be configured to test the performance of newly created natural language model, according to some embodiments. Testing the performance of the natural language model may allow the natural language platform (and/or a user) to determine later on whether the natural language model should be refined and improved before being finalized.
  • In some embodiments, testing the performance of the natural language model at block 1230 includes performing a cross validation process. For example, the cross validation process may involve creating a first “test” natural language model, using the same adaptive machine learning training process to generate the original natural language model, but based on only a subset of the training data, e.g., 90% of the documents in the training data (e.g., supposing there are 1000 documents in the training data, then documents 1-900). This first “test” natural language model is tested by applying the first “test” natural language model on the remaining 10% of the documents not used to create the first “test” natural language model, but still already having annotations. In this way, the veracity of the first test natural language model may be compared against “true answers” provided by the original annotations of the remaining 10% of the documents. The cross validation process may also involve creating a second “test” natural language model, using the same adaptive machine learning processes to generate the original natural language model, but based on a different subset of the training data, e.g., a different 90% of the training data (e.g., documents 1-800 and 901-1000). This second test natural language model may be tested in the same way as the first test natural language model, and therefore may assess the veracity of the second test natural language model against the second 10% subset of the training data (e.g. documents 801-900). This cross validation process may continue each for a different 90% of the training data (for a total of 10 different iterations), so that the veracity of the machine learning process used to the create original natural language model (and each of the test natural language models) may be verified against different subsets of the training data, and ultimately against all of the training data divided into separate parts, where the “true answers” are known. Because all of the training data is being tested against, and because the test natural language models are generated using the same machine learning process as the one used to generate the original natural language model in question, the aggregate performances of the test natural language models is used as a reliable approximation to the original natural language model. In general, subdividing the training data into 10% increments is merely one example, and other partitions may be used.
  • At block 1235, after having tested the performance of the natural language model, either through the example cross validation process just described or through other validation processes apparent to those with skill in the art, the natural language platform may be configured to generate a performance metric of the natural language model, based on the tests run against it. For example, if the performance of the natural language model was determined using the after mentioned cross validation techniques, one or more metrics quantifying a degree of accuracy may be supplied. In other cases, one accuracy measurement for each of the “test” models in the cross validation process may be supplied, and the natural language platform may be configured to provide an aggregate score at block 1235. In some embodiments, these performance metrics may be displayed, such as in application modules 225. Other examples of these performance metrics are described in application (Attorney Docket No. 1402805.00012_IDB012), which is again incorporated herein by reference. Example performance metrics include: model precision, recall, and f1-measure, derived through cross-validation; area under the receiver operating characteristic curve (AUC); rate of improvement in precision, recall, and/or f1-measure as a function of the number of annotations; and percentage of the feature vocabulary in the total collection that is represented in the annotated data set.
  • From here, the natural language platform may determine if the performance of the natural language model satisfies one or more performance criteria. The one or more performance criteria may include a pre-determined threshold, or in other cases may be specified by a user. In the event that the natural language model satisfies the performance criteria, the process may proceed to block 1245. Otherwise, the process may proceed to block 1240, where the natural language model is to be refined through one or more optimization techniques according to some embodiments.
  • At block 1240, supposing that the natural language model fails to satisfy the one or more performance criteria, the natural language platform may be configured to perform various optimization techniques to improve the performance of the natural language model, according to some embodiments. Numerous examples of optimization techniques are described in FIG. 13 below. Some or all of the techniques may be performed, and embodiments are not so limited.
  • In some embodiments, one of the optimization techniques includes conducting another iteration of annotations on the same or different documents among the collection of documents. In other words, the process in flowchart 1200 may cycle back to block 1210, consistent with the iterative adaptive machine learning process described above. For example, through the tested performance of the natural language model at block 1230, it may have been determined that the natural language model accurately predicts documents for certain labels, while it struggles to predict documents into other labels in the ontology. The natural language platform may therefore be configured to conduct a second iteration of intelligent queuing that aims to obtain annotations from documents to help resolve labels that the original natural language model showed unsatisfactory performance in. Further example details of how the tested performance of the previous iteration of the natural language model may be used to guide a subsequent instance of intelligent queuing is described in application (Attorney Docket No. 1402805.00012_IDB012), which is again incorporated herein by reference.
  • To complete one or more iterations of refinement of the natural model, the adaptive machine learning process may continue through blocks 1215 through 1235, consistent with the descriptions above but with new annotations and possibly other optimization techniques performed to generate the revised natural language model. In other cases, after one or more optimization techniques are performed at block 1240, rather than repeat through block 1215 to 1220, the model generation process may proceed straight to block 1225 to generate an updated natural language model using the updated parameters from the one or more optimization techniques and the existing aggregated annotations. Ultimately, it may be determined that the natural language model satisfies the one or more performance criteria, at which point the process may proceed to block 1245. Here, the finalized natural language model may be transmitted to make predictions on untested data. The finalized natural language model may be transmitted to a system exclusively controlled by a client, such as in some of the examples shown in FIG. 2A, or in other cases may be transmitted to specific devices, such as devices 120 or 130.
  • In some embodiments, the transmitted natural language model may be stored in a format that is universally portable. For example, the natural language model may be stored in memory efficiently such that the same model may be executable in mobile devices as well as conventional computers. Additional details of the universal portability aspect according to some embodiments are described further in application (Attorney Docket No. 1402805.00018_IDB018), which is incorporated herein by reference in its entirety.
  • At block 1250, the natural language model may be used to analyze untested data. The untested data may be based on data obtained from the client not used as part of the training data, while in other cases the untested data may include live or real-time data, such as live streams of tweets or customer service emails. In general, the natural language model may analyze untested data by classifying each unit of the untested data into one or more of the tasks or labels as organized in the ontology. The natural language model may be used on millions or even billions of units of data.
  • Referring to FIG. 13, illustration 1300 provides further details of different examples of optimization techniques that may be applied in block 1240 of FIG. 12 for performing optimization techniques to improve the natural language model, according to some embodiments. The following are some descriptions of each of the different examples listed, although some embodiments may use only some of these examples, and embodiments are not so limited.
  • At block 1305, one example for performing optimization techniques to improve the natural language model may include using the results from cross validation or other performance evaluation techniques to guide the selection of further documents to be annotated when performing a next iteration of intelligent queuing. As previously mentioned, additional annotations to documents, either to some of the same documents or even new documents, may be sought after to help resolve some ambiguities in the natural language model or to help improve areas that the natural language model has shown difficulty in classifying. To determine which documents should be used in this next round of annotations, the intelligent queuing process according to aspects of the present disclosure may base these decisions on the performance results generated by performing the cross validation or other performance evaluation techniques from block 1230 of FIG. 12. Further example details for how the intelligent queuing process utilizes the performance evaluation techniques from block 1230 are described in application (Attorney Docket No. 1402805.00012_IDB012), which is again incorporated herein by reference. Block 1305 may also include a process for what type of annotation questions should be presented to an annotator to achieve the proper level of detail for improving the natural language model.
  • At block 1310, another example technique includes performing the intelligent queuing process itself, which is to say that the process proceeds back to block 1210, which has already been discussed.
  • At block 1315, another example optimization technique includes performing feature selection. Feature selection may include applying one or more transformations to each annotated document to generate a set of machine learning features for that document, applying one or more evaluation metrics to each feature to estimate the relative importance of each feature to the natural language model, and finally selecting an appropriate subset of features for inclusion in the natural language model. For example, every token in each document may be designated as a feature, and each feature may be ordered based on the frequency it appears in the total set. Other examples may use statistical co-occurrence criteria (such as chi square and mutual information) or information theoretic criteria (such as information gain) to perform feature evaluation. The feature transformation process is described in more detail in application (Attorney Docket No. 1402805.00017_IDB017), again incorporated herein by reference.
  • At block 1320, another example includes performing padding and rebalancing of the data in a natural language model. This may include duplicating or deleting specific annotations to create an artificial mix of training items to improve model performance in specific areas. Padding and rebalancing may include identifying individual items in the data that are representative of rare but significant labels in the ontology and artificially duplicating these items so that they make up a larger proportion of the training data used to develop the natural language models. This technique can also be referred to as “upsampling.” For example, if positive items are rare relative to negative items in a set of training data used to develop a sentiment classification model, the positive items identified during annotation can be copied one or more times and included in the training data for the model. The number of times that the positive items are copied can be determined through a number of optimization techniques in machine learning, including cross validation. Cross validation can identify the number of positive and negative items to include in training that will achieve the best model performance (as measured by accuracy, precision, recall or f-score) by cyclically adjusting the numbers of items in each category, retraining the model, and testing the performance of the resulting model. Alternatively, individual items in the data that cause the natural language model to make erroneous classifications may be removed from the training data to improve model performance. This technique can also be referred to as “downsampling.”
  • At block 1325, another example includes performing pruning techniques. Pruning techniques may include processes to reduce the size and complexity of certain types of statistical models such as decision trees that are used to develop natural language models. Pruning has the benefit of reducing the tendency of a model to represent noise in the data (overfitting) by removing parts of the model that are useful in classifying only a small number of items in the training data. For example, a natural language model that is trained to distinguish five different topical categories of news articles using decision trees and individual word features, may perform well on training data, but poorly on unseen data. In order to improve the generalizability of the model and to make performance more consistent, certain nodes or features used by the model that only help in classifying less than 1% of the data may be trimmed away or removed to simplify the model and improve performance across different data sets.
  • At block 1330, another example includes performing ensemble or voting methods. This may include combining multiple pre-trained natural language processing models to generate multiple predictions for each unlabeled item in the data set, identifying the items in the dataset that meet specific prediction matching criteria (e.g., unanimous, majority, at least n), and using these items to train a new natural language processing model. This approach may sometimes be referred to as semi-supervised machine learning. In some embodiments, the pre-trained models have been trained using a subset of training items from the target data set. In other embodiments, the pre-trained models have been trained using training items from external data sets. A variety of matching criteria can be applied to the resulting predictions to identify items that meet the minimum requirements for being included as training items in a new natural language model. In some embodiments, a majority vote of the models will determine the label that is assigned to an item in the dataset. For example, in an ensemble of six different sentiment models, if four models predict that an unlabeled tweet is positive, and two models predict that that same tweet is negative, the item will be assigned a positive label and subsequently used for training a new sentiment model. On the other hand, if three of the models predict that an unlabeled tweet is positive, and three models predict that it is negative, the item will not be assigned a label and will not be used for training a new sentiment model. In other embodiments, an unlabeled item will receive a label only if that label corresponds to at least n matching predictions from the ensemble of classifiers. In other embodiments, an unlabeled item will receive a label only if that label corresponds to a unanimous set of predictions from the ensemble of classifiers.
  • At block 1335, another example includes performing feature discovery. This may include identifying the set of feature transformations and feature selection criteria that provide the highest model performance for the provided data. In some embodiments, the set of selected feature transformations and feature selection criteria can be represented as a string of independent binary values. For example, each supported feature transformation or selection criteria may be assigned a position within the string, where a 1 value corresponds to the feature or criteria being enabled and a 0 corresponds to disabled. In such embodiments, a genetic search process may then be used to iteratively define numerous configurations, training and evaluating each in sequence according to the training and testing processes 1225, 1230 and 1235 in FIG. 12, refining and improving the configuration each iteration. In such embodiments, the final configuration of feature transformations and selection criteria may be saved to a database so that later applications of the feature selection process 1315 use the optimized configuration.
  • At block 1340, another example includes performing model interpolation. This may include combining model parameters from at least one previous model with a provided set of model parameters for the desired model to create an interpolated model which has greater performance than would be achieved from the provided set of model parameters. In some embodiments, the parameters incorporated from the at least one previous model may include the annotations and documents used to train each previous model, and the provided parameters for the desired model may include a set of label names, documents, annotations, and rules. In some embodiments, the interpolated model is generated by selecting the parameters from the at least one previous model which do not conflict with the provided parameters, and proceeding with the model training and optimization processes using the combined data set. For example, a previous model may include a large set of English social media posts from Twitter® discussing a variety of topics of general interest (movie releases, consumer electronics, recent personal experiences, and current events), and annotations imputed to each document according to the sentiment (“positive” or “negative”) conveyed by that document. Conventionally, such models perform with poor accuracy when applied to process text that uses highly specific and unique terminology; for example, social media posts about the finance industry. According to some embodiments, an interpolated model may be generated by providing a small amount of parameters (for example, a rule defining “overweight” as conveying “positive” sentiment), combined with the documents and annotations from the previous model parameters that do not conflict with this. In this way, highly accurate domain-specific models may be interpolated from a large amount of previous parameters and a comparatively small amount of provided parameters.
  • At block 1345, another example includes developing new rules for use in defining the ontology or additional annotations. Further descriptions for how the new rules may be implemented are described in FIG. 14, below.
  • At block 1350, another example includes tuning model training hyper-parameters. Examples of model hyper-parameters include regularization constraints applied to the entire model training process (such as a constraint on the L1 or L2 norms), e.g., at block 1355, or adding a smoothing term to each feature individually during the feature selection process (such as a Laplacian smoothing term), e.g., at block 1360. Conventional machine learning techniques produce models which attempt to closely match their training data. For example, in a set of documents about “Bill Clinton” that each have associated annotations imputing “positive” or “negative” sentiment, a small subset of such documents may be found to also discuss “Kenneth Starr,” with an overwhelming percentage of this subset imputed as conveying “negative” sentiment. In such a case, a conventional model training process may learn to interpret “Kenneth Starr” as strongly conveying “negative” sentiment, although this learned interpretation is an artifact of limited training examples rather than an actual conveyance of “negative” sentiment; such a case is known to those with skill in the art as an example of “over-fitting.” In such cases, introducing hyper-parameter constraints can be used to penalize rare terms, such that the model training process will not treat such terms as strong conveyances of any label. In some embodiments, each hyper-parameter may assume an infinite number of possible values (for example, a scale factor or additive term that may be assigned any numerical value). In such embodiments, selecting appropriate hyper-parameter values involves training and evaluating a large number of models with different hyper-parameter configurations. In some embodiments, Monte Carlo techniques, e.g., block 1365, such as the Metropolis-Hastings algorithm may be used to iteratively select and refine hyper-parameter values until a suitably accurate model is created or a maximum number of iterations have been performed.
  • In general, the iterative process as described in flowchart 1200 may efficiently develop a natural language model by identifying and focusing on problem areas of the natural language model through multiple annotation and test phases. In this way, whereas conventional processes for developing a natural language model may take up to six months, achieving the same performance through aspects of the present disclosure may take only three weeks. In general, the processes described herein for producing a natural language model drastically reduce time to convergence and efficiently utilize time spent by humans to provide input for the machine learning components.
  • Referring to FIG. 14, flowchart 1400 illustrates a modified process for generating natural language models, including iterating processes between some processes within the create ontology block and the conduct adaptive machine learning block of FIG. 5, according to some embodiments. In previous descriptions, iterative processes were described to occur within the create ontology block 510 and within the conduct adaptive machine learning block 515. In some embodiments, a third relationship of iterations may also be allowed, such as processes within the adaptive machine learning block 515 iterating back to certain blocks within the create ontology block 510. For example, as shown in flowchart 1400, while blocks 1405, 1410, 1415, and 1425 may include descriptions that are consistent with a first iteration of processes described in previous figures, such as in FIGS. 5, 10, and 12, additional iterative loops may also be possible from the adaptive machine learning block 1425.
  • For example, after at least a first iteration of the natural language model is generated at block 1425, consistent with the descriptions in FIG. 12, it may be determined that the ontology could be modified to improve model performance. Thus, from block 1425, the natural language platform may be configured to proceed back to block 1420 to allow for the ontology to be modified. For example, the natural language model may access inputs provided by analysts to modify one or more of the labels or tasks in the ontology.
  • As another example of an iterative loop, it may be determined that rules may be modified to improve performance of the natural language model. Thus, from block 1425, the natural language platform may be configured to proceed back to block 1415, which in some cases is a sub process within the overall process for creating the ontology. To modify some of the rules, the natural language platform may be configured to apply a weight to each of the rules, where the weight represents a degree to which the rules should be heeded in comparison to other rules. For example, a weight of 1.0 for a rule may signify that the rule should always be followed, whereas a weight of 0.5 may signify that the rule should be followed according to a probabilistic formula that may factor in additional statistical parameters, resulting in a composite statistical score to determine how to classify a document. Thus, modifying the rules may also include adjusting any weights to the rules. Modifying the rules may also involve creating new rules or modifying the logical parameters to which the rules should be applied. After any rules are modified at block 1415, the process may again proceed to block 1420, and continue down the process in another iteration.
  • As another example of an iterative loop, and may be determined that new topics should be discovered or topics may need to be refined or modified in some way. Thus, from block 1425, the natural language platform may be configured to proceed back to block 1410 to conduct topic modeling again, which in some cases is a sub process within the overall process for creating the ontology. To modify any topics, the natural language platform may be configured to accept various parameters for discovering new topics. For example, new documents may be accessed, or certain documents may be given more weight. In other cases, the natural language platform may receive inputs to merge certain topics together into a single composite task or label within the ontology. In other cases, the natural language platform may be configured to receive inputs to divide one or more topics into multiple topics. Analysts may be offered displays of various metrics that express levels of agreement or levels of performance of the natural language model when the topics are organized in one configuration versus another configuration. The analysts may then be able to compare how the model may improve if various topics are modified. Example displays for these various interfaces for modifying the topics, the rules, or the ontology may be described further in application (Attorney Docket No. 1402805.00013_IDB013), which again is incorporated herein by reference.
  • In some cases, after modifying any of the topics at block 1410, the process may proceed directly to block 1420 to modify the ontology in light of the newly modified topics. In other cases, and may also be desirable to modify any rules, so the modification iteration may move from block 1410 to block 1415 before proceeding to block 1420.
  • In some embodiments, these iterative processes may cease once performance of the natural language model is determined to satisfy some threshold criterion. At that point, the natural language model may be available for use to analyze untested data at block 1430.
  • Referring to FIG. 15, flowchart 1500 provides an additional variant for generating natural language models, this time including a maintenance process for updating currently operating natural language models, according to some embodiments. As shown, the various functional blocks in flowchart 1400 remain the same, but in addition, after the natural language model is in use to analyze untested data at block 1430, an additional block 1510 is provided to modify the currently running natural language model with updated data. The model may need to be updated periodically in order to adjust to changing language that may occur in certain collections of documents. For example, if the model is analyzing communications on Twitter, new idioms and slang may inevitably be communicated in tweets over time. A model trained six months ago may therefore lack proper training for how to handle any new changes in language.
  • Rather than generate a completely new model, the existing natural language model may simply be retrained with updated data. The retraining may include revising or discovering new topics, modifying or creating new rules, or modifying the ontology structure to account for the updated data. Thus, the natural language platform may be configured to retrain the model with updated data by proceeding to either block 1410, 1415, or 1420, depending on the need. In some embodiments, modifications may include simply conducting one or more rounds of adaptive machine learning at block 1425, first by performing intelligent queuing to generate new annotations for the updated data.
  • In general, generating natural language processing models using the descriptions provided herein allows for: production of the models in a substantially shorter time period than conventional methods (e.g., highly accurate and fast performing models generated in a period of weeks rather than months), highly adaptive models that can evolve with changing input data, maintaining of consistent levels of model performance over time, the ability to leverage existing models and training data to develop domain-specific models with fewer new annotations, and the ability to add new labels and categorizations as requirements change.
  • Referring to FIG. 16, the block diagram illustrates components of a machine 1600, according to some example embodiments, able to read instructions 1624 from a machine-readable medium 1622 (e.g., a non-transitory machine-readable medium, a machine-readable storage medium, a computer-readable storage medium, or any suitable combination thereof) and perform any one or more of the methodologies discussed herein, in whole or in part. Specifically, FIG. 16 shows the machine 1600 in the example form of a computer system (e.g., a computer) within which the instructions 1624 (e.g., software, a program, an application, an applet, an app, or other executable code) for causing the machine 1600 to perform any one or more of the methodologies discussed herein may be executed, in whole or in part.
  • In alternative embodiments, the machine 1600 operates as a standalone device or may be connected (e.g., networked) to other machines. In a networked deployment, the machine 1600 may operate in the capacity of a server machine 110 or a client machine in a server-client network environment, or as a peer machine in a distributed (e.g., peer-to-peer) network environment. The machine 1600 may include hardware, software, or combinations thereof, and may, as example, be a server computer, a client computer, a personal computer (PC), a tablet computer, a laptop computer, a netbook, a cellular telephone, a smartphone, a set-top box (STB), a personal digital assistant (PDA), a web appliance, a network router, a network switch, a network bridge, or any machine capable of executing the instructions 1624, sequentially or otherwise, that specify actions to be taken by that machine. Further, while only a single machine 1600 is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute the instructions 1624 to perform all or part of any one or more of the methodologies discussed herein.
  • The machine 1600 includes a processor 1602 (e.g., a central processing unit (CPU), a graphics processing unit (GPU), a digital signal processor (DSP), an application specific integrated circuit (ASIC), a radio-frequency integrated circuit (RFIC), or any suitable combination thereof), a main memory 1604, and a static memory 1606, which are configured to communicate with each other via a bus 1608. The processor 1602 may contain microcircuits that are configurable, temporarily or permanently, by some or all of the instructions 1624 such that the processor 1602 is configurable to perform any one or more of the methodologies described herein, in whole or in part. For example, a set of one or more microcircuits of the processor 1602 may be configurable to execute one or more modules (e.g., software modules) described herein.
  • The machine 1600 may further include a video display 1610 (e.g., a plasma display panel (PDP), a light emitting diode (LED) display, a liquid crystal display (LCD), a projector, a cathode ray tube (CRT), or any other display capable of displaying graphics or video). The machine 1600 may also include an alphanumeric input device 1612 (e.g., a keyboard or keypad), a cursor control device 1614 (e.g., a mouse, a touchpad, a trackball, a joystick, a motion sensor, an eye tracking device, or other pointing instrument), a storage unit 1616, a signal generation device 1618 (e.g., a sound card, an amplifier, a speaker, a headphone jack, or any suitable combination thereof), and a network interface device 1620.
  • The storage unit 1616 includes the machine-readable medium 1622 (e.g., a tangible and non-transitory machine-readable storage medium) on which are stored the instructions 1624 embodying any one or more of the methodologies or functions described herein, including, for example, any of the descriptions of FIGS. 1A-15. The instructions 1624 may also reside, completely or at least partially, within the main memory 1604, within the processor 1602 (e.g., within the processor's cache memory), or both, before or during execution thereof by the machine 1600. The instructions 1624 may also reside in the static memory 1606.
  • Accordingly, the main memory 1604 and the processor 1602 may be considered machine-readable media 1622 (e.g., tangible and non-transitory machine-readable media). The instructions 1624 may be transmitted or received over a network 1626 via the network interface device 1620. For example, the network interface device 1620 may communicate the instructions 1624 using any one or more transfer protocols (e.g., HTTP). The machine 1600 may also represent example means for performing any of the functions described herein, including the processes described in FIGS. 1A-15.
  • In some example embodiments, the machine 1600 may be a portable computing device, such as a smart phone or tablet computer, and have one or more additional input components (e.g., sensors or gauges) (not shown). Examples of such input components include an image input component (e.g., one or more cameras), an audio input component (e.g., a microphone), a direction input component (e.g., a compass), a location input component (e.g., a GPS receiver), an orientation component (e.g., a gyroscope), a motion detection component (e.g., one or more accelerometers), an altitude detection component (e.g., an altimeter), and a gas detection component (e.g., a gas sensor). Inputs harvested by any one or more of these input components may be accessible and available for use by any of the modules described herein.
  • As used herein, the term “memory” refers to a machine-readable medium 1622 able to store data temporarily or permanently and may be taken to include, but not be limited to, random-access memory (RAM), read-only memory (ROM), buffer memory, flash memory, and cache memory. While the machine-readable medium 1622 is shown in an example embodiment to be a single medium, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database 115, or associated caches and servers) able to store instructions 1624. The term “machine-readable medium” shall also be taken to include any medium, or combination of multiple media, that is capable of storing the instructions 1624 for execution by the machine 1600, such that the instructions 1624, when executed by one or more processors of the machine 1600 (e.g., processor 1602), cause the machine 1600 to perform any one or more of the methodologies described herein, in whole or in part. Accordingly, a “machine-readable medium” refers to a single storage apparatus or device 120 or 130, as well as cloud-based storage systems or storage networks that include multiple storage apparatus or devices 120 or 130. The term “machine-readable medium” shall accordingly be taken to include, but not be limited to, one or more tangible (e.g., non-transitory) data repositories in the form of a solid-state memory, an optical medium, a magnetic medium, or any suitable combination thereof.
  • Furthermore, the machine-readable medium 1622 is non-transitory in that it does not embody a propagating signal. However, labeling the tangible machine-readable medium 1622 as “non-transitory” should not be construed to mean that the medium is incapable of movement; the medium should be considered as being transportable from one physical location to another. Additionally, since the machine-readable medium 1622 is tangible, the medium may be considered to be a machine-readable device.
  • Throughout this specification, plural instances may implement components, operations, or structures described as a single instance. Although individual operations of one or more methods are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently, and nothing requires that the operations be performed in the order illustrated. Structures and functionality presented as separate components in example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein.
  • Certain embodiments are described herein as including logic or a number of components, modules, or mechanisms. Modules may constitute software modules (e.g., code stored or otherwise embodied on a machine-readable medium 1622 or in a transmission medium), hardware modules, or any suitable combination thereof. A “hardware module” is a tangible (e.g., non-transitory) unit capable of performing certain operations and may be configured or arranged in a certain physical manner. In various example embodiments, one or more computer systems (e.g., a standalone computer system, a client computer system, or a server computer system) or one or more hardware modules of a computer system (e.g., a processor 1602 or a group of processors 1602) may be configured by software (e.g., an application or application portion) as a hardware module that operates to perform certain operations as described herein.
  • In some embodiments, a hardware module may be implemented mechanically, electronically, or any suitable combination thereof. For example, a hardware module may include dedicated circuitry or logic that is permanently configured to perform certain operations. For example, a hardware module may be a special-purpose processor, such as a field programmable gate array (FPGA) or an ASIC. A hardware module may also include programmable logic or circuitry that is temporarily configured by software to perform certain operations. For example, a hardware module may include software encompassed within a general-purpose processor 1602 or other programmable processor 1602. It will be appreciated that the decision to implement a hardware module mechanically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) may be driven by cost and time considerations.
  • Hardware modules can provide information to, and receive information from, other hardware modules. Accordingly, the described hardware modules may be regarded as being communicatively coupled. Where multiple hardware modules exist contemporaneously, communications may be achieved through signal transmission (e.g., over appropriate circuits and buses 1608) between or among two or more of the hardware modules. In embodiments in which multiple hardware modules are configured or instantiated at different times, communications between such hardware modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple hardware modules have access. For example, one hardware module may perform an operation and store the output of that operation in a memory device to which it is communicatively coupled. A further hardware module may then, at a later time, access the memory device to retrieve and process the stored output. Hardware modules may also initiate communications with input or output devices, and can operate on a resource (e.g., a collection of information).
  • The various operations of example methods described herein may be performed, at least partially, by one or more processors 1602 that are temporarily configured (e.g., by software) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors 1602 may constitute processor-implemented modules that operate to perform one or more operations or functions described herein. As used herein, “processor-implemented module” refers to a hardware module implemented using one or more processors 1602.
  • Similarly, the methods described herein may be at least partially processor-implemented, a processor 1602 being an example of hardware. For example, at least some of the operations of a method may be performed by one or more processors 1602 or processor-implemented modules. As used herein, “processor-implemented module” refers to a hardware module in which the hardware includes one or more processors 1602. Moreover, the one or more processors 1602 may also operate to support performance of the relevant operations in a “cloud computing” environment or as a “software as a service” (SaaS). For example, at least some of the operations may be performed by a group of computers (as examples of machines 1600 including processors 1602), with these operations being accessible via a network 1626 (e.g., the Internet) and via one or more appropriate interfaces (e.g., an API).
  • The performance of certain operations may be distributed among the one or more processors 1602, not only residing within a single machine 1600, but deployed across a number of machines 1600. In some example embodiments, the one or more processors 1602 or processor-implemented modules may be located in a single geographic location (e.g., within a home environment, an office environment, or a server farm). In other example embodiments, the one or more processors 1602 or processor-implemented modules may be distributed across a number of geographic locations.
  • Unless specifically stated otherwise, discussions herein using words such as “processing,” “computing,” “calculating,” “determining,” “presenting,” “displaying,” or the like may refer to actions or processes of a machine 1600 (e.g., a computer) that manipulates or transforms data represented as physical (e.g., electronic, magnetic, or optical) quantities within one or more memories (e.g., volatile memory, non-volatile memory, or any suitable combination thereof), registers, or other machine components that receive, store, transmit, or display information. Furthermore, unless specifically stated otherwise, the terms “a” or “an” are herein used, as is common in patent documents, to include one or more than one instance. Finally, as used herein, the conjunction “or” refers to a non-exclusive “or,” unless specifically stated otherwise.
  • The present disclosure is illustrative and not limiting. Further modifications will be apparent to one skilled in the art in light of this disclosure and are intended to fall within the scope of the appended claims.

Claims (20)

What is claimed is:
1. A system for generating a natural language model, the system comprising:
at least one memory and at least one processor communicatively coupled to the at least one memory; and
a database module, an application program interface (API) module, a background processing module, and an applications module, each stored on the at least one memory and executable by the at least one processor;
the API module configured to ingest training data representative of documents to be analyzed by the natural language model and to store the training data in the database module;
the background processing module configured to:
generate a hierarchical data structure, the hierarchical data structure comprising at least two topical nodes, wherein the at least two topical nodes represent partitions organized by two or more topical themes among the topical content of the training data within which the training data is to be subdivided into;
select among the training data a plurality of documents to be annotated;
generate at least one annotation prompt for each document among the plurality of documents to be annotated, said annotation prompt configured to elicit an annotation about said document indicating which node among the at least two topical nodes of the hierarchal data structure said document is to be classified into;
the application module configured to cause display of the at least one annotation prompt for each document among the plurality of documents to be annotated;
the API module further configured to receive for each document among the plurality of documents to be annotated, the annotation in response to the displayed annotation prompt; and
the background processing module further configured to generate the natural language model using an adaptive machine learning process configured to determine, among the received annotations, patterns for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
2. The system of claim 1, wherein the background processing module is further configured to test performance of the natural language model using a subset of the documents among the training data that received annotations.
3. The system of claim 2, wherein the background processing module is further configured to:
compute a performance metric of the natural language model, based on results of the testing; and
determine whether the natural language model satisfies at least one performance criterion based on the computed performance metric.
4. The system of claim 3, wherein the background processing module is further configured to:
performing one or more optimization techniques configured to improve performance of the natural language platform, in response to determining that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric.
5. The system of claim 4, wherein the one or more optimization techniques comprises at least one of: a feature selection process, a padding and rebalancing process of the natural language model, a pruning process of the natural language model, a feature discovery process, a smoothing process of the natural language model, or a model interpolation process.
6. The system of claim 3, wherein the background processing module is further configured to:
determine that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric;
in response to said determining:
identify a topical node among the two or more topical nodes of the hierarchal data structure that the natural language model fails to accurately categorize documents into;
select a second plurality of documents to be annotated, the second plurality comprising documents associated with said topical node that the natural language model failed to accurately categorize documents into; and
generate a second set of at least one annotation prompt for each document among the second plurality of documents to be annotated, said annotation prompt among the second set configured to elicit an annotation about said document to improve the natural language model in accurately categorizing documents into said topical node;
wherein the applications module is further configured to cause display of the second set of the at least one annotation prompt for each document among the second plurality of documents to be annotated;
wherein the API module is further configured to receive for each document among the second plurality of documents to be annotated, a second set of annotations in response to the second set of displayed annotation prompts; and
wherein the background processing module is further configured to generate a refined natural language model using the adaptive machine learning process and based on the hierarchical data structure, the training data and the second set of annotations.
7. The system of claim 1, wherein generating the hierarchical data structure comprises:
performing a topic modeling process configured to identify two or more topics among the content of the training data that is configured to define the two or more topical nodes of the hierarchical data structure.
8. The system of claim 1, wherein the background processing module is further configured to access one or more rules configured to instruct the natural language model how to categorize one or more documents into the two or more nodes of the hierarchical data structure.
9. The system of claim 8, wherein generating the hierarchical data structure comprises:
conducting a rules generation process configured to evaluate logical consistency among the one or more rules.
10. The system of claim 1, wherein generating the hierarchical data structure comprises:
generating at least one annotation prompt for each topical node among the two or more topical nodes in the hierarchical data structure, said annotation prompt configured to elicit an annotation about said topical node indicating a level of accuracy of placement of the node within the hierarchical data structure;
causing display of the at least one annotation prompt for each topical node; and
receiving for each topical node, the annotation in response to the displayed annotation prompt.
11. The system of claim 10, wherein the background processing module is further configured to evaluate performance of the hierarchical data structure based on the annotations.
12. The system of claim 11, wherein the background processing module is further configured to determine that the hierarchical data structure fails to satisfy at least one performance criterion in response to the evaluating; and
modify a logical relationship among the two or more topical nodes based on the annotations and in response to determining that the data structure fails to satisfy the at least one performance criterion.
13. The system of claim 9, wherein the API module is further configured to receive a training guideline based on the annotations to the nodes, the training guideline configured to provide instructions to an annotator for answering one or more annotation prompts for each document among the plurality of documents to be annotated.
14. The system of claim 1, wherein the hierarchical data structure comprises at least a third topical node and a fourth topical node, wherein the third and fourth topical nodes both represent sub-partitions within the topical theme of the first node and organized by a third and fourth topical theme, respectively, among the topical content of the training data within which the training data is to be subdivided into.
15. A method for generating a natural language platform system configured to generate a natural language model, the method comprising:
deriving from an analogous natural language platform system, parameters configured to optimize performance of said analogous system to generate an analogous natural language model configured to analyze similar but not identical documents as the natural language model;
interpolating said parameters to be optimized for documents to be analyzed by the natural language model; and
implementing the interpolated parameters in the natural language platform system such that the interpolated parameters are configured to generate the natural language model.
16. A system for generating natural language models, the system comprising:
a full service natural language platform geographically located at a remote host location and configured to:
receive training data through a network connection;
train a natural language model based on the received training data; and
generate predictions about untested data using the natural language model; and
a connector module geographically located at a local client host location and communicatively coupled to the full service natural language platform through the network connection and configured to:
access the training data stored in a client data store at the local client host location;
format the training data in a uniform manner;
transmit the training data to the full service natural language platform through the network connection;
receive the predictions about the untested data; and
store the predictions about the untested data in a memory at the local client host location.
17. The system of claim 16, further comprising a text extraction module communicatively coupled to the connector module and configured to attach to the client data store and extract textual data for use as the training data.
18. The system of claim 16, wherein the full service natural language platform is further configured to:
generate a hierarchical data structure, the hierarchical data structure comprising at least two topical nodes, wherein the at least two topical nodes represent partitions organized by two or more topical themes among the topical content of the training data within which the training data is to be subdivided into;
select among the training data a plurality of documents to be annotated; and
generate at least one annotation prompt for each document among the plurality of documents to be annotated, said annotation prompt configured to elicit an annotation about said document indicating which node among the at least two topical nodes of the hierarchal data structure said document is to be classified into.
19. The system of claim 18, wherein the full service natural language platform is further configured to:
cause display of the at least one annotation prompt for each document among the plurality of documents to be annotated;
receive for each document among the plurality of documents to be annotated, the annotation in response to the displayed annotation prompt; and
generate the natural language model using an adaptive machine learning process configured to determine, among the received annotations, patterns for how the documents in the training data are to be subdivided according to the at least two topical nodes of the hierarchical data structure.
20. The system of claim 19, wherein the full service natural language platform is further configured to:
test performance of the natural language model using a subset of the documents among the training data that received annotations;
compute a performance metric of the natural language model, based on results of the testing;
determine that the natural language platform fails to satisfy the at least one performance criterion based on the computed performance metric;
in response to said determining:
identify a topical node among the two or more topical nodes of the hierarchal data structure that the natural language model fails to accurately categorize documents into;
select a second plurality of documents to be annotated, the second plurality comprising documents associated with said topical node that the natural language model failed to accurately categorize documents into;
generate a second set of at least one annotation prompt for each document among the second plurality of documents to be annotated, said annotation prompt among the second set configured to elicit an annotation about said document to improve the natural language model in accurately categorizing documents into said topical node;
cause display of the second set of the at least one annotation prompt for each document among the second plurality of documents to be annotated;
receive for each document among the second plurality of documents to be annotated, a second set of annotations in response to the second set of displayed annotation prompts; and
generate a refined natural language model using the adaptive machine learning process and based on the hierarchical data structure, the training data and the second set of annotations.
US16/289,481 2014-12-09 2019-02-28 Architectures for natural language processing Abandoned US20200034737A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US16/289,481 US20200034737A1 (en) 2014-12-09 2019-02-28 Architectures for natural language processing
US17/166,493 US20210232762A1 (en) 2014-12-09 2021-02-03 Architectures for natural language processing

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US201462089736P 2014-12-09 2014-12-09
US201462089745P 2014-12-09 2014-12-09
US201462089742P 2014-12-09 2014-12-09
US201462089747P 2014-12-09 2014-12-09
US201514964518A 2015-12-09 2015-12-09
US16/289,481 US20200034737A1 (en) 2014-12-09 2019-02-28 Architectures for natural language processing

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US201514964518A Continuation 2014-12-09 2015-12-09

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/166,493 Continuation US20210232762A1 (en) 2014-12-09 2021-02-03 Architectures for natural language processing

Publications (1)

Publication Number Publication Date
US20200034737A1 true US20200034737A1 (en) 2020-01-30

Family

ID=56094482

Family Applications (20)

Application Number Title Priority Date Filing Date
US14/964,510 Abandoned US20160162569A1 (en) 2014-12-09 2015-12-09 Methods and systems for improving machine learning performance
US14/964,522 Abandoned US20160162458A1 (en) 2014-12-09 2015-12-09 Graphical systems and methods for human-in-the-loop machine intelligence
US14/964,511 Active - Reinstated US9495345B2 (en) 2014-12-09 2015-12-09 Methods and systems for modeling complex taxonomies with natural language understanding
US14/964,528 Abandoned US20160162464A1 (en) 2014-12-09 2015-12-09 Techniques for combining human and machine learning in natural language processing
US14/964,520 Abandoned US20160162457A1 (en) 2014-12-09 2015-12-09 Optimization techniques for artificial intelligence
US14/964,517 Active US10127214B2 (en) 2014-12-09 2015-12-09 Methods for generating natural language processing systems
US15/294,156 Abandoned US20170235813A1 (en) 2014-12-09 2016-10-14 Methods and systems for modeling complex taxonomies with natural language understanding
US16/125,343 Abandoned US20190243886A1 (en) 2014-12-09 2018-09-07 Methods and systems for improving machine learning performance
US16/181,102 Abandoned US20190303428A1 (en) 2014-12-09 2018-11-05 Methods for generating natural language processing systems
US16/185,843 Abandoned US20190311024A1 (en) 2014-12-09 2018-11-09 Techniques for combining human and machine learning in natural language processing
US16/197,190 Abandoned US20190311025A1 (en) 2014-12-09 2018-11-20 Methods and systems for modeling complex taxonomies with natural language understanding
US16/198,453 Abandoned US20200234002A1 (en) 2014-12-09 2018-11-21 Optimization techniques for artificial intelligence
US16/289,481 Abandoned US20200034737A1 (en) 2014-12-09 2019-02-28 Architectures for natural language processing
US16/749,689 Abandoned US20200184146A1 (en) 2014-12-09 2020-01-22 Techniques for combining human and machine learning in natural language processing
US16/796,812 Abandoned US20210150130A1 (en) 2014-12-09 2020-02-20 Methods for generating natural language processing systems
US16/811,737 Active 2036-04-16 US11599714B2 (en) 2014-12-09 2020-03-06 Methods and systems for modeling complex taxonomies with natural language understanding
US17/119,902 Active US11288444B2 (en) 2014-12-09 2020-12-11 Optimization techniques for artificial intelligence
US17/131,486 Abandoned US20210232761A1 (en) 2014-12-09 2020-12-22 Methods and systems for improving machine learning performance
US17/166,493 Abandoned US20210232762A1 (en) 2014-12-09 2021-02-03 Architectures for natural language processing
US17/182,178 Abandoned US20210232763A1 (en) 2014-12-09 2021-02-22 Graphical systems and methods for human-in-the-loop machine intelligence

Family Applications Before (12)

Application Number Title Priority Date Filing Date
US14/964,510 Abandoned US20160162569A1 (en) 2014-12-09 2015-12-09 Methods and systems for improving machine learning performance
US14/964,522 Abandoned US20160162458A1 (en) 2014-12-09 2015-12-09 Graphical systems and methods for human-in-the-loop machine intelligence
US14/964,511 Active - Reinstated US9495345B2 (en) 2014-12-09 2015-12-09 Methods and systems for modeling complex taxonomies with natural language understanding
US14/964,528 Abandoned US20160162464A1 (en) 2014-12-09 2015-12-09 Techniques for combining human and machine learning in natural language processing
US14/964,520 Abandoned US20160162457A1 (en) 2014-12-09 2015-12-09 Optimization techniques for artificial intelligence
US14/964,517 Active US10127214B2 (en) 2014-12-09 2015-12-09 Methods for generating natural language processing systems
US15/294,156 Abandoned US20170235813A1 (en) 2014-12-09 2016-10-14 Methods and systems for modeling complex taxonomies with natural language understanding
US16/125,343 Abandoned US20190243886A1 (en) 2014-12-09 2018-09-07 Methods and systems for improving machine learning performance
US16/181,102 Abandoned US20190303428A1 (en) 2014-12-09 2018-11-05 Methods for generating natural language processing systems
US16/185,843 Abandoned US20190311024A1 (en) 2014-12-09 2018-11-09 Techniques for combining human and machine learning in natural language processing
US16/197,190 Abandoned US20190311025A1 (en) 2014-12-09 2018-11-20 Methods and systems for modeling complex taxonomies with natural language understanding
US16/198,453 Abandoned US20200234002A1 (en) 2014-12-09 2018-11-21 Optimization techniques for artificial intelligence

Family Applications After (7)

Application Number Title Priority Date Filing Date
US16/749,689 Abandoned US20200184146A1 (en) 2014-12-09 2020-01-22 Techniques for combining human and machine learning in natural language processing
US16/796,812 Abandoned US20210150130A1 (en) 2014-12-09 2020-02-20 Methods for generating natural language processing systems
US16/811,737 Active 2036-04-16 US11599714B2 (en) 2014-12-09 2020-03-06 Methods and systems for modeling complex taxonomies with natural language understanding
US17/119,902 Active US11288444B2 (en) 2014-12-09 2020-12-11 Optimization techniques for artificial intelligence
US17/131,486 Abandoned US20210232761A1 (en) 2014-12-09 2020-12-22 Methods and systems for improving machine learning performance
US17/166,493 Abandoned US20210232762A1 (en) 2014-12-09 2021-02-03 Architectures for natural language processing
US17/182,178 Abandoned US20210232763A1 (en) 2014-12-09 2021-02-22 Graphical systems and methods for human-in-the-loop machine intelligence

Country Status (1)

Country Link
US (20) US20160162569A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022155539A1 (en) * 2021-01-14 2022-07-21 Virtuosource Llc A rules-based decision support system for assessment of digital content involving natural language
US20220253607A1 (en) * 2021-02-08 2022-08-11 International Business Machines Corporation Question based chatbot generator from web api specifications
US11443082B2 (en) 2020-05-27 2022-09-13 Accenture Global Solutions Limited Utilizing deep learning and natural language processing to convert a technical architecture diagram into an interactive technical architecture diagram
US11556788B2 (en) * 2020-06-15 2023-01-17 International Business Machines Corporation Text-based response environment action selection

Families Citing this family (319)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10002189B2 (en) 2007-12-20 2018-06-19 Apple Inc. Method and apparatus for searching using an active ontology
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US10430506B2 (en) 2012-12-10 2019-10-01 International Business Machines Corporation Utilizing classification and text analytics for annotating documents to allow quick scanning
KR102380145B1 (en) 2013-02-07 2022-03-29 애플 인크. Voice trigger for a digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
CN110442699A (en) 2013-06-09 2019-11-12 苹果公司 Operate method, computer-readable medium, electronic equipment and the system of digital assistants
KR101749009B1 (en) 2013-08-06 2017-06-19 애플 인크. Auto-activating smart responses based on activities from remote devices
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
WO2015184186A1 (en) 2014-05-30 2015-12-03 Apple Inc. Multi-command single utterance input method
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US10169826B1 (en) 2014-10-31 2019-01-01 Intuit Inc. System and method for generating explanations for tax calculations
BR112017021842A2 (en) 2014-11-20 2018-07-10 Univ Arizona State systems and methods for generating liquid water from the air
US10387970B1 (en) 2014-11-25 2019-08-20 Intuit Inc. Systems and methods for analyzing and generating explanations for changes in tax return results
US20160162569A1 (en) 2014-12-09 2016-06-09 Idibon, Inc. Methods and systems for improving machine learning performance
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US10872384B1 (en) 2015-03-30 2020-12-22 Intuit Inc. System and method for generating explanations for year-over-year tax changes
US10460227B2 (en) 2015-05-15 2019-10-29 Apple Inc. Virtual assistant in a communication session
US10200824B2 (en) 2015-05-27 2019-02-05 Apple Inc. Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10318591B2 (en) * 2015-06-02 2019-06-11 International Business Machines Corporation Ingesting documents using multiple ingestion pipelines
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
EP3311311A1 (en) * 2015-06-18 2018-04-25 Aware, Inc. Automatic entity resolution with rules detection and generation system
US20160378747A1 (en) 2015-06-29 2016-12-29 Apple Inc. Virtual assistant for media playback
US10607298B1 (en) 2015-07-30 2020-03-31 Intuit Inc. System and method for indicating sections of electronic tax forms for which narrative explanations can be presented
US10140983B2 (en) * 2015-08-28 2018-11-27 International Business Machines Corporation Building of n-gram language model for automatic speech recognition (ASR)
US10331312B2 (en) 2015-09-08 2019-06-25 Apple Inc. Intelligent automated assistant in a media environment
US10740384B2 (en) 2015-09-08 2020-08-11 Apple Inc. Intelligent automated assistant for media search and playback
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US20180241848A1 (en) * 2015-09-11 2018-08-23 Hewlett Packard Enterprise Development Lp Human-readable cloud structures
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US20170098172A1 (en) 2015-10-06 2017-04-06 Evolv Technologies, Inc. Training Artificial Intelligence
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10956666B2 (en) 2015-11-09 2021-03-23 Apple Inc. Unconventional virtual assistant interactions
US10102201B2 (en) * 2015-11-30 2018-10-16 Soundhound, Inc. Natural language module store
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US9760556B1 (en) 2015-12-11 2017-09-12 Palantir Technologies Inc. Systems and methods for annotating and linking electronic documents
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10346546B2 (en) * 2015-12-23 2019-07-09 Oath Inc. Method and system for automatic formality transformation
US10740573B2 (en) 2015-12-23 2020-08-11 Oath Inc. Method and system for automatic formality classification
US10572524B2 (en) * 2016-02-29 2020-02-25 Microsoft Technology Licensing, Llc Content categorization
US9747143B1 (en) * 2016-03-30 2017-08-29 International Business Machines Corporation Multi platform based event processing
TWI718284B (en) 2016-04-07 2021-02-11 美商零質量純水股份有限公司 Solar thermal unit
RU2632143C1 (en) * 2016-04-11 2017-10-02 Общество С Ограниченной Ответственностью "Яндекс" Training method of rating module using the training selection with the interference labels
RU2628431C1 (en) * 2016-04-12 2017-08-16 Общество с ограниченной ответственностью "Аби Продакшн" Selection of text classifier parameter based on semantic characteristics
GB201608576D0 (en) * 2016-05-16 2016-06-29 Pro Intex It Ltd Functional behaviour test system and method
CN115228249B (en) 2016-05-20 2024-09-13 环球源公司 System and method for water extraction control
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
RU2637883C1 (en) * 2016-06-20 2017-12-07 Общество С Ограниченной Ответственностью "Яндекс" Method of establishing training object for training machine training algorithm
US10635704B2 (en) * 2016-07-15 2020-04-28 Cisco Technology, Inc. Automatic ontology generation for internet of things applications
US11055794B1 (en) 2016-07-27 2021-07-06 Intuit Inc. Methods, systems and computer program products for estimating likelihood of qualifying for benefit
US10872315B1 (en) 2016-07-27 2020-12-22 Intuit Inc. Methods, systems and computer program products for prioritization of benefit qualification questions
US10762472B1 (en) 2016-07-27 2020-09-01 Intuit Inc. Methods, systems and computer program products for generating notifications of benefit qualification change
US10769592B1 (en) 2016-07-27 2020-09-08 Intuit Inc. Methods, systems and computer program products for generating explanations for a benefit qualification change
US10579941B2 (en) * 2016-09-01 2020-03-03 Facebook, Inc. Systems and methods for recommending pages
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10650086B1 (en) * 2016-09-27 2020-05-12 Palantir Technologies Inc. Systems, methods, and framework for associating supporting data in word processing
US12020334B2 (en) * 2016-10-26 2024-06-25 Intuit Inc. Methods, systems and computer program products for generating and presenting explanations for tax questions
WO2018081633A1 (en) 2016-10-28 2018-05-03 Roam Analytics, Inc. Semantic parsing engine
US11507629B2 (en) 2016-10-28 2022-11-22 Parexel International, Llc Dataset networking and database modeling
US10963789B2 (en) 2016-11-28 2021-03-30 Conduent Business Services, Llc Long-term memory networks for knowledge extraction from text and publications
US10878488B2 (en) * 2016-11-29 2020-12-29 Samsung Electronics Co., Ltd. Electronic apparatus and method for summarizing content thereof
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10679008B2 (en) * 2016-12-16 2020-06-09 Microsoft Technology Licensing, Llc Knowledge base for analysis of text
US11373752B2 (en) * 2016-12-22 2022-06-28 Palantir Technologies Inc. Detection of misuse of a benefit system
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
US11954098B1 (en) * 2017-02-03 2024-04-09 Thomson Reuters Enterprise Centre Gmbh Natural language processing system and method for documents
US10298757B2 (en) * 2017-02-23 2019-05-21 Accenture Global Solutions Limited Integrated service centre support
US20180268004A1 (en) * 2017-03-17 2018-09-20 Microsoft Technology Licensing, Llc Rule hierarchies for graph adaptation
US11514058B2 (en) 2017-03-17 2022-11-29 Microsoft Technology Licensing, Llc Context rules for a graph
US11270686B2 (en) 2017-03-28 2022-03-08 International Business Machines Corporation Deep language and acoustic modeling convergence and cross training
US10679009B2 (en) * 2017-04-21 2020-06-09 Tata Consultancy Services Limited System and method for belief based human-bot conversation
US11880746B1 (en) * 2017-04-26 2024-01-23 Hrb Innovations, Inc. Interface for artificial intelligence training
US10963501B1 (en) * 2017-04-29 2021-03-30 Veritas Technologies Llc Systems and methods for generating a topic tree for digital information
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
DK180048B1 (en) 2017-05-11 2020-02-04 Apple Inc. MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK201770427A1 (en) 2017-05-12 2018-12-20 Apple Inc. Low-latency intelligent automated assistant
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK201770411A1 (en) 2017-05-15 2018-12-20 Apple Inc. Multi-modal interfaces
DK179560B1 (en) 2017-05-16 2019-02-18 Apple Inc. Far-field extension for digital assistant services
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
US20180336275A1 (en) 2017-05-16 2018-11-22 Apple Inc. Intelligent automated assistant for media exploration
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
US20180336892A1 (en) 2017-05-16 2018-11-22 Apple Inc. Detecting a trigger of a digital assistant
US11520610B2 (en) 2017-05-18 2022-12-06 Peloton Interactive Inc. Crowdsourced on-boarding of digital assistant operations
US11340925B2 (en) 2017-05-18 2022-05-24 Peloton Interactive Inc. Action recipes for a crowdsourced digital assistant system
US11056105B2 (en) 2017-05-18 2021-07-06 Aiqudo, Inc Talk back from actions in applications
US11043206B2 (en) 2017-05-18 2021-06-22 Aiqudo, Inc. Systems and methods for crowdsourced actions and commands
US10466963B2 (en) 2017-05-18 2019-11-05 Aiqudo, Inc. Connecting multiple mobile devices to a smart home assistant account
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US10949502B2 (en) 2017-06-02 2021-03-16 Thinkspan, LLC Universal data scaffold based data management platform
AU2018300250B2 (en) 2017-07-14 2024-04-18 Source Global, PBC Systems for controlled treatment of water with ozone and related methods therefor
CN109272003A (en) * 2017-07-17 2019-01-25 华东师范大学 A kind of method and apparatus for eliminating unknown error in deep learning model
US11139048B2 (en) 2017-07-18 2021-10-05 Analytics For Life Inc. Discovering novel features to use in machine learning techniques, such as machine learning techniques for diagnosing medical conditions
US11062792B2 (en) 2017-07-18 2021-07-13 Analytics For Life Inc. Discovering genomes to use in machine learning techniques
CN110019770A (en) * 2017-07-24 2019-07-16 华为技术有限公司 The method and apparatus of train classification models
US11531998B2 (en) * 2017-08-30 2022-12-20 Qualtrics, Llc Providing a conversational digital survey by generating digital survey questions based on digital survey responses
US11384517B2 (en) 2017-09-05 2022-07-12 Source Global, PBC Systems and methods to produce liquid water extracted from air
AU2018329665B2 (en) 2017-09-05 2023-11-16 Source Global, PBC Systems and methods for managing production and distribution of liquid water extracted from air
US10565444B2 (en) 2017-09-07 2020-02-18 International Business Machines Corporation Using visual features to identify document sections
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US20190102697A1 (en) * 2017-10-02 2019-04-04 International Business Machines Corporation Creating machine learning models from structured intelligence databases
US11555421B2 (en) 2017-10-06 2023-01-17 Source Global, PBC Systems for generating water with waste heat and related methods therefor
KR102365621B1 (en) * 2017-10-20 2022-02-21 구글 엘엘씨 Capturing detailed structures in patient-physician conversations for use in clinical documentation
CN107832305A (en) * 2017-11-28 2018-03-23 百度在线网络技术(北京)有限公司 Method and apparatus for generating information
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US11281997B2 (en) 2017-12-06 2022-03-22 Source Global, PBC Systems for constructing hierarchical training data sets for use with machine-learning and related methods therefor
US20190179883A1 (en) * 2017-12-08 2019-06-13 International Business Machines Corporation Evaluating textual annotation model performance
US10979457B2 (en) 2017-12-20 2021-04-13 Check Point Public Cloud Security Ltd Cloud security assessment system using near-natural language compliance rules
US10803108B2 (en) * 2017-12-20 2020-10-13 International Business Machines Corporation Facilitation of domain and client-specific application program interface recommendations
US10942955B2 (en) * 2017-12-21 2021-03-09 Shanghai Xiaoi Robot Technology Co., Ltd. Questioning and answering method, method for generating questioning and answering system, and method for modifying questioning and answering system
US10942954B2 (en) * 2017-12-22 2021-03-09 International Business Machines Corporation Dataset adaptation for high-performance in specific natural language processing tasks
US20190197433A1 (en) * 2017-12-22 2019-06-27 Wipro Limited Methods for adaptive information extraction through adaptive learning of human annotators and devices thereof
US10599783B2 (en) * 2017-12-26 2020-03-24 International Business Machines Corporation Automatically suggesting a temporal opportunity for and assisting a writer in writing one or more sequel articles via artificial intelligence
US11768852B2 (en) * 2017-12-27 2023-09-26 Marlabs Incorporated System and method for data analysis and presentation of data
US10963495B2 (en) 2017-12-29 2021-03-30 Aiqudo, Inc. Automated discourse phrase discovery for generating an improved language model of a digital assistant
US10963499B2 (en) * 2017-12-29 2021-03-30 Aiqudo, Inc. Generating command-specific language model discourses for digital assistant interpretation
US10929613B2 (en) 2017-12-29 2021-02-23 Aiqudo, Inc. Automated document cluster merging for topic-based digital assistant interpretation
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US11160223B2 (en) 2018-02-18 2021-11-02 Source Global, PBC Systems for generating water for a container farm and related methods therefor
US11449762B2 (en) 2018-02-20 2022-09-20 Pearson Education, Inc. Real time development of auto scoring essay models for custom created prompts
EP3756148A4 (en) * 2018-02-20 2021-11-17 Pearson Education, Inc. Systems and methods for automated evaluation model customization
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10579737B2 (en) * 2018-03-06 2020-03-03 Adobe Inc. Natural language image editing annotation framework
US10963972B1 (en) * 2018-03-06 2021-03-30 Wells Fargo Bank, N.A. Adaptive life advisor system
US10685655B2 (en) 2018-03-07 2020-06-16 International Business Machines Corporation Leveraging natural language processing
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US11474836B2 (en) 2018-03-13 2022-10-18 Microsoft Technology Licensing, Llc Natural language to API conversion
US10747560B2 (en) * 2018-03-20 2020-08-18 Microsoft Technology Licensing, Llc Computerized task guidance across devices and applications
RU2691855C1 (en) * 2018-03-23 2019-06-18 Общество с ограниченной ответственностью "Аби Продакшн" Training classifiers used to extract information from natural language texts
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US20190332789A1 (en) * 2018-04-27 2019-10-31 Microsoft Technology Licensing, Llc Hierarchical access rights and role based access
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
EP3791389A4 (en) * 2018-05-08 2022-01-26 3M Innovative Properties Company Hybrid batch and live natural language processing
WO2019217974A1 (en) 2018-05-11 2019-11-14 Zero Mass Water, Inc. Systems for generating water using exogenously generated heat, exogenously generated electricity, and exhaust process fluids and related methods therefor
US10951482B2 (en) 2018-05-16 2021-03-16 Microsoft Technology Licensing, Llc Device identification on a building automation control network
US11281995B2 (en) * 2018-05-21 2022-03-22 International Business Machines Corporation Finding optimal surface for hierarchical classification task on an ontology
US11456915B2 (en) 2018-05-21 2022-09-27 Microsoft Technology Licensing, Llc Device model templates
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
US10783328B2 (en) * 2018-06-04 2020-09-22 International Business Machines Corporation Semi-automatic process for creating a natural language processing resource
US10629205B2 (en) * 2018-06-12 2020-04-21 International Business Machines Corporation Identifying an accurate transcription from probabilistic inputs
US11232255B2 (en) * 2018-06-13 2022-01-25 Adobe Inc. Generating digital annotations for evaluating and training automatic electronic document annotation models
CN108733359B (en) * 2018-06-14 2020-12-25 北京航空航天大学 Automatic generation method of software program
US20190385711A1 (en) * 2018-06-19 2019-12-19 Ellipsis Health, Inc. Systems and methods for mental health assessment
US10664472B2 (en) * 2018-06-27 2020-05-26 Bitdefender IPR Management Ltd. Systems and methods for translating natural language sentences into database queries
US11456082B2 (en) * 2018-07-03 2022-09-27 International Business Machines Corporation Patient engagement communicative strategy recommendation
US10839618B2 (en) 2018-07-12 2020-11-17 Honda Motor Co., Ltd. Applied artificial intelligence for natural language processing automotive reporting system
US11399030B2 (en) * 2018-07-23 2022-07-26 Kyndryl, Inc. Ontology based control of access to resources in a computing system
EP3827339A1 (en) * 2018-07-23 2021-06-02 Google LLC Intelligent home screen of cloud-based content management platform
US11144581B2 (en) * 2018-07-26 2021-10-12 International Business Machines Corporation Verifying and correcting training data for text classification
US11636333B2 (en) 2018-07-26 2023-04-25 Tesla, Inc. Optimizing neural network structures for embedded systems
US11392794B2 (en) * 2018-09-10 2022-07-19 Ca, Inc. Amplification of initial training data
US11183195B2 (en) * 2018-09-27 2021-11-23 Snackable Inc. Audio content processing systems and methods
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
CN110969330A (en) * 2018-09-30 2020-04-07 曹西军 Enterprise competitive barrier assessment method and system
AU2019359894A1 (en) 2018-10-19 2021-06-10 Source Global, PBC Systems and methods for generating liquid water using highly efficient techniques that optimize production
US20200124566A1 (en) 2018-10-22 2020-04-23 Zero Mass Water, Inc. Systems and methods for detecting and measuring oxidizing compounds in test fluids
US10943185B1 (en) 2018-10-23 2021-03-09 Bank Of America Corporation Supervised machine-learning training platform with isolated update testing and bi-directional update propogation
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US10979448B2 (en) * 2018-11-02 2021-04-13 KnowBe4, Inc. Systems and methods of cybersecurity attack simulation for incident response training and awareness
US10341491B1 (en) * 2018-11-26 2019-07-02 Capital One Services, Llc Identifying unreported issues through customer service interactions and website analytics
US10956671B2 (en) 2018-11-30 2021-03-23 International Business Machines Corporation Supervised machine learning models of documents
US12008439B2 (en) * 2018-11-30 2024-06-11 Jpmorgan Chase Bank, N.A. Methods for sharing machine learning based web service models
US11216892B1 (en) * 2018-12-06 2022-01-04 Meta Platforms, Inc. Classifying and upgrading a content item to a life event item
US11093708B2 (en) * 2018-12-13 2021-08-17 Software Ag Adaptive human to machine interaction using machine learning
EP3903241A4 (en) 2018-12-24 2022-09-14 Roam Analytics, Inc. Constructing a knowledge graph employing multiple subgraphs and a linking layer including multiple linking nodes
JP7376593B2 (en) * 2018-12-31 2023-11-08 インテル・コーポレーション Security system using artificial intelligence
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11328007B2 (en) * 2019-02-04 2022-05-10 International Business Machines Corporation Generating a domain-specific phrasal dictionary
US11676043B2 (en) 2019-03-04 2023-06-13 International Business Machines Corporation Optimizing hierarchical classification with adaptive node collapses
US11157702B2 (en) * 2019-03-06 2021-10-26 International Business Machines Corporation Utilizing varying coordinates related to a target event to provide contextual outputs
US11263391B2 (en) * 2019-03-11 2022-03-01 Parexel International, Llc Methods, apparatus and systems for annotation of text documents
US20200294083A1 (en) * 2019-03-12 2020-09-17 Jacob M. Del Hagen Interface method and system for enabling an advertisement sponsor to input data concerning leads generated in response to advertisements
US11645110B2 (en) * 2019-03-13 2023-05-09 International Business Machines Corporation Intelligent generation and organization of user manuals
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11715030B2 (en) 2019-03-29 2023-08-01 Red Hat, Inc. Automatic object optimization to accelerate machine learning training
SG11202112327VA (en) 2019-04-03 2021-12-30 Singularity Systems Inc Hierarchical machine learning architecture including master engine supported by distributed light-weight real-time edge engines
CN113924582A (en) * 2019-04-04 2022-01-11 美国元初智能公司 Machine learning processing pipeline optimization
US10614345B1 (en) 2019-04-12 2020-04-07 Ernst & Young U.S. Llp Machine learning based extraction of partition objects from electronic documents
BR112021021014A2 (en) 2019-04-22 2021-12-14 Source Global Pbc Air drying system and method by water vapor adsorption to generate liquid water from air
US11392757B2 (en) * 2019-04-26 2022-07-19 Figure Eight Technologies, Inc. Management of annotation jobs
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
DK201970509A1 (en) 2019-05-06 2021-01-15 Apple Inc Spoken notifications
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
US11748613B2 (en) * 2019-05-10 2023-09-05 Baidu Usa Llc Systems and methods for large scale semantic indexing with deep level-wise extreme multi-label learning
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11277453B2 (en) * 2019-05-24 2022-03-15 International Business Machines Corporation Media communication management
DK201970511A1 (en) 2019-05-31 2021-02-15 Apple Inc Voice identification in digital assistant systems
DK180129B1 (en) 2019-05-31 2020-06-02 Apple Inc. User activity shortcut suggestions
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
US11468890B2 (en) 2019-06-01 2022-10-11 Apple Inc. Methods and user interfaces for voice-based control of electronic devices
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11113518B2 (en) 2019-06-28 2021-09-07 Eygs Llp Apparatus and methods for extracting data from lineless tables using Delaunay triangulation and excess edge removal
US10489454B1 (en) * 2019-06-28 2019-11-26 Capital One Services, Llc Indexing a dataset based on dataset tags and an ontology
US11531703B2 (en) 2019-06-28 2022-12-20 Capital One Services, Llc Determining data categorizations based on an ontology and a machine-learning model
US11354504B2 (en) * 2019-07-10 2022-06-07 International Business Machines Corporation Multi-lingual action identification
US20210026897A1 (en) * 2019-07-23 2021-01-28 Microsoft Technology Licensing, Llc Topical clustering and notifications for driving resource collaboration
US11521019B2 (en) 2019-08-06 2022-12-06 Bank Of America Corporation Systems and methods for incremental learning and autonomous model reconfiguration in regulated AI systems
US11188517B2 (en) 2019-08-09 2021-11-30 International Business Machines Corporation Annotation assessment and ground truth construction
US11294884B2 (en) 2019-08-09 2022-04-05 International Business Machines Corporation Annotation assessment and adjudication
US11915465B2 (en) 2019-08-21 2024-02-27 Eygs Llp Apparatus and methods for converting lineless tables into lined tables using generative adversarial networks
US12026630B2 (en) 2019-09-23 2024-07-02 Adaptai Ltd. Artificial intelligence methods and systems for improving user adaptability using textual communication
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
US11442944B2 (en) 2019-10-18 2022-09-13 Thinkspan, LLC Algorithmic suggestions based on a universal data scaffold
WO2021077038A1 (en) 2019-10-18 2021-04-22 Taylor Brian Samuel Scalable scaffolding and bundled data
CN114651305A (en) 2019-11-18 2022-06-21 谷歌有限责任公司 Automatic mining of real world audio training data
US10810709B1 (en) 2019-11-21 2020-10-20 Eygs Llp Systems and methods for improving the quality of text documents using artificial intelligence
US10922476B1 (en) * 2019-12-13 2021-02-16 Microsoft Technology Licensing, Llc Resource-efficient generation of visual layout information associated with network-accessible documents
TWI798513B (en) * 2019-12-20 2023-04-11 國立清華大學 Training method of natural language corpus for the decision making model of machine learning
CN111104514B (en) * 2019-12-23 2023-04-25 北京百度网讯科技有限公司 Training method and device for document tag model
US11775862B2 (en) * 2020-01-14 2023-10-03 Microsoft Technology Licensing, Llc Tracking provenance in data science scripts
US20210233008A1 (en) * 2020-01-28 2021-07-29 Schlumberger Technology Corporation Oilfield data file classification and information processing systems
US11625934B2 (en) 2020-02-04 2023-04-11 Eygs Llp Machine learning based end-to-end extraction of tables from electronic documents
US11416684B2 (en) * 2020-02-06 2022-08-16 Adobe Inc. Automated identification of concept labels for a set of documents
US11354513B2 (en) 2020-02-06 2022-06-07 Adobe Inc. Automated identification of concept labels for a text fragment
US11250128B2 (en) 2020-02-18 2022-02-15 Bank Of America Corporation System and method for detecting source code anomalies
US11176329B2 (en) 2020-02-18 2021-11-16 Bank Of America Corporation Source code compiler using natural language input
US11568153B2 (en) 2020-03-05 2023-01-31 Bank Of America Corporation Narrative evaluator
US20210326718A1 (en) * 2020-04-16 2021-10-21 Microsoft Technology Licensing, Llc Machine learning techniques to shape downstream content traffic through hashtag suggestion during content creation
US11934441B2 (en) 2020-04-29 2024-03-19 International Business Machines Corporation Generative ontology learning and natural language processing with predictive language models
US11061543B1 (en) 2020-05-11 2021-07-13 Apple Inc. Providing relevant data items based on context
US11038934B1 (en) 2020-05-11 2021-06-15 Apple Inc. Digital assistant hardware abstraction
US11341339B1 (en) * 2020-05-14 2022-05-24 Amazon Technologies, Inc. Confidence calibration for natural-language understanding models that provides optimal interpretability
US11704580B2 (en) 2020-05-31 2023-07-18 International Business Machines Corporation Automated combination of predictions made by different prediction systems
US12118297B2 (en) * 2020-06-15 2024-10-15 Canon Kabushiki Kaisha Information processing apparatus
US11393456B1 (en) * 2020-06-26 2022-07-19 Amazon Technologies, Inc. Spoken language understanding system
CN111522891A (en) * 2020-07-03 2020-08-11 支付宝(杭州)信息技术有限公司 Method and device for determining relationship closeness between two entities and electronic equipment
US11490204B2 (en) 2020-07-20 2022-11-01 Apple Inc. Multi-device audio adjustment coordination
US11438683B2 (en) 2020-07-21 2022-09-06 Apple Inc. User identification using headphones
US20220027578A1 (en) * 2020-07-27 2022-01-27 Nvidia Corporation Text string summarization
US11829720B2 (en) 2020-09-01 2023-11-28 Apple Inc. Analysis and validation of language models
US12080289B2 (en) * 2020-12-22 2024-09-03 Samsung Electronics Co., Ltd. Electronic apparatus, system comprising electronic apparatus and server and controlling method thereof
WO2022159443A1 (en) 2021-01-19 2022-07-28 Source Global, PBC Systems and methods for generating water from air
US12073323B2 (en) 2021-03-10 2024-08-27 Rockspoon, Inc. System and method for intelligent service intermediation
US11645464B2 (en) 2021-03-18 2023-05-09 International Business Machines Corporation Transforming a lexicon that describes an information asset
CN115237856A (en) * 2021-04-23 2022-10-25 伊姆西Ip控股有限责任公司 Method, device and computer program product for marking files
US20230008868A1 (en) * 2021-07-08 2023-01-12 Nippon Telegraph And Telephone Corporation User authentication device, user authentication method, and user authentication computer program
US11443102B1 (en) * 2021-08-13 2022-09-13 Pricewaterhousecoopers Llp Methods and systems for artificial intelligence-assisted document annotation
US11645462B2 (en) 2021-08-13 2023-05-09 Pricewaterhousecoopers Llp Continuous machine learning method and system for information extraction
US12050858B2 (en) * 2021-09-21 2024-07-30 Bank Of America Corporation Personal data discovery
US20230088315A1 (en) * 2021-09-22 2023-03-23 Motorola Solutions, Inc. System and method to support human-machine interactions for public safety annotations
US11516157B1 (en) * 2021-09-30 2022-11-29 Dell Products, L.P. Machine learning facilitated drafting of response to communication
US20230134796A1 (en) * 2021-10-29 2023-05-04 Glipped, Inc. Named entity recognition system for sentiment labeling
US11977843B2 (en) * 2022-01-24 2024-05-07 S&P Global Inc. Dynamic ontology for intelligent data discovery
US12106049B1 (en) * 2022-03-16 2024-10-01 Yseop Sa Ontology mapping and adaptation systems and methods
US20230409838A1 (en) * 2022-05-31 2023-12-21 International Business Machines Corporation Explaining natural-language-to-flow pipelines
US20240037598A1 (en) * 2022-07-31 2024-02-01 Klaviyo Inc. Method and system for exemplary campaign message generation
US11546323B1 (en) 2022-08-17 2023-01-03 strongDM, Inc. Credential management for distributed services
US11736531B1 (en) 2022-08-31 2023-08-22 strongDM, Inc. Managing and monitoring endpoint activity in secured networks
US11765159B1 (en) 2022-09-28 2023-09-19 strongDM, Inc. Connection revocation in overlay networks
CN115294773B (en) * 2022-09-29 2023-02-14 深圳市城市交通规划设计研究中心股份有限公司 Bus lane optimal configuration method, electronic device and storage medium
US12079452B2 (en) * 2022-09-30 2024-09-03 Atlassian Pty Ltd. Label management system for an electronic document management service
US12038985B2 (en) * 2022-11-29 2024-07-16 Sap Se Batch user interface calls and response processing
CN115904482B (en) * 2022-11-30 2023-09-26 杭州巨灵兽智能科技有限公司 Interface document generation method, device, equipment and storage medium
US11916885B1 (en) 2023-01-09 2024-02-27 strongDM, Inc. Tunnelling with support for dynamic naming resolution
US20240273309A1 (en) * 2023-02-15 2024-08-15 Casetext, Inc. Text generation interface system
WO2024177487A1 (en) * 2023-02-20 2024-08-29 서울대학교병원 Medical prompt generation device and method
US11765207B1 (en) * 2023-03-17 2023-09-19 strongDM, Inc. Declaring network policies using natural language

Family Cites Families (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5970490A (en) * 1996-11-05 1999-10-19 Xerox Corporation Integration platform for heterogeneous databases
AU3689901A (en) * 2000-02-10 2001-08-20 Involve Technology Llc System for creating and maintaining a database of information utilizing user opinions
US7730113B1 (en) * 2000-03-07 2010-06-01 Applied Discovery, Inc. Network-based system and method for accessing and processing emails and other electronic legal documents that may include duplicate information
WO2002005137A2 (en) * 2000-07-07 2002-01-17 Criticalpoint Software Corporation Methods and system for generating and searching ontology databases
US7376635B1 (en) * 2000-07-21 2008-05-20 Ford Global Technologies, Llc Theme-based system and method for classifying documents
US6513059B1 (en) * 2000-08-24 2003-01-28 Cambira Corporation Adaptive collaborative intelligent network system
US20020152202A1 (en) * 2000-08-30 2002-10-17 Perro David J. Method and system for retrieving information using natural language queries
US7146399B2 (en) * 2001-05-25 2006-12-05 2006 Trident Company Run-time architecture for enterprise integration with transformation generation
US20050154690A1 (en) * 2002-02-04 2005-07-14 Celestar Lexico-Sciences, Inc Document knowledge management apparatus and method
US7548847B2 (en) * 2002-05-10 2009-06-16 Microsoft Corporation System for automatically annotating training data for a natural language understanding system
US8543564B2 (en) * 2002-12-23 2013-09-24 West Publishing Company Information retrieval systems with database-selection aids
US20040254950A1 (en) * 2003-06-13 2004-12-16 Musgrove Timothy A. Catalog taxonomy for storing product information and system and method using same
WO2005008358A2 (en) * 2003-07-22 2005-01-27 Kinor Technologies Inc. Information access using ontologies
US20050278362A1 (en) * 2003-08-12 2005-12-15 Maren Alianna J Knowledge discovery system
GB0320205D0 (en) * 2003-08-28 2003-10-01 British Telecomm Method and apparatus for storing and retrieving data
US20050060140A1 (en) * 2003-09-15 2005-03-17 Maddox Paul Christopher Using semantic feature structures for document comparisons
US7533090B2 (en) * 2004-03-30 2009-05-12 Google Inc. System and method for rating electronic documents
US7809548B2 (en) * 2004-06-14 2010-10-05 University Of North Texas Graph-based ranking algorithms for text processing
US20070106499A1 (en) * 2005-08-09 2007-05-10 Kathleen Dahlgren Natural language search system
US7664746B2 (en) * 2005-11-15 2010-02-16 Microsoft Corporation Personalized search and headlines
US20070150802A1 (en) * 2005-12-12 2007-06-28 Canon Information Systems Research Australia Pty. Ltd. Document annotation and interface
US7835911B2 (en) * 2005-12-30 2010-11-16 Nuance Communications, Inc. Method and system for automatically building natural language understanding models
EP2159717A3 (en) * 2006-03-30 2010-03-17 Sony France S.A. Hybrid audio-visual categorization system and method
US7664644B1 (en) * 2006-06-09 2010-02-16 At&T Intellectual Property Ii, L.P. Multitask learning for spoken language understanding
US8131756B2 (en) 2006-06-21 2012-03-06 Carus Alwin B Apparatus, system and method for developing tools to process natural language text
AR062635A1 (en) * 2006-09-01 2008-11-19 Thomson Global Resources SYSTEM, METHODS, SOFTWARE AND INTERFASES TO FORMAT APPOINTMENTS OF LEGISLATION
JP4311432B2 (en) * 2006-09-29 2009-08-12 ブラザー工業株式会社 Information processing apparatus and program
US7774198B2 (en) * 2006-10-06 2010-08-10 Xerox Corporation Navigation system for text
WO2008086189A2 (en) * 2007-01-04 2008-07-17 Wide Angle Llc Relevancy rating of tags
US8738606B2 (en) * 2007-03-30 2014-05-27 Microsoft Corporation Query generation using environment configuration
US8214743B2 (en) * 2007-08-07 2012-07-03 International Business Machines Corporation Data management techniques
JP2011501847A (en) * 2007-10-17 2011-01-13 アイティーアイ・スコットランド・リミテッド Computer-implemented method
JP4518165B2 (en) * 2008-03-11 2010-08-04 富士ゼロックス株式会社 Related document presentation system and program
US20100153318A1 (en) * 2008-11-19 2010-06-17 Massachusetts Institute Of Technology Methods and systems for automatically summarizing semantic properties from documents with freeform textual annotations
US8165974B2 (en) * 2009-06-08 2012-04-24 Xerox Corporation System and method for assisted document review
US8515957B2 (en) * 2009-07-28 2013-08-20 Fti Consulting, Inc. System and method for displaying relationships between electronically stored information to provide classification suggestions via injection
US20110029517A1 (en) * 2009-07-31 2011-02-03 Shihao Ji Global and topical ranking of search results using user clicks
US8682649B2 (en) * 2009-11-12 2014-03-25 Apple Inc. Sentiment prediction from textual data
US20120064501A1 (en) * 2010-04-08 2012-03-15 Sukkarieh Jana Z Systems and Methods for Evaluation of Automatic Content Scoring Technologies
US8498950B2 (en) * 2010-10-15 2013-07-30 Yahoo! Inc. System for training classifiers in multiple categories through active learning
US8725739B2 (en) * 2010-11-01 2014-05-13 Evri, Inc. Category-based content recommendation
US20120271774A1 (en) * 2011-04-21 2012-10-25 Hirevue, Inc. Interview frameworks
WO2012166581A2 (en) * 2011-05-27 2012-12-06 Ctc Tech Corp. Creation, use and training of computer-based discovery avatars
JP5866728B2 (en) * 2011-10-14 2016-02-17 サイバーアイ・エンタテインメント株式会社 Knowledge information processing server system with image recognition system
US11410072B2 (en) * 2011-10-21 2022-08-09 Educational Testing Service Computer-implemented systems and methods for detection of sentiment in writing
JP2015505082A (en) * 2011-12-12 2015-02-16 インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation Generation of natural language processing model for information domain
US9652452B2 (en) * 2012-01-06 2017-05-16 Yactraq Online Inc. Method and system for constructing a language model
US9170715B1 (en) * 2012-02-03 2015-10-27 CounterPointe Digital Technologies LLC System and method for mass visualization of real estate properties
US8832162B2 (en) * 2012-03-25 2014-09-09 Think Computer Corporation Method and system for storing, categorizing and distributing information concerning relationships between data
US8943135B2 (en) * 2012-07-24 2015-01-27 Fard Johnmar System and method for measuring the positive or negative impact of digital and social media content on intent and behavior
US20140188665A1 (en) 2013-01-02 2014-07-03 CrowdChunk LLC CrowdChunk System, Method, and Computer Program Product for Searching Summaries of Online Reviews of Products
GB2509539A (en) * 2013-01-08 2014-07-09 Ibm Production rule engine
US20140257990A1 (en) * 2013-03-06 2014-09-11 TipTap, Inc. Method and system for determining correlations between personality traits of a group of consumers and a brand/product
US9122681B2 (en) * 2013-03-15 2015-09-01 Gordon Villy Cormack Systems and methods for classifying electronic information using advanced active learning techniques
US9189540B2 (en) * 2013-04-05 2015-11-17 Hewlett-Packard Development Company, L.P. Mobile web-based platform for providing a contextual alignment view of a corpus of documents
EP2994846A4 (en) 2013-05-09 2017-01-25 Moodwire, Inc. Hybrid human machine learning system and method
US9348815B1 (en) * 2013-06-28 2016-05-24 Digital Reasoning Systems, Inc. Systems and methods for construction, maintenance, and improvement of knowledge representations
US9772994B2 (en) 2013-07-25 2017-09-26 Intel Corporation Self-learning statistical natural language processing for automatic production of virtual personal assistants
US9990422B2 (en) * 2013-10-15 2018-06-05 Adobe Systems Incorporated Contextual analysis engine
US9547640B2 (en) * 2013-10-16 2017-01-17 International Business Machines Corporation Ontology-driven annotation confidence levels for natural language processing
US10754925B2 (en) * 2014-06-04 2020-08-25 Nuance Communications, Inc. NLU training with user corrections to engine annotations
US20160019282A1 (en) * 2014-07-16 2016-01-21 Axiom Global Inc. Discovery management method and system
US9886247B2 (en) * 2014-10-30 2018-02-06 International Business Machines Corporation Using an application programming interface (API) data structure in recommending an API composite
US9860308B2 (en) * 2014-11-25 2018-01-02 International Business Machines Corporation Collaborative creation of annotation training data
US9852132B2 (en) * 2014-11-25 2017-12-26 Chegg, Inc. Building a topical learning model in a content management system
US20160162569A1 (en) * 2014-12-09 2016-06-09 Idibon, Inc. Methods and systems for improving machine learning performance
US11295071B2 (en) * 2014-12-09 2022-04-05 100.Co, Llc Graphical systems and methods for human-in-the-loop machine intelligence

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11443082B2 (en) 2020-05-27 2022-09-13 Accenture Global Solutions Limited Utilizing deep learning and natural language processing to convert a technical architecture diagram into an interactive technical architecture diagram
US11556788B2 (en) * 2020-06-15 2023-01-17 International Business Machines Corporation Text-based response environment action selection
WO2022155539A1 (en) * 2021-01-14 2022-07-21 Virtuosource Llc A rules-based decision support system for assessment of digital content involving natural language
US20220253607A1 (en) * 2021-02-08 2022-08-11 International Business Machines Corporation Question based chatbot generator from web api specifications
US11741313B2 (en) * 2021-02-08 2023-08-29 International Business Machines Corporation Question based chatbot generator from web API specifications

Also Published As

Publication number Publication date
US20190311024A1 (en) 2019-10-10
US20190243886A1 (en) 2019-08-08
US20210232761A1 (en) 2021-07-29
US20210165955A1 (en) 2021-06-03
US20170235813A1 (en) 2017-08-17
US20160162476A1 (en) 2016-06-09
US10127214B2 (en) 2018-11-13
US20200234002A1 (en) 2020-07-23
US9495345B2 (en) 2016-11-15
US20210150130A1 (en) 2021-05-20
US20160162458A1 (en) 2016-06-09
US20160162569A1 (en) 2016-06-09
US20210232762A1 (en) 2021-07-29
US20210232760A1 (en) 2021-07-29
US11599714B2 (en) 2023-03-07
US20200184146A1 (en) 2020-06-11
US20160162457A1 (en) 2016-06-09
US20160162464A1 (en) 2016-06-09
US20190311025A1 (en) 2019-10-10
US11288444B2 (en) 2022-03-29
US20190303428A1 (en) 2019-10-03
US20210232763A1 (en) 2021-07-29
US20160162456A1 (en) 2016-06-09

Similar Documents

Publication Publication Date Title
US20210232762A1 (en) Architectures for natural language processing
US20240078386A1 (en) Methods and systems for language-agnostic machine learning in natural language processing using feature extraction
US11816439B2 (en) Multi-turn dialogue response generation with template generation
CN111753060B (en) Information retrieval method, apparatus, device and computer readable storage medium
US11295071B2 (en) Graphical systems and methods for human-in-the-loop machine intelligence
US9268766B2 (en) Phrase-based data classification system
US12086548B2 (en) Event extraction from documents with co-reference
US10977155B1 (en) System for providing autonomous discovery of field or navigation constraints
KR102169382B1 (en) Artificial Intelligence-Based Personalized Expert Cross Matching and Proposal System
CN112101042A (en) Text emotion recognition method and device, terminal device and storage medium
US10896034B2 (en) Methods and systems for automated screen display generation and configuration
US11314488B2 (en) Methods and systems for automated screen display generation and configuration
US20240256597A1 (en) Machine learning selection of images
US20240104506A1 (en) Using unsupervised machine learning to identify attribute values as related to an input
Reda Intelligent Assistant Agents: Comparative Analysis of Chatbots through Diverse Methodologies
Upreti A Comparative Analysis of NLP Algorithms for Implementing AI Conversational Assistants
Upreti A COMPARATIVE ANALYSIS OF NLP ALGORITHMS FOR IMPLEMENTING AI CONVERSATIONAL ASSISTANTS: Comparative Analysis of NLP Algorithms for NLI
CA3199880A1 (en) Deep technology innovation management by cross-pollinating innovations dataset
WO2024233222A1 (en) Ensemble learning enhanced prompting for open relation extraction
CN117033540A (en) Report generation method, report generation device, electronic equipment and medium

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: IDIBON, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MUNRO, ROBERT J.;ERLE, SCHUYLER D.;WALKER, CHRISTOPHER;AND OTHERS;SIGNING DATES FROM 20160119 TO 20160322;REEL/FRAME:055945/0300

AS Assignment

Owner name: IDIBON (ASSIGNMENT FOR THE BENEFIT OF CREDITORS), LLC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:IDIBON, INC.;REEL/FRAME:055978/0362

Effective date: 20160519

AS Assignment

Owner name: HEALY, TREVOR, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:IDIBON (ASSIGNMENT FOR THE BENEFIT OF CREDITORS), LLC;REEL/FRAME:056057/0325

Effective date: 20161010

AS Assignment

Owner name: AIPARC HOLDINGS PTE. LTD., SINGAPORE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HEALY, TREVOR;REEL/FRAME:056083/0123

Effective date: 20181006

AS Assignment

Owner name: AI IP INVESTMENTS LTD, VIRGIN ISLANDS, BRITISH

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AIPARC HOLDINGS PTE. LTD.;REEL/FRAME:056096/0278

Effective date: 20210114

AS Assignment

Owner name: 100.CO, LLC, FLORIDA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AI IP INVESTMENTS LTD.;REEL/FRAME:056145/0509

Effective date: 20210414

AS Assignment

Owner name: 100.CO TECHNOLOGIES, INC., FLORIDA

Free format text: NUNC PRO TUNC ASSIGNMENT;ASSIGNOR:100.CO, LLC;REEL/FRAME:062131/0714

Effective date: 20221214

AS Assignment

Owner name: DAASH INTELLIGENCE, INC., FLORIDA

Free format text: CHANGE OF NAME;ASSIGNOR:100.CO TECHNOLOGIES, INC.;REEL/FRAME:064347/0117

Effective date: 20230118

AS Assignment

Owner name: 100.CO GLOBAL HOLDINGS, LLC, FLORIDA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DAASH INTELLIGENCE, INC.;REEL/FRAME:064420/0108

Effective date: 20230713

AS Assignment

Owner name: AI IP INVESTMENTS LTD., UNITED KINGDOM

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:100.CO GLOBAL HOLDINGS, LLC;REEL/FRAME:066636/0583

Effective date: 20231220