US10545648B2 - Evaluating conversation data based on risk factors - Google Patents
Evaluating conversation data based on risk factors Download PDFInfo
- Publication number
- US10545648B2 US10545648B2 US16/242,639 US201916242639A US10545648B2 US 10545648 B2 US10545648 B2 US 10545648B2 US 201916242639 A US201916242639 A US 201916242639A US 10545648 B2 US10545648 B2 US 10545648B2
- Authority
- US
- United States
- Prior art keywords
- intent
- user
- conversation
- user input
- intent unit
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G06F17/2785—
-
- G06F17/279—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
- G06F40/35—Discourse or dialogue representation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G06F17/2755—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/268—Morphological analysis
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1822—Parsing for meaning understanding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/183—Speech classification or search using natural language modelling using context dependencies, e.g. language models
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/66—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition
Definitions
- a growing number of people are using smart devices, such as smart phones, tablet computers, laptop computers, and so on, to perform a variety of functionality.
- the users interact with their devices through a virtual assistant.
- the virtual assistant may communicate with a user to perform a desired service or task, such as searching for content, checking-in to a flight, setting a calendar appointment, and so on.
- a desired service or task such as searching for content, checking-in to a flight, setting a calendar appointment, and so on.
- FIG. 1 illustrates an example architecture in which techniques described herein may be implemented.
- FIG. 2 illustrates example details of a virtual assistant service.
- FIG. 3 illustrates example details of a smart device.
- FIG. 4 illustrates an example intent unit selection interface that may be presented to facilitate selection of intent units to review for health status.
- FIG. 5 illustrates an example chart for displaying health status associated with intent units.
- FIG. 6 illustrates an example feedback interface to allow a user to review and provide feedback on mapping of inputs to intent units.
- FIG. 7 illustrates an example risk factor interface to allow a user to configure factors for determining a risk that an input is incorrectly mapped to an intent unit.
- FIG. 8 illustrates an example feedback results interface for reviewing feedback on virtual assistant conversations.
- FIG. 9 illustrates an example process for determining a measure of confidence that input received from a conversation involving a virtual assistant is mapped to a correct intent unit, and receiving feedback from a voter regarding accuracy of the mapping.
- FIGS. 10 and 11 illustrate example processes for determining and utilizing confidence values of intents associated with user inputs.
- This disclosure describes techniques and architectures for evaluating conversations.
- conversations with users, virtual assistants, and others may be analyzed to identify potential risks within a language model that is employed by the virtual assistants and other entities.
- the potential risks may be evaluated by administrators, users, systems, and others to identify potential issues with the language model that need to be addressed. This may allow the language model to be improved and enhance user experience with the virtual assistants and others that employ the language model.
- the techniques and architectures described herein may analyze conversation data for one or more conversations.
- the conversation data may generally include back-and-forth communications between a user and a virtual assistant or other system that employs Natural Language Processing (NLP) techniques, such as customer service agents.
- NLP Natural Language Processing
- the NLP techniques may map user input to one or more intents (e.g., intent units) that are defined by a language model. Based on the mapping, a task may be performed, such as presenting a response (e.g., textual response, audio response, etc.), performing an action (e.g., booking a hotel room, scheduling an appointment, etc.), and so on.
- the conversation data may be analyzed to determine a confidence value for an identified intent.
- the confidence value may indicate a level of confidence, or measure of confidence, that the intent is accurately determined for the input (e.g., does the intent satisfy the input).
- the NLP techniques may map the user input to an intent unit that is associated with “flight status.” Based on the identified intent unit of “flight status,” a task to may be performed by the virtual assistant, such as providing a response of “Your flight is scheduled to leave on-time at 3:30.”
- the techniques and architectures may analyze the conversation to determine a level of confidence that the “flight status” intent unit is the correct intent unit that should have been identified.
- a confidence value may be determined based on a variety of risk factors.
- the one or more risk factors may include various forms of information.
- a risk factor may indicate whether or not an intent is identified for user input, whether or not user input proceeded a failure to identify an intent for other user input, whether or not user input is involved in a conversation that included a failure to identify an intent, whether or not a same intent is identified in a conversation, a tone of user's voice, a facial expression of a user, and so on.
- risk factors are discussed in further detail herein.
- a “risk factor” may sometimes be referred to as a “risk indicator,” while a confidence value may sometimes be referred to as a “risk score.” In some instances, multiple risk factors may be used and weighted to generate a confidence value.
- a confidence value may be used to evaluate an intent unit.
- the confidence value may be used to determine a health status (e.g., unit risk) that indicates a level of risk associated with the intent unit.
- a health status e.g., unit risk
- the health status of the intent unit may indicate that the intent unit is relatively risky. In other words, the health status may indicate that the intent unit needs to be updated. In some instances, the health status may be used to rank the intent unit relative to other intent units.
- the ranking and/or health status may be presented to an administrator of a language model or others, so that a language model may be evaluated.
- the user may update the intent unit and/or other elements of the language model in order to address the risky intent unit.
- the user may cause that the intent unit and/or corresponding user input be released to other users for further evaluation.
- the intent unit may be released to voters to obtain feedback (e.g., votes) from users regarding an accuracy of matching user input to the intent unit.
- the voters may view the intent unit and user input that hit the intent unit and provide an opinion as to whether or not they agree that the intent unit should have been identified for the user input.
- the results of the voting may be provided to the administrator so that further action may be taken (e.g., update the intent unit if the voters agree that the intent unit is risky).
- a confidence value may be used to determine whether or not a task should be performed during a conversation. For example, a confidence value may be generated as the conversation is occurring (e.g., in real-time). The confidence value may become a form of context to conversation that effects how the conversation proceeds. For example, if the confidence value is relatively high (e.g., more than a threshold), a task associated with the intent may be performed (e.g., in response to input of “How can I make one more reservation, then add it to the ticket,” provide an answer “you need to login and click add flight to reservation” if the confidence value is relatively high).
- the user may be prompted for additional information (e.g., in response to input of “How can I make one more reservation, then add it to the ticket,” provide a follow-up question of “what ticket are you referring to” if the confidence value is relatively low).
- a risky element of a language model may be identified. This may provide a relatively broad understanding of the language model and/or help prioritize what elements to update in the language model. Further, by improving the language model, this may ultimately enhance user interactions with virtual assistants and other entities that employ the language model.
- the techniques may evaluate relatively large data sets that include thousands or hundreds-of-thousands of pieces of user input in an efficient and timely manner. Further, the techniques and architectures may evaluate a language model that may include thousands or hundreds-of-thousands of elements (e.g., intent units, etc.).
- FIG. 1 illustrates an example architecture 100 in which the techniques described herein may be implemented.
- the architecture 100 includes one or more smart devices 102 (hereinafter “the smart device 102 ”) to present a virtual assistant to one or more end-users 104 (hereinafter “the user 104 ”) to perform tasks for the user 104 .
- the virtual assistant may be implemented in cooperation with a service provider 106 that generally manages access to and/or functionality associated with the virtual assistant.
- the service provider 106 operates in cooperation with a conversation voting community 108 to evaluate conversation data.
- the conversation voting community 108 may include one or more voters 110 (hereinafter “the voter 110 ”) to interact with one or more computing devices 112 (hereinafter “the computing devices 112 ”) to provide feedback regarding conversations.
- the feedback may be provided to the service provider 106 for analysis.
- any of the processes may be performed by the computing device 112 and/or the smart device 102 .
- the service provider 106 is illustrated as a single service, the service provider 106 may be implemented as a variety of services that are implemented separately from each other (e.g., a virtual assistant service that is separate from a conversation analysis service).
- the smart device 102 , the service provider 106 , and/or the computing device 112 may communicate via one or more networks 114 .
- the one or more networks 114 may include any one or combination of multiple different types of networks, such as cellular networks, wireless networks, Local Area Networks (LANs), Wide Area Networks (WANs), the Internet, and so on.
- the service provider 106 may include one or more computing devices.
- the one or more computing devices may be implemented as one or more desktop computers, laptop computers, servers, and so on.
- the one or more computing devices may be configured in a cluster, data center, cloud computing environment, or a combination thereof.
- the service provider 106 provides cloud computing resources, including computational resources, storage resources, networking resources, and the like, that operate remotely to the smart device 102 and/or the computing device 112 .
- the service provider 106 may perform a variety of operations. For example, the service provider 106 may analyze conversation data based on a variety of risk factors to generate confidence values for user input. A confidence value may then be associated with the corresponding user input and/or an intent unit that was hit for the user input. Additionally, or alternatively, the service provider 106 may determine a health status of an intent unit based on user input that is mapped to the intent unit. For example, the health status of an intent unit may indicate a number of user inputs (that have been mapped to the intent unit and) that have less than a threshold confidence value relative to a number of user inputs that have been mapped to the intent unit overall.
- the service provider 106 may provide a variety of user interfaces to assist in evaluating conversation data, such as any of the interfaces of FIGS. 4-8 .
- the service provider 116 may provide an evaluation interface 116 to an administrator of the service provider 106 .
- the administrator may view potentially risky intent units (e.g., in a ranking based on health status) and select an intent unit to release to voters.
- the service provider 106 may provide a voting interface 118 via the computing device 112 , so that the voter 110 may provide feedback regarding an accuracy of matching user input to an intent unit.
- the voter 110 may select “Yes” or “No” to indicate whether or not the voter agrees that the identified intent unit should map to the user input.
- the service provider 106 may collect feedback from a variety of voters over time. Although not illustrated in FIG. 1 , the service provider 106 may be associated with an output device, such as a computer monitor, speaker, projector, computing device, and so on. The service provider 106 may provide user interfaces to, for example, an administrator via the output device.
- an output device such as a computer monitor, speaker, projector, computing device, and so on.
- the service provider 106 may provide user interfaces to, for example, an administrator via the output device.
- the smart device 102 , the computing device 112 , and/or an output device associated with the service provider 106 may comprise any type of computing device that is configured to perform an operation.
- the smart device 102 and/or the computing device 112 may be implemented as a laptop computer, a desktop computer, a server, a smart phone, an electronic reader device, a mobile handset, a personal digital assistant (PDA), a portable navigation device, a portable gaming device, a tablet computer, a wearable computer (e.g., a watch, optical head-mounted display (e.g., a pair of glass(es) with computing capabilities), etc.), a portable media player, a television, a set-top box, a computer system in a car, an appliance, a camera, a robot, a hologram system, a security system, a home-based computer system (e.g., intercom system, home media system, etc.), a projector, an automated teller machine (ATM), and so on.
- ATM automated
- the smart device 102 may output a virtual assistant to the user 104 via a conversation user interface.
- the virtual assistant may interact with the user 104 in a conversational manner to perform tasks. For example, in response to a query from the user 104 to “find the nearest restaurant,” a virtual assistant may provide information through the conversation user interface that identifies the nearest restaurant. As such, the user 104 and/or the virtual assistant may communicate in a natural language format.
- a virtual assistant may be configured for multi-modal input/output (e.g., receive and/or respond in audio or speech, text, touch, gesture, etc.), multi-language communication (e.g., receive and/or respond according to any type of human language), multi-channel communication (e.g., carry out conversations through a variety of computing devices, such as continuing a conversation as a user transitions from using one computing device to another), and other types of input/output or communication.
- multi-modal input/output e.g., receive and/or respond in audio or speech, text, touch, gesture, etc.
- multi-language communication e.g., receive and/or respond according to any type of human language
- multi-channel communication e.g., carry out conversations through a variety of computing devices, such as continuing a conversation as a user transitions from using one computing device to another
- other types of input/output or communication e.g., multi-modal input/output (e.g., receive and/or respond in audio or speech, text
- a virtual assistant may comprise an intelligent personal assistant.
- a virtual assistant may generally perform tasks for users and act as an interface to information of a service provider, information of the smart device 102 , information of the service provider 106 , and/or any type of information. For example, in response to input from the user 104 , a virtual assistant may access content items stored on a service provider and provide a content item to the user.
- a virtual assistant may embody a human-like persona (e.g., human emulation) and/or artificial intelligence (AI).
- a virtual assistant may be represented by an image or avatar that is displayed on the smart device 102 .
- An avatar may comprise an animated character that may take on any number of shapes and appearances, and/or resemble a human talking to a user.
- the avatar may be arranged as a representative of a service provider or the service provider 106 , while in other instances the avatar may be a dedicated personal assistant to a user.
- Example virtual assistants are described in U.S. application Ser. No. 14/293,586, filed Jun. 2, 2014, which is incorporated herein by reference.
- the smart device 102 may be equipped with one or more processors, memory, and/or one or more network interfaces.
- the smart device 102 may also include one or more cameras, one or more displays, one or more microphones, one or more speakers, and/or one or more sensors. These components may be communicatively coupled to the one or more processors.
- the one or more processors may include a central processing unit (CPU), a graphics processing unit (GPU), a microprocessor, a digital signal processor and so on.
- the one or more cameras may include a front facing camera and/or a rear facing camera.
- the one or more displays may include a touch screen, a Liquid-crystal Display (LCD), a Light-emitting Diode (LED) display, an organic LED display, a plasma display, an electronic paper display or any other type of technology.
- the one or more sensors may include an accelerometer, compass, gyroscope, magnetometer, Global Positioning System (GPS), olfactory sensor (e.g., for smell), heart rate sensor, light sensor, capacitive sensor, inductive sensor, eye tracking device, or other sensor.
- GPS Global Positioning System
- FIG. 2 illustrates details of the example the service provider 106 of FIG. 1 .
- the service provider 106 may be implemented as one or more computing devices.
- the one or more computing devices may include one or more processors 202 , memory 204 , and one or more network interfaces 206 .
- the one or more processors 202 may include a central processing unit (CPU), a graphics processing unit (GPU), a microprocessor, a digital signal processor, and so on.
- CPU central processing unit
- GPU graphics processing unit
- microprocessor a digital signal processor
- the memory 204 may include software functionality configured as one or more “modules.”
- module is intended to represent example divisions of the software for purposes of discussion, and is not intended to represent any type of requirement or required method, manner or necessary organization. Accordingly, while various “modules” are discussed, their functionality and/or similar functionality could be arranged differently (e.g., combined into a fewer number of modules, broken into a larger number of modules, etc.). Further, while certain functions are described herein as being implemented as software modules configured for execution by a processor, in other embodiments, any or all of the functions may be implemented (e.g., performed) in whole or in part by hardware logic components.
- FPGAs Field-programmable Gate Arrays
- ASICs Application-specific Integrated Circuits
- ASSPs Program-specific Standard Products
- SOCs System-on-a-chip systems
- CPLDs Complex Programmable Logic Devices
- the memory 204 may include one or a combination of computer-readable media (e.g., computer storage media).
- Computer-readable media includes volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data.
- Computer-readable media includes, but is not limited to, phase change memory (PRAM), static random-access memory (SRAM), dynamic random-access memory (DRAM), other types of random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, compact disk read-only memory (CD-ROM), digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transitory medium that can be used to store information for access by a computing device.
- PRAM phase change memory
- SRAM static random-access memory
- DRAM dynamic random-access memory
- RAM random access memory
- ROM read-only memory
- EEPROM electrically erasable programmable read-only memory
- flash memory or other memory technology
- CD-ROM compact disk read-only memory
- DVD digital versatile disks
- magnetic cassettes magnetic tape
- magnetic disk storage or other magnetic storage devices or any other non-transitory medium that can be used
- the memory 204 includes an input processing module 208 , a task and response module 210 , a risk indicator module 212 , a voting module 214 , and a conversation review module 216 .
- the input processing module 208 may be configured to perform various techniques to process input received from a user. For instance, input that is received from the user 104 during a conversation with a virtual assistant may be sent to the input processing module 208 for processing. If the input is speech input, the input processing module 208 may perform speech recognition techniques to convert the input into a format that is understandable by a computing device, such as text. Additionally, or alternatively, the input processing module 208 may utilize Natural Language Processing (NLP) to interpret or derive a meaning and/or concept of the input.
- NLP Natural Language Processing
- the speech recognition and/or NLP techniques may include known or new techniques.
- the input processing module 208 may map user input to an intent unit based on a pattern of components for the user input.
- the pattern may include an order and/or proximity of the components to each other.
- Each term in the user input may be mapped to a component that represents a variation(s) of a term(s).
- the pattern of components of the user input may then be mapped to a pattern defined within an intent unit.
- the intent unit may be associated with a task (described below), which may be performed to satisfy the user input.
- the task may include providing a response to a user, performing an action (e.g., booking a flight), and so on.
- a component for the term “car” may include synonyms and/or spelling variations of the term, such as “automobile,” “kar,” etc.
- Example components include Vocab Terms (e.g., vocabulary synonyms and/or spelling variations), Helper Terms (e.g., words or phrases that are commonly used, but have only one meaning, such as “and,” “or,” “but,” etc.), Wild Cards (e.g., placeholders for any random word or words), and so on.
- Example components are described in U.S. application Ser. No. 12/014,229, filed on Jan. 15, 2008, which is incorporated herein by reference.
- the task and response module 210 may be configured to identify and/or perform tasks and/or formulate a response to input and based on an intent unit the input has been mapped to.
- the triggering of an intent unit may cause a tasks and/or responses to be provided by the task and response module 210 .
- users may interact with virtual assistants to cause tasks to be performed by the virtual assistants.
- a task may be performed in response to explicit user input, such as playing music in response to “please play music.” In other instances, a task may be performed in response to inferred user input requesting that that the task be performed, such as providing weather information in response to “the weather looks nice today.” In yet further instances, a task may be performed when an event has occurred (and possibly when no input has been received), such as providing flight information an hour before a flight, presenting flight information upon arrival of a user at an airport, and so on.
- a task may include any type of operation that is performed at least in part by a computing device.
- a task may include logging a user into a site, setting a calendar appointment, resetting a password for a user, purchasing an item, opening an application, sending an instruction to a device to perform an act, sending an email, navigating to a web site, upgrading a user's seat assignment, outputting content (e.g., outputting audio (an audible answer), video, an image, text, a hyperlink, etc.), and so on.
- a task may include performing an operation according to one or more criteria (e.g., one or more default settings), such as sending an email through a particular email account, providing directions with a particular mobile application, searching for content through a particular search engine, and so on.
- a task may include or be associated with a response to a user (e.g., “here is your requested information,” outputting content, etc.).
- a response may be provided through a conversation user interface associated with a virtual assistant.
- a response may be addressed to or otherwise tailored to a user (e.g., “Yes, John, as a Gold Customer you are entitled to a seat upgrade, and I have provided some links below that may be of interest to you . . . ”).
- Input and/or output between users and virtual assistants (e.g., conversations) may be stored in a virtual assistant conversation data store 218 .
- the risk indicator module 212 may be configured to determine a measure of confidence (e.g., confidence value) that input received from the user 104 has been mapped to the correct intent unit of a language model (stored in a virtual assistant data store 220 ).
- the risk indicator module 212 may use one or more risk indicators, or risk factors, from a risk indicator data store 222 in order to identify the measure of confidence that the input was mapped to the correct intent unit.
- the one or more factors may include various forms of information.
- a factor may sometimes be referred to as a “risk indicator,” while a confidence value may sometimes be referred to as a “risk score,” “measure of confidence,” or “degree of confidence.”
- multiple factors may be used and weighted to generate a confidence value.
- the weighted factors may be combined to form the confidence value.
- Example factors that may be used to evaluate conversation data include:
- a confidence value may be generated after a conversation has occurred. For example, a confidence value may be generated for an action that was performed for user input (e.g., evaluate previous conversations between users and virtual assistants). While in other instances, a confidence value may be generated in real-time as a conversation is taking place.
- the confidence value may have various uses. For example, the confidence value may be used to determine whether or not an input received from a user corresponds to the intent unit that the input was mapped to. For instance, if the confidence value is below (or above) a predetermined threshold, it may be determined that there is a risk the input was incorrectly (or alternatively correctly) mapped to the corresponding intent unit. In another example, the confidence value may be used during a conversation.
- a predetermined response could be provided (e.g., Input: “How high is the Eiffel tower?” Answer: “I don't know a lot about world attractions, but I can do a Google® search for you”) and/or a follow-up question may be provided to obtain further information (e.g., for input of “what's the status,” a follow-up question of “are you referring to the status of a flight” may be presented).
- a task associated with the intent unit may be automatically performed. As such, the confidence value may provide context to interpret input and/or formulate a response.
- the risk indicator module 212 may further allow a user of the service provider 106 to configure a weight associated with the individual risk indicators, or factors, stored in the risk indicator data store 222 . For example, a risk indicator that an administrator determines to be more important, or determinative, in determining whether an input is mapped to the appropriate intent unit may be weighted more heavily than other risk indicators.
- the techniques and architectures of the risk indicator module 212 may be utilized to evaluate conversation data in a variety of contexts.
- conversation data that is obtained from a conversation between a user and a virtual assistant may be analyzed.
- conversation data from a conversation between a user and another user may be analyzed.
- a conversation between a chat bot and a user may be analyzed.
- the voting module 214 may be configured to allow voters to vote on whether or not user input was correctly mapped to an intent unit.
- the voters may answer “yes,” “no,” or “unsure” when asked whether the input was correctly mapped to the intent unit.
- the voters may be selected from a pre-approved voter list. The voters may have been determined to have knowledge sufficient to vote on mappings between intent units and inputs. In some examples, a particular voter may only be approved for certain intent units. In some instances, a predetermined number of voters may vote on a same item (e.g., a same intent unit mapped to user input).
- two or more voters may be prompted for their feedback regarding mapping of user input to intent units. This may help eliminate bias between individual voters. Further details regarding interfaces of the voting module 214 are described in greater detail below with respect to FIG. 6 .
- the voting module 214 may access conversations that are stored in a virtual assistant conversation data store 218 .
- the virtual assistance conversation data store 218 may include conversations that have inputs mapped to a same intent unit.
- the conversation review module 216 may allow a user, voter, or an administrator to review inputs and their respective intent units (e.g., tasks and responses), as well as votes received via the voting module 214 from voters.
- the conversation review module 216 may allow a user (e.g., administrator) to review a conversation including an input and which intent unit the input was mapped to. Additional information may be provided to an administrator, such as a type of input (e.g., test question, current input, etc,). Further, the conversation review module 216 may present to an administrator the voting results collected from the voting module 214 , and suggest an action to be taken based on the voting results.
- Various implementations of the conversation review module 216 are discussed further with respect to FIG. 8 .
- the risk indicator module 212 may be employed to evaluate weights associated with the risk indicators. For example, based on votes cast by voters, and the risk indicators used, a determination can be made to weight a risk indicator move heavily, or less heavily. For example, after viewing one or more conversations, it may be determined that a particular risk indicator generally indicates that input is correctly mapped to the intent unit and determined that the voters also indicate that the input should be mapped to the intent unit. In cases such as this, the particular risk indicator may be identified as being an effective indicator of a measure of risk. Thus, the risk indicator may be relied on more heavily, or weighted more heavily, in determining risk in a language model.
- modules 208 - 216 are illustrated as being included in the service provider 106 , in some instances one or more of these modules may be included in the computing device 112 , the smart device 102 , or elsewhere. As such, in some examples the service provider 106 may be eliminated entirely, such as in the case when all processing is performed locally. In addition, in some instances any of the data stores 218 - 222 may be included elsewhere.
- FIG. 3 illustrates example details of an example computing device 300 , such as the computing device 110 and/or a computing device associated with the service provider 106 (e.g., an output device).
- the computing device 300 may be equipped with one or more processors 302 , memory 304 , one or more cameras 306 , one or more displays 308 , one or more microphones 310 , one or more projectors 312 , one or more speakers 314 , and/or one or more sensors 316 .
- the components 304 - 316 may be communicatively coupled to the one or more processors 302 .
- the one or more processors 302 may include a central processing unit (CPU), a graphics processing unit (GPU), a microprocessor, a digital signal processor, and so on.
- the one or more cameras 306 may include a front facing camera and/or a rear facing camera.
- the one or more displays 308 may include a touch screen, a Liquid-crystal Display (LCD), a Light-emitting Diode (LED) display, an organic LED display, a plasma display, an electronic paper display, or any other type of technology.
- the one or more sensors 316 may include an accelerometer, compass, gyroscope, magnetometer, Global Positioning System (GPS), olfactory sensor (e.g., for smell), or other sensor.
- the components 306 - 316 may be configured to receive user input, such as gesture input (e.g., through the camera), touch input, audio or speech input, and so on, and/or may be configured to output content, such as audio, images, video, and so on.
- user input such as gesture input (e.g., through the camera), touch input, audio or speech input, and so on
- content such as audio, images, video, and so on.
- the one or more displays 308 , the one or more projectors 312 , and/or the one or more speakers 314 may comprise a content output device configured to output content and/or a virtual assistant.
- the computing device 300 may also include one or more network interfaces.
- the memory 304 may include a client application 318 , such as a web browser or application (e.g., mobile application, desktop application, etc.).
- client application 318 may be configured to output a voting interface and enable voters to provide feedback.
- the client application 318 may output an interface to evaluate risk associated with an intent input and/or votes.
- the client application 318 may be configured to facilitate any of the interfaces described below with respect to FIGS. 4-8 .
- FIGS. 4-8 illustrate example interfaces that may be presented to users, voters, administrators, and others.
- the interfaces may be provided via a web browser, an application (e.g., mobile application, desktop application, etc.), and so on.
- an application e.g., mobile application, desktop application, etc.
- FIGS. 4-8 illustrate example interfaces that may be presented to users, voters, administrators, and others.
- the interfaces may be provided via a web browser, an application (e.g., mobile application, desktop application, etc.), and so on.
- an application e.g., mobile application, desktop application, etc.
- FIG. 4 illustrates an example intent unit selection interface 400 that may be presented to a user to select one of intent units 402 to evaluate, or to be output for evaluation.
- the user may be an administrator of the service provider 106 , while in other instances the user may be a voter or any other user.
- the intent units 402 may be included in a natural language model, and may comprise logic, rules, or algorithms used to map tasks, actions, or responses to input received during a conversation between a virtual agent and a user. Stated otherwise, the intent units 402 may comprise a class, or grouping, of responses that are used to respond to one or more user purposes or goals.
- the intent units 402 may be used to analyze a dataset of information, such as a chat/conversation history between a human and a virtual assistant.
- a dataset may be created that is composed of conversations involving a virtual assistant where input received from a human is mapped to an intent unit to provide a response or action.
- various factors e.g., risk indicators
- risk indicators may be used to determine a measure of confidence that the input was mapped to was the correct intent unit.
- health statuses 404 ( 1 ), 404 ( 2 ) . . . 404 (N) may be calculated for individual units.
- the health statuses 404 may comprise a ratio of a number of user inputs that have a confidence value below a threshold relative to a total number of user inputs that have been mapped to the intent unit. As illustrated, the health status 404 ( 1 ) for the intent unit “Missing flight credit” may have an overall health status of 4.49%.
- each of the intent units 402 may be sorted (ranked) in a health status column 408 based on an associated health status 404 to identify which intent unit(s) are associated with the highest percentage of risk. Intent units that rank towards the top may have the highest need to be evaluated and/or modified. Thus, by ordering the intent units 402 based on health statuses 404 , users may quickly identify which intent unit(s) to select for review and evaluation.
- the intent unit selection interface 408 may further have an “add to voting” option 408 for of the intent units 404 .
- a user may check, or select, the intent units 402 that they would like to have analyzed further.
- the user may select a submit button 410 to output the selected intent unit for voting.
- the selected intent units may be output to a group of voters.
- the voters may be selected from approved lists of voters. For examples, the voters may have already been determined to be competent or qualified to evaluate conversations specific to various intent units.
- FIG. 5 illustrates an example chart 500 for displaying health status associated with intent units.
- the chart 500 may comprise a bar graph.
- other charts may be used such as a pie chart, line graph, or any other type of chart for displaying information.
- the chart 500 may plot information, such as intent unit(s), along an x-axis.
- information such as inputs (e.g., number of user inputs) may be plotted on a y-axis of the chart 500 .
- the chart 500 may display intent unit(s) versus inputs contained in a dataset of conversations mapped to intent unit(s).
- intent unit(s) may each have a bar 502 indicates a total number of user inputs that have mapped to that intent unit and a bar 504 indicates a number of user inputs that have mapped to the intent unit and are associated with less than a threshold level of confidence (or higher than a threshold level of risk).
- the bar 504 may indicate a percentage of the inputs that are risky. In this way, it may be determined visually which intent unit(s) are associated with the largest number of inputs, as well as the riskiest user inputs.
- the chart 500 may further have one or more tabs 506 that, when selected, change the information, or a display of information, contained in the chart 500 . For example, by selecting the tab “voting queue” of the one or more tabs 506 , the chart 500 may switch to displaying information in another arrangement, such as the arrangement shown in the intent unit selection interface 400 .
- FIG. 6 illustrates an example feedback interface 600 to allow a user to review and provide feedback on mapping of inputs to intent units.
- the feedback interface 600 may allow a user to be prompted as to whether an intent unit is correctly mapped to user input 602 .
- a virtual agent may have determined that the user input 602 should be mapped to a response of intent unit.
- the intent unit corresponds to “change a reservation,” and the user input 602 may be a question from a user, such as “can we change our flight reservations to leave for home sooner from Maui to Portland today?”
- additional information relating to the intent unit may be presented, such as the intent 604 , sample questions 606 , and related units 608 .
- the intent 604 may comprise a summary of the intent deemed to be associated with the displayed intent unit. Using this displayed intent 604 , a user may more easily be able to determine whether the user input 602 was correctly mapped to the intent unit.
- the sample questions 606 may comprise inputs that represent the type of language that the intent unit is meant to answer.
- the intent unit in this example may be represented by inputs such as “I want to change a flight plan,” or “can I use my ticket on a different date?”
- the related units 608 may comprise other intent units which are similar to the one the user is viewing.
- a user may select a related unit to view its associated intent.
- the intent 604 , the sample questions 606 , and the related units 608 may be displayed to help a user, or voter, make an informed decision about whether or not the input 602 is correctly mapped to the displayed intent unit.
- the user may determine, based on the intent 604 , the sample questions 606 , and/or the related units 608 , whether or not the user input 602 was incorrectly mapped to intent unit.
- the user may select one or more vote buttons 610 to specify a “yes,” “unsure,” and “no” answer.
- a user may use navigation buttons 612 to navigate to a next, or previous, unit. Thus, using the navigation buttons 612 , a user may navigate to a next item to evaluate.
- the feedback interface 600 may be provided to any number of users so that a relatively large number of votes may be obtained. This may potentially avoid bias across users. Further, by providing an intent unit instead of a response that is provided by a virtual assistant, the users may better evaluate the interactions with the virtual assistant. This may also avoid potential user bias.
- FIG. 7 illustrates an example risk factor interface 700 to allow a user to configure risk factors.
- the risk factor interface 700 may allow a user, such as an administrator of a language processing model, to weight risk factors.
- the risk factors may be various characteristics associated with a conversation between a human and a virtual assistant that indicate a measure of confidence that an input was mapped to a correct intent unit.
- the risk factor interface 700 may allow a user to select various weights for factors that may be specific to certain datasets. For example, a dataset that contains a relatively small set of conversations, or that is “immature,” may output more IDK responses as there is less training data available to aid in mapping of inputs to intent units. Thus, a user may want to weight IDK response with less importance for a particular dataset as it tends to occur often in an immature dataset that does not have a lot of training data (e.g., test questions).
- the risk factor interface 700 may be used to configure factors for a newly created dataset. In other examples, the risk factor interface 700 may be used to reconfigure factors for an existing dataset.
- a user may select one of “length of refinement” buttons 702 to identify a range of time for conversations, test questions, or other stored items. For example, a user may select a “0-12 month” time period, which may indicate that only test questions and/or conversations that have been obtained in the previous 12 months may be considered.
- the test questions 704 , the potential new inputs 706 , and the current inputs 708 may have slider bars associated with one or more factors, or predictors, where the slider bars allow a user to configure an amount of importance (e.g., weight) they would like to assign to the factors or predictors.
- the amount of importance may be identified as “low,” “medium,” or “high.” Alternatively, the amount of importance may be on any scale.
- the predictors, or panel of predictors may assign risk to an input based on the predictor's confidence that the intent unit the input was mapped to is correct, or appropriate.
- a user may wish to add the test questions 704 to a dataset.
- the test questions 704 may have a single indication of risk when evaluating the measure of risk.
- the test questions 704 may assign risk to an input based on the predictor's measure of confidence that the intent unit to which an input is mapped.
- a slider bar may enable a user to select an amount of importance assigned to the test questions.
- an icon e.g., an “x” or a “ ⁇ ” may be next to one or more slider bars to allow a user to turn on, or off, the evaluation of test questions, or other categories.
- the current inputs 708 may have a panel of to adjust weighting of risk indicators, such as “IDK triggered” and “sequential hits.”
- a slider bar associated with “IDK triggered” may configure a weighting applied to input that triggered an IDK response.
- a slide bar for “sequential hits” may configure a weighting applied to input that is associated sequential hits, or sequential mappings, to an intent unit during the conversation.
- a slider bar for “ties” may be used to configure a weighting applied to ties of mapping an input in a conversation to intent units.
- IDK triggered may have weightings assigned to sub-items, such as “input triggers IDK” and “input preceded IDK.” In this way, “IDK triggered” may be weighted more heavily when an input triggers the IDK, or when the input precedes the IDK.
- the corresponding dataset may be updated based on the weighted factors. While the example illustrates manually assigning weights to factors using a scroll bar, or slider bar, it is understood that in other examples, the weighting may be configured automatically based on feedback received from human voters. For example, if human voters consistently determine that an input is correctly mapped to an intent unit, and that a particular risk factor has consistently determined that the input was mapped to that intent unit, the weight associated with that particular risk factor may be increased as it shows accuracy in determining risk.
- FIG. 8 illustrates an example feedback results interface 800 for reviewing feedback received from voters regarding conversations.
- the feedback results interface 800 may display voting results received from one or more voters.
- the feedback results interface 800 may be viewed by a user, such as an administrator, to allow the administrator to view results and take action on mappings between inputs and intent units.
- one or more columns of information may be displayed, such as an input column 802 , a unit hit column 804 (e.g., for an intent unit), an input type column 806 , a voting results column 808 , a voting majority column 810 , and an action column 812 .
- the input column 802 may identify the input text.
- the unit hit column 804 may identify the unit hit by (e.g., mapped to) the input.
- the input type column 806 may identify whether the input was a test question, current input, or potential new input.
- the voting results column 808 may show the outcome of votes for determining whether the input was correctly mapped to an intent unit.
- the voting results may be indicated by three boxes indicating the amount of voters who said “yes,” “no,” or were “unsure” as to whether the input was correctly mapped to the intent unit.
- the voting majority column 810 may identify whether a majority had come to a decision regarding whether the input was correctly mapped to the intent unit.
- the action column 812 may provide a recommended action for the user, or administrator to take.
- the input type 806 has been identified as a “potential new input.”
- the voting results indicate that the voters, in this case three voters, all agreed that the potential new input was correctly mapped to the intent unit. While the boxes are illustrated as being ordered in the voting results column in order of “yes,” “no,” or “unsure” (from left to right), other variations may be presented. Additionally, the boxes may be color coordinated (e.g., “yes” box is green, “no” box is red, etc.) to visually illustrate how the voters cast their votes.
- the action 812 suggested to the administrator is to add the new input and adjust the language model.
- an input type may 806 may be a test question
- the voting results 808 may indicate that the three voters could not come to a majority as to whether the test question had correctly mapped to an appropriate intent unit.
- an action 812 may be to remove the test question because it may not be an appropriate or helpful question to train a dataset.
- row 818 may have an action such as “adjust model.” This may occur in examples where the input type 806 is current input and the voting results 808 indicate that all three voters agree that the input was incorrectly mapped to the corresponding intent unit. In instances such as this, because the input type 806 is a current input, and/or the voters unanimously agree that the input was incorrectly mapped to an intent unit, the language model may require updates. In instances such as these, an administrator may need to perform further analysis of the language model to determine what errors caused the mistake in mapping.
- row 820 may have an action 812 such as “no action needed.” This may occur in examples where the input type 806 is a current input, and the voting results indicate that all three voters agree that the input was correctly mapped to the corresponding intent unit. In instances such as this, because the input type 806 is a current input, and/or the voters unanimously agree that the input was correctly mapped to an intent unit, no action needs to be taken.
- FIGS. 9-11 illustrate example processes 900 , 1000 , and 1100 for employing the techniques described herein.
- the processes 900 , 1000 , and 1100 are described as being performed in the architecture 100 of FIG. 1 .
- one or more of the individual operations of the processes 900 , 1000 , and 1100 may be performed by the service provider 106 , the computing device 112 , and/or the smart device 102 .
- the processes 900 , 1000 , and 1100 may be performed in other architectures.
- the architecture 100 may be used to perform other processes.
- the processes 900 , 1000 , and 1100 are illustrated as a logical flow graph, each operation of which represents a sequence of operations that can be implemented in hardware, software, or a combination thereof.
- the operations represent computer-readable instructions stored on one or more computer-readable storage media that, when executed by one or more processors, perform the recited operations.
- computer-readable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types.
- the order in which the operations are described is not intended to be construed as a limitation, and any number of the described operations can be combined in any order and/or in parallel to implement the process. Further, any number of the described operations may be modified or omitted.
- FIG. 9 is a flowchart of an illustrative process 900 of for determining a measure of confidence that input received from a conversation involving a virtual assistant is mapped to a correct intent unit, and receiving feedback from a voter regarding accuracy of the mapping.
- a conversation user interface may be output to a device to enable a conversation.
- the user interface may be displayed on a display device.
- the conversation may take place between a user and a virtual assistant, two or more users, two or more virtual assistants, or combinations thereof.
- user input may be received, for example, through the conversation user interface.
- the user input may be processed to identify an intent of the user input.
- an intent may be characterized as an intent unit.
- the processing may comprise processing the user input with one or more natural language processing techniques.
- an intent unit may be associated with a language model for one or more of the natural language processing techniques.
- an intent unit may be associated with (i) an action to be performed at least partly by a virtual assistant, (ii) a pattern of components for triggering the intent unit, or (iii) combinations thereof.
- a weighting may be applied to one or more risk factors. For example, a plurality of risk indicators may be identified and a weighting may be applied to each risk indicator generating a weighted risk indicators.
- a measure of confidence that the intent unit is correctly identified for the user input may be determined. For example, the determining may be based at least in part on the presence of one or more risk indicators for the user input. Additionally or alternatively, the determining may be based at least in part on the weighted risk indicators.
- a measure of confidence may be associated with the user input.
- a health status of the intent unit may be determined where the health status may indicate a level of risk associated with the intent unit.
- a health status of the intent unit may be determined based at least in part on one or more measures of confidence.
- one or more intent units may be ranked with respect to each other. For example, an intent unit and another intent unit may be ranked based at least in part on the health status of the intent unit and a health status of the other intent unit. For example, a first intent unit may have a health status that indicates a relatively lower risk when compared to a second intent unit with a health status that indicates a relatively higher risk.
- the health statuses may be presented. For example, the health statuses may be displayed to a user. One or more health statuses may be presented. All of the health statuses may be presented, for example, in ranked order. Various embodiments contemplate that only health statuses above or below a threshold will be presented. Various embodiments contemplate that only a set number of highest or lowest health statuses will be presented. Various embodiments contemplate that a user may use the rankings to select which health status and associated intent units and inputs to focus on. For example, a health status that indicates a relatively large number of hits and relatively large amount of risk may be selected to work on over a health status that indicates a lower number of hits and or a lower amount of risk.
- the health status of the intent unit may be presented, for example, via an output device associated with an administrator.
- a selection of the intent unit may be received to obtain feedback regarding the intent unit.
- a feedback interface may be presented that enables a voter to provide feedback regarding matching of the user input to the intent unit.
- feedback may be received for the voter regarding an accuracy of matching the user input to the intent unit.
- the feedback may comprise a vote indicating whether or not the user input matches the intent unit.
- the feedback may be evaluated to determine whether the feedback indicates that the matching of the user input to the intent unit is accurate. For example, if yes, at 930 , a weight of a risk indicator may be reduced, for example, based at least in part on the determining that the feedback indicates that the matching of the user input to the intent unit is accurate. However, if no, then at 932 , a weight of a risk indicator may be increased, for example, based at least in part on the determining that the feedback indicates that the matching of the user input to the intent unit is not accurate. However, various embodiments contemplate that based on how the implemented system is configured, at 930 , the weight may be held steady or increased. Similarly, at 932 , the weight may be held steady or decreased.
- the weighting may be applied to the risk indicator.
- the intent unit may be evaluated based at least in part on the feedback.
- the evaluation my comprise presenting a feedback results interface that indicates at least one of a number of votes that are associated with an accurate matching of the user input to the intent unit or a number of votes that are associated with an inaccurate matching of the user input to the intent unit.
- a measure of confidence may be determined based at least in part on the presence and/or measure of the one or more weighted risk indicators.
- FIG. 10 illustrates an example process 1000 for determining and utilizing confidence values of intents associated with user inputs.
- conversation data may be received.
- the conversation data may be received as part of a conversation.
- one or more risk factors may be identified to evaluate conversation data associated with one or more users.
- the conversation data may represent at least one conversation.
- the conversation is between a user and a virtual assistant, one or more users and one or more virtual assistants, two or more users, two or more virtual assistants, or combinations thereof.
- the risk factors may include any and all risk indicators including, but not limited to, the following: user feedback from a user regarding an evaluation of at least a portion of the least one conversation; user feedback from an administrator regarding an evaluation of at least a portion of the at least one conversation; a tone of a user's voice during the at least one conversation; a gesture of a user during the at least one conversation; a facial expression of a user during the at least one conversation; a sensor signal obtained from monitoring user response; a message from a user about the at least one conversation; a confidence value determined for the at least one conversation from at least one of a probabilistic model or statistical model, or combinations thereof.
- risk indicators including, but not limited to, the following: user feedback from a user regarding an evaluation of at least a portion of the least one conversation; user feedback from an administrator regarding an evaluation of at least a portion of the at least one conversation; a tone of a user's voice during the at least one conversation; a gesture of a user during the at least one conversation; a facial expression of
- the risk factors may indicate, but is not limited to, any of the following: whether or not the processing identifies an intent for the user input; whether or not the user input directly preceded other user input for which the natural language processing system failed to identify an intent; whether or not the user input is involved in a conversation in which the natural language processing system failed to identify an intent; whether or not the user input is one of multiple user inputs for which the natural language processing system identifies a same intent two or more times in a row; whether or not the user input is involved in a conversation in which the natural language processing system identifies a same intent two or more times in a row; whether or not the user input is involved in a conversation in which the natural language processing system identifies a same intent two or more times, or combinations thereof.
- the conversation data may be processed with a natural language processing system to identify an intent for the conversation data.
- a weighting may be applied to each risk factor.
- a confidence value may be determined for the conversation data based at least in part on the one or more risk factors.
- the confidence value indicates a level of confidence that the intent is accurately identified for the conversation data. Additionally or alternatively, determining the confidence value may be based at least in part on one or more of the plurality of weighted risk factors.
- the confidence value may be utilized. For example, at 1014 , it may be determined that multiple pieces of user input are each associated with a confidence value that is below a threshold amount.
- the conversation data includes the multiple pieces of user input.
- user characteristic includes a demographic characteristic of a user.
- a user characteristic includes context that is used to interpret user input of the conversation data.
- context includes at least one of a type of device that is used by a user to provide the user input, purchase history, content that has been viewed by the user, or combinations thereof.
- a correlation between a user characteristic and the multiple pieces of user input may be determined.
- the correlation may be presented, for example, via an output device.
- the confidence value may be utilized to, for example, determine whether an action associated with the intent should be performed.
- the confidence value may be utilized to, for example, determine a health status that indicates a level of risk associated with the intent.
- the intent may be ranked with other intents.
- the ranking may be based at least in part on a respective health status of each of the intents.
- the ranks may be presented.
- the confidence value may be utilized to update a language model associated with the natural language processing system.
- FIG. 11 shows an illustrative process 1100 to utilize confidence values. For example, at 1102 , a confidence value may be compared to a threshold value. If the confidence value is below the threshold, various actions may be taken.
- a feedback interface may be presented.
- the feedback interface may enable a voter to vote regarding matching a user input, for example, from conversation data, to an intent.
- feedback from the voter may be received.
- the feedback may comprise a vote regarding the accuracy of matching the user input with the intent unit.
- a profile factor may be determined for the voter.
- the profile factor may be based at least in part on a comparison of a previous number of votes from the voter to votes from a majority of other voters.
- a first voter may have a history of providing votes that are consistent with a majority of votes for each voting opportunity.
- the voter profile factor may be relatively high (or increased to be higher than a threshold).
- the first voter may have a history of providing votes that are consistent with a minority.
- the voter profile factor may be relatively low (or decreased to be lower than a threshold).
- the first voter may have a relatively recent history of voting with the majority or minority.
- a predetermined number of historical votes, a predetermined time period of historical votes, a recent streak of votes consistent with the majority or minority, or combinations thereof may be used in the comparison.
- a majority of voter may comprise a predetermined number of voters.
- a weight may be applied to the vote of the voter.
- the weight may be based at least in part on the profile factor of the voter.
- the weight applied to the vote may create a weighted vote.
- whether the feedback indicates that the matching is accurate may be determined.
- a weight may be adjusted and/or applied to a risk factor.
- the weight may be adjusted or modified based at least in part on the determining whether the feedback indicates that that matching of the user input to the intent is accurate. Additionally or alternatively, the weight may be adjusted based at least in part on one or more weighted votes. Additionally or alternatively, the weight may be adjusted or applied to the risk factors based at least in part on a maturity of intent units associated with the natural language processing system, or an amount of user inputs for the conversation data, or a combination thereof. Additionally or alternatively, the weight may be adjusted or applied based at least in part on recent changes to the language model, underlying knowledge base, services of the system, among others, or combinations thereof. For example, a relatively mature intent unit may act relatively immature if the application of the language model or knowledge base (among others) changes.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- User Interface Of Digital Computer (AREA)
- Machine Translation (AREA)
- Signal Processing (AREA)
Abstract
Description
-
- user feedback from a user regarding an evaluation of at least a portion of a conversation (e.g., a rating provided by a user during the conversation or after of how well the user felt his questions were answered, information from a survey regarding a conversation, etc.)—the feedback may cover a single response for a single input or may cover multiple responses and/or inputs;
- user feedback from an administrator (e.g., a user associated with a virtual assistant service) regarding an evaluation of at least a portion of a conversation;
- a user feedback from a business entity (e.g., a business who uses the virtual assistant service for their business) regarding an evaluation of at least a portion of a conversation;
- a tone of a user's voice during a conversation (e.g., the user sounds upset or happy, the user uses a loud voice when communicating (above a threshold volume), etc.);
- a gesture of a user during a conversation (e.g., the user shakes his head when a response was provided, the user provides a thumbs-down, the user provides other body movement, etc.)—the gesture may be obtained from a camera or other sensor;
- a facial expression of a user during a conversation (e.g., the user raised an eye brow, frowned, etc.);
- a sensor signal obtained from monitoring user response (e.g., a hear rate of the user, eye movement, movement of device, placing a phone into a pocket (identified from light sensor), closing an application, ending a conversation, etc.)—a sensory may include a heart rate monitor, eye tracking device, and so on;
- a message from a user about a conversation (e.g., the user posts a comment on a blog or social media site about a bad experience with a virtual assistant, the user sends a text message about a conversation, etc.);
- a confidence value determined by one or more Natural Language Processing (NLP) systems including, for example, a comparison of natural language understanding of user input (e.g., a confidence value from a probabilistic model or statistical model (sometimes referred to as a “classifier”))—an intent unit that is identified for an input may be analyzed by a model that generates its own confidence value of how accurately the intent unit satisfies the input;
- whether or not a classifier, trained via machine learning using an independent dataset, maps the input to an intent unit different than that of an intent unit mapped to the input by the natural language processing system;
- whether or not the input triggered a particular intent unit (e.g., insult or apology intent unit) indicating that a user is frustrated or upset;
- whether or not the input is involved in a conversation that includes an escalation (e.g., a virtual assistant was unable to provide a response and the conversation transferred to a human assistant to continue the conversation);
- whether or not the input preceded an escalation;
- whether or not the input is mapped to an intent unit associated with answering threats, swearing, or cursing;
- whether or not the input contained a word unknown to the natural language processing system;
- whether or not the input is mapped to an intent unit (e.g., whether or not an intent unit is found) (a failure to identify an intent unit is referred to as an “I don't know” (IDK) response indicating that a response in a language model is not identified);
- whether or not the user input directly preceded other user input for which a natural language processing system failed to identify an intent;
- whether or not the user input is involved in a conversation in which the natural language processing system failed to identify an intent (e.g., whether or not the input is involved in a conversation that included a particular response indicating that a response in a language model is not identified for the input);
- whether or not the user input is one of multiple user inputs for which the natural language processing system identifies a same intent two or more times overall and/or in a row (e.g., whether or not the input triggers a response that is involved in two or more sequential hits to a same intent unit);
- whether or not the input is involved in a conversation in which a natural language processing system identifies a same intent two or more times overall and/or in a row;
- whether or not the input is involved in a conversation in which two or more intent units are identified for the same input (e.g., a tie);
- whether or not the input is mapped to an intent unit that is designated as being associated with personality (e.g., personality intent unit—an intent unit that is not directly related to the subject matter being discussed, such as an intent unit related to your favorite color when the subject matter is booking a flight);
- whether or not a threshold number of personality intent units are hit (e.g., within a window of time, in a row, within a number of input interactions, etc.);
- a combination of any of the above factors; or
- any other factor.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/242,639 US10545648B2 (en) | 2014-09-09 | 2019-01-08 | Evaluating conversation data based on risk factors |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462048144P | 2014-09-09 | 2014-09-09 | |
US201462049982P | 2014-09-12 | 2014-09-12 | |
US14/849,541 US20160071517A1 (en) | 2014-09-09 | 2015-09-09 | Evaluating Conversation Data based on Risk Factors |
US15/857,160 US10175865B2 (en) | 2014-09-09 | 2017-12-28 | Evaluating conversation data based on risk factors |
US16/242,639 US10545648B2 (en) | 2014-09-09 | 2019-01-08 | Evaluating conversation data based on risk factors |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/857,160 Continuation US10175865B2 (en) | 2014-09-09 | 2017-12-28 | Evaluating conversation data based on risk factors |
Publications (2)
Publication Number | Publication Date |
---|---|
US20190138190A1 US20190138190A1 (en) | 2019-05-09 |
US10545648B2 true US10545648B2 (en) | 2020-01-28 |
Family
ID=55438072
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/849,541 Abandoned US20160071517A1 (en) | 2014-09-09 | 2015-09-09 | Evaluating Conversation Data based on Risk Factors |
US15/857,160 Active 2035-09-14 US10175865B2 (en) | 2014-09-09 | 2017-12-28 | Evaluating conversation data based on risk factors |
US16/242,639 Active US10545648B2 (en) | 2014-09-09 | 2019-01-08 | Evaluating conversation data based on risk factors |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/849,541 Abandoned US20160071517A1 (en) | 2014-09-09 | 2015-09-09 | Evaluating Conversation Data based on Risk Factors |
US15/857,160 Active 2035-09-14 US10175865B2 (en) | 2014-09-09 | 2017-12-28 | Evaluating conversation data based on risk factors |
Country Status (1)
Country | Link |
---|---|
US (3) | US20160071517A1 (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210237757A1 (en) * | 2020-01-31 | 2021-08-05 | Toyota Jidosha Kabushiki Kaisha | Information processing device, information processing method, and storage medium storing information processing program |
US11328016B2 (en) | 2018-05-09 | 2022-05-10 | Oracle International Corporation | Constructing imaginary discourse trees to improve answering convergent questions |
US11373632B2 (en) | 2017-05-10 | 2022-06-28 | Oracle International Corporation | Using communicative discourse trees to create a virtual persuasive dialogue |
US11386274B2 (en) * | 2017-05-10 | 2022-07-12 | Oracle International Corporation | Using communicative discourse trees to detect distributed incompetence |
US11436416B2 (en) | 2019-06-06 | 2022-09-06 | Verint Americas Inc. | Automated conversation review to surface virtual assistant misunderstandings |
US20220284194A1 (en) * | 2017-05-10 | 2022-09-08 | Oracle International Corporation | Using communicative discourse trees to detect distributed incompetence |
US11455494B2 (en) | 2018-05-30 | 2022-09-27 | Oracle International Corporation | Automated building of expanded datasets for training of autonomous agents |
US20220382990A1 (en) * | 2018-10-10 | 2022-12-01 | Verint Americas Inc. | System for minimizing repetition in intelligent virtual assistant conversations |
US11586827B2 (en) * | 2017-05-10 | 2023-02-21 | Oracle International Corporation | Generating desired discourse structure from an arbitrary text |
US11615145B2 (en) | 2017-05-10 | 2023-03-28 | Oracle International Corporation | Converting a document into a chatbot-accessible form via the use of communicative discourse trees |
US11694037B2 (en) | 2017-05-10 | 2023-07-04 | Oracle International Corporation | Enabling rhetorical analysis via the use of communicative discourse trees |
US11748572B2 (en) | 2017-05-10 | 2023-09-05 | Oracle International Corporation | Enabling chatbots by validating argumentation |
US11783126B2 (en) | 2017-05-10 | 2023-10-10 | Oracle International Corporation | Enabling chatbots by detecting and supporting affective argumentation |
US11797773B2 (en) | 2017-09-28 | 2023-10-24 | Oracle International Corporation | Navigating electronic documents using domain discourse trees |
US11842144B1 (en) * | 2022-07-26 | 2023-12-12 | Rammer Technologies, Inc. | Summarizing conversational speech |
US11960844B2 (en) | 2017-05-10 | 2024-04-16 | Oracle International Corporation | Discourse parsing using semantic and syntactic relations |
US12001805B2 (en) * | 2022-04-25 | 2024-06-04 | Gyan Inc. | Explainable natural language understanding platform |
US12141535B2 (en) | 2017-05-10 | 2024-11-12 | Oracle International Corporation | Techniques for maintaining rhetorical flow |
Families Citing this family (147)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US10489434B2 (en) | 2008-12-12 | 2019-11-26 | Verint Americas Inc. | Leveraging concepts with information retrieval techniques and knowledge bases |
US20120311585A1 (en) | 2011-06-03 | 2012-12-06 | Apple Inc. | Organizing task items that represent tasks to perform |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US9634855B2 (en) | 2010-05-13 | 2017-04-25 | Alexander Poltorak | Electronic personal interactive device that determines topics of interest using a conversational agent |
US20180143989A1 (en) * | 2016-11-18 | 2018-05-24 | Jagadeshwar Nomula | System to assist users of a software application |
US11068954B2 (en) | 2015-11-20 | 2021-07-20 | Voicemonk Inc | System for virtual agents to help customers and businesses |
US9836177B2 (en) | 2011-12-30 | 2017-12-05 | Next IT Innovation Labs, LLC | Providing variable responses in a virtual-assistant environment |
US10417037B2 (en) | 2012-05-15 | 2019-09-17 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
AU2014214676A1 (en) | 2013-02-07 | 2015-08-27 | Apple Inc. | Voice trigger for a digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
KR101922663B1 (en) | 2013-06-09 | 2018-11-28 | 애플 인크. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
DE112014003653B4 (en) | 2013-08-06 | 2024-04-18 | Apple Inc. | Automatically activate intelligent responses based on activities from remote devices |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US20160071517A1 (en) | 2014-09-09 | 2016-03-10 | Next It Corporation | Evaluating Conversation Data based on Risk Factors |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10460227B2 (en) | 2015-05-15 | 2019-10-29 | Apple Inc. | Virtual assistant in a communication session |
US10200824B2 (en) | 2015-05-27 | 2019-02-05 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device |
US20160378747A1 (en) | 2015-06-29 | 2016-12-29 | Apple Inc. | Virtual assistant for media playback |
US10331312B2 (en) | 2015-09-08 | 2019-06-25 | Apple Inc. | Intelligent automated assistant in a media environment |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10740384B2 (en) | 2015-09-08 | 2020-08-11 | Apple Inc. | Intelligent automated assistant for media search and playback |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10269372B1 (en) * | 2015-09-24 | 2019-04-23 | United Services Automobile Association (Usaa) | System for sound analysis and recognition |
US11170172B1 (en) | 2015-09-28 | 2021-11-09 | Press Ganey Associates, Inc. | System and method for actionizing comments |
US10528671B1 (en) * | 2015-09-28 | 2020-01-07 | NarrativeDX Inc. | System and method for actionizing comments using voice data |
US9635167B2 (en) | 2015-09-29 | 2017-04-25 | Paypal, Inc. | Conversation assistance system |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US20170109671A1 (en) * | 2015-10-19 | 2017-04-20 | Adapt Ready Inc. | System and method to identify risks and provide strategies to overcome risks |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10956666B2 (en) | 2015-11-09 | 2021-03-23 | Apple Inc. | Unconventional virtual assistant interactions |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US11501175B2 (en) * | 2016-02-08 | 2022-11-15 | Micro Focus Llc | Generating recommended inputs |
US10885461B2 (en) | 2016-02-29 | 2021-01-05 | Oracle International Corporation | Unsupervised method for classifying seasonal patterns |
US10867421B2 (en) | 2016-02-29 | 2020-12-15 | Oracle International Corporation | Seasonal aware method for forecasting and capacity planning |
US11012719B2 (en) * | 2016-03-08 | 2021-05-18 | DISH Technologies L.L.C. | Apparatus, systems and methods for control of sporting event presentation based on viewer engagement |
US10984036B2 (en) | 2016-05-03 | 2021-04-20 | DISH Technologies L.L.C. | Providing media content based on media element preferences |
US12223282B2 (en) | 2016-06-09 | 2025-02-11 | Apple Inc. | Intelligent automated assistant in a home environment |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
US12197817B2 (en) | 2016-06-11 | 2025-01-14 | Apple Inc. | Intelligent device arbitration and control |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
US10019988B1 (en) * | 2016-06-23 | 2018-07-10 | Intuit Inc. | Adjusting a ranking of information content of a software application based on feedback from a user |
US20180018965A1 (en) * | 2016-07-12 | 2018-01-18 | Bose Corporation | Combining Gesture and Voice User Interfaces |
US10573299B2 (en) | 2016-08-19 | 2020-02-25 | Panasonic Avionics Corporation | Digital assistant and associated methods for a transportation vehicle |
US9972312B2 (en) * | 2016-08-19 | 2018-05-15 | Panasonic Avionics Corporation | Digital assistant and associated methods for a transportation vehicle |
US10754958B1 (en) * | 2016-09-19 | 2020-08-25 | Nopsec Inc. | Vulnerability risk mitigation platform apparatuses, methods and systems |
US10135989B1 (en) | 2016-10-27 | 2018-11-20 | Intuit Inc. | Personalized support routing based on paralinguistic information |
US11196826B2 (en) * | 2016-12-23 | 2021-12-07 | DISH Technologies L.L.C. | Communications channels in media systems |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
JP6800249B2 (en) * | 2017-01-20 | 2020-12-16 | 本田技研工業株式会社 | Conversation processing server, conversation processing server control method, and terminal |
US11227230B2 (en) * | 2017-03-17 | 2022-01-18 | International Business Machines Corporation | Automated technical content conversion based on user understanding level |
US10599885B2 (en) | 2017-05-10 | 2020-03-24 | Oracle International Corporation | Utilizing discourse structure of noisy user-generated content for chatbot learning |
DK180048B1 (en) | 2017-05-11 | 2020-02-04 | Apple Inc. | MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | Low-latency intelligent automated assistant |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK201770411A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | MULTI-MODAL INTERFACES |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
US20180336892A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Detecting a trigger of a digital assistant |
US10902840B2 (en) * | 2017-05-31 | 2021-01-26 | SAI Society for Advanced Scientific Research | Method and system for thought-to-speech |
CN107016996B (en) * | 2017-06-06 | 2020-11-10 | 广东小天才科技有限公司 | Audio data processing method and device |
US10949754B2 (en) * | 2017-07-26 | 2021-03-16 | Sap Portals Israel Ltd. | Decision support tool with interactive sliders |
DE102017213946B4 (en) * | 2017-08-10 | 2022-11-10 | Audi Ag | Method for processing a recognition result of an automatic online speech recognizer for a mobile terminal |
JP6479916B1 (en) * | 2017-09-07 | 2019-03-06 | ヤフー株式会社 | Information processing apparatus, information processing method, and program |
US10673787B2 (en) * | 2017-10-03 | 2020-06-02 | Servicenow, Inc. | Virtual agent conversation service |
US10574598B2 (en) * | 2017-10-18 | 2020-02-25 | International Business Machines Corporation | Cognitive virtual detector |
US11016729B2 (en) | 2017-11-08 | 2021-05-25 | International Business Machines Corporation | Sensor fusion service to enhance human computer interactions |
US10685648B2 (en) * | 2017-11-08 | 2020-06-16 | International Business Machines Corporation | Sensor fusion model to enhance machine conversational awareness |
US10951761B1 (en) | 2017-12-20 | 2021-03-16 | Wells Fargo Bank, N.A. | System and method for live and virtual support interaction |
US10733375B2 (en) * | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10521462B2 (en) * | 2018-02-27 | 2019-12-31 | Accenture Global Solutions Limited | Virtual services rapid deployment tool |
US10460734B2 (en) | 2018-03-08 | 2019-10-29 | Frontive, Inc. | Methods and systems for speech signal processing |
US20210358478A1 (en) * | 2018-03-19 | 2021-11-18 | Verint Americas Inc. | Model-agnostic visualizations using linear programming approximation |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US11843719B1 (en) * | 2018-03-30 | 2023-12-12 | 8X8, Inc. | Analysis of customer interaction metrics from digital voice data in a data-communication server system |
US11113473B2 (en) * | 2018-04-02 | 2021-09-07 | SoundHound Inc. | Interpreting expressions having potentially ambiguous meanings in different domains |
WO2019207379A1 (en) * | 2018-04-26 | 2019-10-31 | Reliance Jio Infocomm Limited | System and method for providing a response to a user query using a visual assistant |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US11169668B2 (en) * | 2018-05-16 | 2021-11-09 | Google Llc | Selecting an input mode for a virtual assistant |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
US11347966B2 (en) * | 2018-07-20 | 2022-05-31 | Samsung Electronics Co., Ltd. | Electronic apparatus and learning method of electronic apparatus |
US12118991B2 (en) * | 2018-07-20 | 2024-10-15 | Sony Corporation | Information processing device, information processing system, and information processing method |
US10698500B2 (en) | 2018-07-24 | 2020-06-30 | Bank Of America Corporation | Touch-based and audio-responsive user interface for elimination of a query's component terms |
CN110874201B (en) * | 2018-08-29 | 2023-06-23 | 斑马智行网络(香港)有限公司 | Interactive method, device, storage medium and operating system |
US10984198B2 (en) * | 2018-08-30 | 2021-04-20 | International Business Machines Corporation | Automated testing of dialog systems |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
CA3114298C (en) * | 2018-09-28 | 2024-06-11 | Element Ai Inc. | Recommendation method and system and method and system for improving a machine learning system |
US12001926B2 (en) | 2018-10-23 | 2024-06-04 | Oracle International Corporation | Systems and methods for detecting long term seasons |
US11374958B2 (en) * | 2018-10-31 | 2022-06-28 | International Business Machines Corporation | Security protection rule prediction and enforcement |
US10884903B1 (en) * | 2018-11-01 | 2021-01-05 | Intuit Inc. | Automatic production testing and validation |
US11004449B2 (en) * | 2018-11-29 | 2021-05-11 | International Business Machines Corporation | Vocal utterance based item inventory actions |
US10839167B2 (en) * | 2018-12-04 | 2020-11-17 | Verizon Patent And Licensing Inc. | Systems and methods for dynamically expanding natural language processing agent capacity |
US11164574B2 (en) * | 2019-01-03 | 2021-11-02 | International Business Machines Corporation | Conversational agent generation |
US11133026B2 (en) * | 2019-01-04 | 2021-09-28 | International Business Machines Corporation | Natural language processor for using speech to cognitively detect and analyze deviations from a baseline |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11960847B2 (en) * | 2019-04-04 | 2024-04-16 | Verint Americas Inc. | Systems and methods for generating responses for an intelligent virtual |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
DK201970509A1 (en) | 2019-05-06 | 2021-01-15 | Apple Inc | Spoken notifications |
US11537940B2 (en) * | 2019-05-13 | 2022-12-27 | Oracle International Corporation | Systems and methods for unsupervised anomaly detection using non-parametric tolerance intervals over a sliding window of t-digests |
US11874861B2 (en) * | 2019-05-17 | 2024-01-16 | International Business Machines Corporation | Retraining a conversation system based on negative feedback |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
CN111985751B (en) * | 2019-05-23 | 2023-09-26 | 百度在线网络技术(北京)有限公司 | Human-computer chat experience assessment system |
DK201970511A1 (en) | 2019-05-31 | 2021-02-15 | Apple Inc | Voice identification in digital assistant systems |
DK180129B1 (en) | 2019-05-31 | 2020-06-02 | Apple Inc. | USER ACTIVITY SHORTCUT SUGGESTIONS |
US11227599B2 (en) | 2019-06-01 | 2022-01-18 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
KR102339085B1 (en) * | 2019-07-22 | 2021-12-14 | 엘지전자 주식회사 | Artificial intelligence apparatus for recognizing speech of user in consideration of user's application usage log and method for the same |
US11423672B2 (en) * | 2019-08-02 | 2022-08-23 | Dish Network L.L.C. | System and method to detect driver intent and employ safe driving actions |
US11221897B2 (en) * | 2019-09-11 | 2022-01-11 | International Business Machines Corporation | Managing device maintenance via artificial intelligence |
US11887015B2 (en) | 2019-09-13 | 2024-01-30 | Oracle International Corporation | Automatically-generated labels for time series data and numerical lists to use in analytic and machine learning systems |
US10878008B1 (en) * | 2019-09-13 | 2020-12-29 | Intuit Inc. | User support with integrated conversational user interfaces and social question answering |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
WO2021074459A1 (en) * | 2019-10-16 | 2021-04-22 | Sigma Technologies, S.L. | Method and system to automatically train a chatbot using domain conversations |
US11423235B2 (en) * | 2019-11-08 | 2022-08-23 | International Business Machines Corporation | Cognitive orchestration of multi-task dialogue system |
US11158308B1 (en) * | 2019-11-27 | 2021-10-26 | Amazon Technologies, Inc. | Configuring natural language system |
KR20210072362A (en) | 2019-12-09 | 2021-06-17 | 엘지전자 주식회사 | Artificial intelligence apparatus and method for generating training data for artificial intelligence model |
US20210209441A1 (en) * | 2020-01-06 | 2021-07-08 | International Business Machines Corporation | Comparing performance of virtual assistants |
US10841251B1 (en) * | 2020-02-11 | 2020-11-17 | Moveworks, Inc. | Multi-domain chatbot |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11183193B1 (en) | 2020-05-11 | 2021-11-23 | Apple Inc. | Digital assistant hardware abstraction |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11363041B2 (en) | 2020-05-15 | 2022-06-14 | International Business Machines Corporation | Protecting computer assets from malicious attacks |
CA3178952A1 (en) | 2020-06-02 | 2021-12-09 | Jeffrey SALTER | Systems and method for intent messaging |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
US11551689B2 (en) * | 2020-09-30 | 2023-01-10 | International Business Machines Corporation | Voice command execution |
US11675820B2 (en) * | 2020-10-27 | 2023-06-13 | International Business Machines Corporation | Building and modifying conversational user journeys |
US11228644B1 (en) | 2020-11-10 | 2022-01-18 | Capital One Services, Llc | Systems and methods to generate contextual threads |
US12014836B2 (en) | 2020-12-10 | 2024-06-18 | International Business Machines Corporation | Stream integrity for artificial intelligence conversations |
US11816437B2 (en) * | 2020-12-15 | 2023-11-14 | International Business Machines Corporation | Automatical process application generation |
WO2023212258A1 (en) * | 2022-04-28 | 2023-11-02 | Theai, Inc. | Relationship graphs for artificial intelligence character models |
US11995457B2 (en) | 2022-06-03 | 2024-05-28 | Apple Inc. | Digital assistant integration with system interface |
US20240193190A1 (en) * | 2022-12-09 | 2024-06-13 | Asapp, Inc. | Automated key-value extraction using natural language intents |
US11990123B1 (en) * | 2023-06-24 | 2024-05-21 | Roy Rosser | Automated training of AI chatbots |
Citations (196)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5278980A (en) | 1991-08-16 | 1994-01-11 | Xerox Corporation | Iterative technique for phrase query formation and an information retrieval system employing same |
US5418948A (en) | 1991-10-08 | 1995-05-23 | West Publishing Company | Concept matching of natural language queries with a database of document concepts |
US5535120A (en) | 1990-12-31 | 1996-07-09 | Trans-Link International Corp. | Machine translation and telecommunications system using user ID data to select dictionaries |
US5615112A (en) | 1993-01-29 | 1997-03-25 | Arizona Board Of Regents | Synthesized object-oriented entity-relationship (SOOER) model for coupled knowledge-base/database of image retrieval expert system (IRES) |
US5677835A (en) | 1992-09-04 | 1997-10-14 | Caterpillar Inc. | Integrated authoring and translation system |
US5682539A (en) | 1994-09-29 | 1997-10-28 | Conrad; Donovan | Anticipated meaning natural language interface |
US5727174A (en) | 1992-03-23 | 1998-03-10 | International Business Machines Corporation | Graphical end-user interface for intelligent assistants |
US6012053A (en) | 1997-06-23 | 2000-01-04 | Lycos, Inc. | Computer system with user-controlled relevance ranking of search results |
US6112177A (en) | 1997-11-07 | 2000-08-29 | At&T Corp. | Coarticulation method for audio-visual text-to-speech synthesis |
US6144938A (en) | 1998-05-01 | 2000-11-07 | Sun Microsystems, Inc. | Voice user interface with personality |
US6175829B1 (en) | 1998-04-22 | 2001-01-16 | Nec Usa, Inc. | Method and apparatus for facilitating query reformulation |
US20010000356A1 (en) | 1995-07-07 | 2001-04-19 | Woods William A. | Method and apparatus for generating query responses in a computer-based document retrieval system |
US6282507B1 (en) | 1999-01-29 | 2001-08-28 | Sony Corporation | Method and apparatus for interactive source language expression recognition and alternative hypothesis presentation and selection |
US6285978B1 (en) | 1998-09-24 | 2001-09-04 | International Business Machines Corporation | System and method for estimating accuracy of an automatic natural language translation |
US20010033298A1 (en) | 2000-03-01 | 2001-10-25 | Benjamin Slotznick | Adjunct use of instant messenger software to enable communications to or between chatterbots or other software agents |
US20010044751A1 (en) | 2000-04-03 | 2001-11-22 | Pugliese Anthony V. | System and method for displaying and selling goods and services |
US20010049688A1 (en) | 2000-03-06 | 2001-12-06 | Raya Fratkina | System and method for providing an intelligent multi-step dialog with a user |
US20010053968A1 (en) | 2000-01-10 | 2001-12-20 | Iaskweb, Inc. | System, method, and computer program product for responding to natural language queries |
US20020008716A1 (en) | 2000-07-21 | 2002-01-24 | Colburn Robert A. | System and method for controlling expression characteristics of a virtual agent |
US6353817B1 (en) | 1998-06-26 | 2002-03-05 | Charles M Jacobs | Multi-user system for creating and maintaining a medical-decision-making knowledge base |
US20020032591A1 (en) | 2000-09-08 | 2002-03-14 | Agentai, Inc. | Service request processing performed by artificial intelligence systems in conjunctiion with human intervention |
US6388665B1 (en) | 1994-07-08 | 2002-05-14 | Microsoft Corporation | Software platform having a real world interface with animated characters |
US6396951B1 (en) | 1997-12-29 | 2002-05-28 | Xerox Corporation | Document-based query data for information retrieval |
US6401061B1 (en) | 1999-05-13 | 2002-06-04 | Yuri L. Zieman | Combinatorial computational technique for transformation phrase text-phrase meaning |
US20020123994A1 (en) | 2000-04-26 | 2002-09-05 | Yves Schabes | System for fulfilling an information need using extended matching techniques |
US20020129031A1 (en) | 2001-01-05 | 2002-09-12 | Lau Lee Min | Managing relationships between unique concepts in a database |
US20020198885A1 (en) | 2001-04-09 | 2002-12-26 | Streepy Larry V. | Method and system for interfacing with a multi-level data structure |
US20030004908A1 (en) | 2001-06-29 | 2003-01-02 | Linthicum Steven Eric | Method and system for automated maintenance and training instruction generation and validation |
US20030041307A1 (en) | 2001-08-24 | 2003-02-27 | Jong-Won Park | Electronic dictionary system with a tree structure and its output |
US20030061029A1 (en) | 2001-08-29 | 2003-03-27 | Efraim Shaket | Device for conducting expectation based mixed initiative natural language dialogs |
US20030088547A1 (en) | 2001-11-06 | 2003-05-08 | Hammond Joel K. | Method and apparatus for providing comprehensive search results in response to user queries entered over a computer network |
US20030126090A1 (en) | 2001-12-28 | 2003-07-03 | Fujitsu Limited | Conversation method, device, program and computer-readable recording medium on which conversation program is recorded |
US20030126089A1 (en) | 2001-12-28 | 2003-07-03 | Fujitsu Limited | Conversation method, device, program and computer-readable recording medium on which conversation program is recorded |
US20030142829A1 (en) | 2001-11-26 | 2003-07-31 | Cristiano Avigni | Systems and methods for determining sound of a moving object |
US20030212544A1 (en) | 2002-05-10 | 2003-11-13 | Alejandro Acero | System for automatically annotating training data for a natural language understanding system |
US6661418B1 (en) | 2001-01-22 | 2003-12-09 | Digital Animations Limited | Character animation system |
US20040107088A1 (en) | 1994-09-30 | 2004-06-03 | Budzinski Robert L. | Memory system for storing and retrieving experience and knowledge with natural language utilizing state representation data, word sense numbers, function codes, directed graphs and/or context memory |
US6757362B1 (en) | 2000-03-06 | 2004-06-29 | Avaya Technology Corp. | Personal virtual assistant |
US20040141013A1 (en) | 2003-01-21 | 2004-07-22 | Microsoft Corporation | System and method for directly accessing functionality provided by an application |
US20040186705A1 (en) | 2003-03-18 | 2004-09-23 | Morgan Alexander P. | Concept word management |
US6826540B1 (en) | 1999-12-29 | 2004-11-30 | Virtual Personalities, Inc. | Virtual human interface for conducting surveys |
US6829603B1 (en) | 2000-02-02 | 2004-12-07 | International Business Machines Corp. | System, method and program product for interactive natural dialog |
US6834120B1 (en) | 2000-11-15 | 2004-12-21 | Sri International | Method and system for estimating the accuracy of inference algorithms using the self-consistency methodology |
US20050027694A1 (en) | 2003-07-31 | 2005-02-03 | Volker Sauermann | User-friendly search results display system, method, and computer program product |
US20050054381A1 (en) | 2003-09-05 | 2005-03-10 | Samsung Electronics Co., Ltd. | Proactive user interface |
US20050120276A1 (en) | 1999-01-06 | 2005-06-02 | Parasoft Corporation | Modularizing a computer program for testing and debugging |
US20060004826A1 (en) | 2004-05-04 | 2006-01-05 | Mark Zartler | Data disambiguation systems and methods |
US6987514B1 (en) | 2000-11-09 | 2006-01-17 | Nokia Corporation | Voice avatars for wireless multiuser entertainment services |
US20060020466A1 (en) | 2004-07-26 | 2006-01-26 | Cousineau Leo E | Ontology based medical patient evaluation method for data capture and knowledge representation |
US6999932B1 (en) | 2000-10-10 | 2006-02-14 | Intel Corporation | Language independent voice-based search system |
US20060037076A1 (en) | 2004-05-04 | 2006-02-16 | Shantu Roy | Methods and systems for enforcing network and computer use policy |
US20060036430A1 (en) | 2004-08-12 | 2006-02-16 | Junling Hu | System and method for domain-based natural language consultation |
US20060047632A1 (en) | 2004-08-12 | 2006-03-02 | Guoming Zhang | Method using ontology and user query processing to solve inventor problems and user problems |
US20060067352A1 (en) | 2004-09-30 | 2006-03-30 | Ajita John | Method and apparatus for providing a virtual assistant to a communication participant |
US20060074689A1 (en) | 2002-05-16 | 2006-04-06 | At&T Corp. | System and method of providing conversational visual prosody for talking heads |
US20060074831A1 (en) | 2004-09-20 | 2006-04-06 | Hyder Andrew D | Virtual assistant |
US20060080107A1 (en) | 2003-02-11 | 2006-04-13 | Unveil Technologies, Inc., A Delaware Corporation | Management of conversations |
US20060092978A1 (en) | 2004-09-30 | 2006-05-04 | Ajita John | Method and apparatus for developing a virtual assistant for a communication |
US7058902B2 (en) | 2002-07-30 | 2006-06-06 | Microsoft Corporation | Enhanced on-object context menus |
US20060161414A1 (en) | 2004-12-15 | 2006-07-20 | C.R.F. Societa Consortile Per Azioni | Event-driven model generated from an ordered natural language interface |
US20060206483A1 (en) | 2004-10-27 | 2006-09-14 | Harris Corporation | Method for domain identification of documents in a document database |
US20060253427A1 (en) | 2005-05-04 | 2006-11-09 | Jun Wu | Suggesting and refining user input based on original user input |
US20070043687A1 (en) | 2005-08-19 | 2007-02-22 | Accenture Llp | Virtual assistant |
US7194483B1 (en) | 2001-05-07 | 2007-03-20 | Intelligenxia, Inc. | Method, system, and computer program product for concept-based multi-dimensional analysis of unstructured information |
US20070100790A1 (en) | 2005-09-08 | 2007-05-03 | Adam Cheyer | Method and apparatus for building an intelligent automated assistant |
US20070106670A1 (en) | 2005-11-08 | 2007-05-10 | Nortel Networks Limited | Interactive communication session cookies |
US20070130112A1 (en) | 2005-06-30 | 2007-06-07 | Intelligentek Corp. | Multimedia conceptual search system and associated search method |
US20070134631A1 (en) | 2005-12-13 | 2007-06-14 | Posit Science Corporation | Progressions in HiFi assessments |
US20070156677A1 (en) | 1999-07-21 | 2007-07-05 | Alberti Anemometer Llc | Database access system |
US20070185702A1 (en) | 2006-02-09 | 2007-08-09 | John Harney | Language independent parsing in natural language systems |
US20070197296A1 (en) | 2004-08-27 | 2007-08-23 | Nhn Corporation | Method and system for providing character having game item functions |
US7263493B1 (en) | 2002-01-11 | 2007-08-28 | P5, Inc. | Delivering electronic versions of supporting documents associated with an insurance claim |
US20070265533A1 (en) | 2006-05-12 | 2007-11-15 | Bao Tran | Cuffless blood pressure monitoring appliance |
US20070294229A1 (en) | 1998-05-28 | 2007-12-20 | Q-Phrase Llc | Chat conversation methods traversing a provisional scaffold of meanings |
US20080010268A1 (en) | 2006-07-06 | 2008-01-10 | Oracle International Corporation | Document ranking with sub-query series |
US20080016040A1 (en) | 2006-07-14 | 2008-01-17 | Chacha Search Inc. | Method and system for qualifying keywords in query strings |
US20080036756A1 (en) | 2006-08-10 | 2008-02-14 | Maria Gaos | System and methods for content conversion and distribution |
US20080091406A1 (en) | 2006-10-16 | 2008-04-17 | Voicebox Technologies, Inc. | System and method for a cooperative conversational voice user interface |
US20080096533A1 (en) | 2006-10-24 | 2008-04-24 | Kallideas Spa | Virtual Assistant With Real-Time Emotions |
US20080133444A1 (en) | 2006-12-05 | 2008-06-05 | Microsoft Corporation | Web-based collocation error proofing |
US20080162498A1 (en) | 2001-06-22 | 2008-07-03 | Nosa Omoigui | System and method for knowledge retrieval, management, delivery and presentation |
US20080222734A1 (en) | 2000-11-13 | 2008-09-11 | Redlich Ron M | Security System with Extraction, Reconstruction and Secure Recovery and Storage of Data |
US7426697B2 (en) | 2005-01-18 | 2008-09-16 | Microsoft Corporation | Multi-application tabbing system |
US20080235604A1 (en) | 2007-03-23 | 2008-09-25 | Peter Ebert | Model-based customer engagement techniques |
US20080305815A1 (en) | 2007-05-18 | 2008-12-11 | Smarttouch, Inc. | System and method for enhanced communications via small data rate communication systems |
US20090006525A1 (en) | 2007-06-26 | 2009-01-01 | Darryl Cynthia Moore | Methods, systems, and products for producing persona-based hosts |
US7483829B2 (en) | 2001-07-26 | 2009-01-27 | International Business Machines Corporation | Candidate synonym support device for generating candidate synonyms that can handle abbreviations, mispellings, and the like |
US20090030800A1 (en) | 2006-02-01 | 2009-01-29 | Dan Grois | Method and System for Searching a Data Network by Using a Virtual Assistant and for Advertising by using the same |
US20090063427A1 (en) | 2007-09-03 | 2009-03-05 | Marc Zuta | Communications System and Method |
US20090070103A1 (en) | 2007-09-07 | 2009-03-12 | Enhanced Medical Decisions, Inc. | Management and Processing of Information |
US20090077488A1 (en) | 2007-01-07 | 2009-03-19 | Bas Ording | Device, Method, and Graphical User Interface for Electronic Document Translation on a Touch-Screen Display |
US20090089100A1 (en) | 2007-10-01 | 2009-04-02 | Valeriy Nenov | Clinical information system |
US20090119587A1 (en) | 2007-11-02 | 2009-05-07 | Allen James F | Interactive complex task teaching system |
US20090119095A1 (en) | 2007-11-05 | 2009-05-07 | Enhanced Medical Decisions. Inc. | Machine Learning Systems and Methods for Improved Natural Language Processing |
US7536413B1 (en) | 2001-05-07 | 2009-05-19 | Ixreveal, Inc. | Concept-based categorization of unstructured objects |
US7548899B1 (en) | 2000-12-08 | 2009-06-16 | Del Favero Jr John P | Method and system for information retrieval based on menu selections |
US20090157386A1 (en) | 2007-08-11 | 2009-06-18 | Microsoft Corporation | Diagnostic evaluation of machine translators |
US20090171923A1 (en) | 2008-01-02 | 2009-07-02 | Michael Patrick Nash | Domain-specific concept model for associating structured data that enables a natural language query |
US7558792B2 (en) | 2004-06-29 | 2009-07-07 | Palo Alto Research Center Incorporated | Automatic extraction of human-readable lists from structured documents |
US20090182702A1 (en) | 2008-01-15 | 2009-07-16 | Miller Tanya M | Active Lab |
US20090204677A1 (en) | 2008-02-11 | 2009-08-13 | Avaya Technology Llc | Context based filter method and apparatus |
US20090216691A1 (en) | 2008-02-27 | 2009-08-27 | Inteliwise Sp Z.O.O. | Systems and Methods for Generating and Implementing an Interactive Man-Machine Web Interface Based on Natural Language Processing and Avatar Virtual Agent Based Character |
US20090225041A1 (en) | 2008-03-04 | 2009-09-10 | Apple Inc. | Language input interface on a device |
US20090227223A1 (en) | 2008-03-05 | 2009-09-10 | Jenkins Nevin C | Versatile personal medical emergency communication system |
US20090228264A1 (en) | 2003-02-11 | 2009-09-10 | Microsoft Corporation | Management of conversations |
US20090235356A1 (en) | 2008-03-14 | 2009-09-17 | Clear Blue Security, Llc | Multi virtual expert system and method for network management |
US20090248399A1 (en) | 2008-03-21 | 2009-10-01 | Lawrence Au | System and method for analyzing text using emotional intelligence factors |
US7599831B2 (en) | 2003-03-14 | 2009-10-06 | Sonum Technologies, Inc. | Multi-stage pattern reduction for natural language processing |
US7610382B1 (en) | 2006-06-30 | 2009-10-27 | Amazon Technologies, Inc. | System and method for marking content |
US20090271205A1 (en) | 2008-04-24 | 2009-10-29 | Finn Peter G | Preferred customer service representative presentation to virtual universe clients |
US20100005122A1 (en) | 2007-01-30 | 2010-01-07 | International Business Machines Corporation | Dynamic information systems |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US20100050237A1 (en) | 2008-08-19 | 2010-02-25 | Brian Ronald Bokor | Generating user and avatar specific content in a virtual world |
US20100070871A1 (en) | 2008-09-12 | 2010-03-18 | International Business Machines Corporation | Extendable Recommender Framework for Web-Based Systems |
US20100070448A1 (en) | 2002-06-24 | 2010-03-18 | Nosa Omoigui | System and method for knowledge retrieval, management, delivery and presentation |
US7711547B2 (en) | 2001-03-16 | 2010-05-04 | Meaningful Machines, L.L.C. | Word association method and apparatus |
US7739604B1 (en) | 2002-09-25 | 2010-06-15 | Apple Inc. | Method and apparatus for managing windows |
US20100153398A1 (en) | 2008-12-12 | 2010-06-17 | Next It Corporation | Leveraging concepts with information retrieval techniques and knowledge bases |
US20100169336A1 (en) | 2008-12-30 | 2010-07-01 | Target Brands Incorporated | Customer Search Utility |
US20100226490A1 (en) | 2009-03-05 | 2010-09-09 | Verizon Patent And Licensing, Inc. | Virtual call center manager |
US7797146B2 (en) | 2003-05-13 | 2010-09-14 | Interactive Drama, Inc. | Method and system for simulated interactive conversation |
US20100235808A1 (en) | 2009-03-12 | 2010-09-16 | Deutsche Telekom Ag | Method and system for task modeling of mobile phone applications |
US7818183B2 (en) | 2007-10-22 | 2010-10-19 | American Well Corporation | Connecting consumers with service providers |
US20100281012A1 (en) | 2009-04-29 | 2010-11-04 | Microsoft Corporation | Automatic recommendation of vertical search engines |
US20100312547A1 (en) | 2009-06-05 | 2010-12-09 | Apple Inc. | Contextual voice commands |
US20110004841A1 (en) | 2007-10-19 | 2011-01-06 | Tyler Gildred | Conversational interface having visual representations for interactive data gathering |
US7912701B1 (en) | 2005-05-04 | 2011-03-22 | IgniteIP Capital IA Special Management LLC | Method and apparatus for semiotic correlation |
US20110071819A1 (en) | 2009-09-22 | 2011-03-24 | Tanya Miller | Apparatus, system, and method for natural language processing |
US20110078105A1 (en) | 2009-09-29 | 2011-03-31 | Pandorabots | Method for personalizing chat bots |
US20110119196A1 (en) | 2009-11-18 | 2011-05-19 | Nohold, Inc. | Confederated knowledge for two or more internet-accessible knowledge management systems |
US7970663B2 (en) | 2007-05-02 | 2011-06-28 | Ganz | Method of calculating an estimated market value of a character |
WO2011088053A2 (en) | 2010-01-18 | 2011-07-21 | Apple Inc. | Intelligent automated assistant |
US20110179126A1 (en) | 2010-01-20 | 2011-07-21 | Aol Inc. | Systems And Methods For Electronic Distribution Of Messages Over Communication Channels |
US20110213642A1 (en) | 2008-05-21 | 2011-09-01 | The Delfin Project, Inc. | Management system for a conversational system |
US20110288947A1 (en) | 2010-05-20 | 2011-11-24 | Biran Dov | Method, system, and apparatus for social network based support platform using content packages and expert interaction |
US20110301982A1 (en) | 2002-04-19 | 2011-12-08 | Green Jr W T | Integrated medical software system with clinical decision support |
US20110307245A1 (en) | 2010-06-14 | 2011-12-15 | Xerox Corporation | Word alignment method and system for improved vocabulary coverage in statistical machine translation |
US20120022872A1 (en) | 2010-01-18 | 2012-01-26 | Apple Inc. | Automatically Adapting User Interfaces For Hands-Free Interaction |
US20120030553A1 (en) | 2008-06-13 | 2012-02-02 | Scrible, Inc. | Methods and systems for annotating web pages and managing annotations and annotated web pages |
US20120041903A1 (en) | 2009-01-08 | 2012-02-16 | Liesl Jane Beilby | Chatbots |
US20120078891A1 (en) | 2010-09-28 | 2012-03-29 | International Business Machines Corporation | Providing answers to questions using multiple models to score candidate answers |
US8160979B1 (en) * | 2006-12-20 | 2012-04-17 | Cisco Technology, Inc. | Method and apparatus for providing a virtual service agent that receives queries, compares questions to a set of queries, and allows a user to confirm a closest match |
US20120110473A1 (en) | 2010-11-02 | 2012-05-03 | Erick Tseng | Avatar-Based Communications Launching System |
US20120117005A1 (en) | 2010-10-11 | 2012-05-10 | Spivack Nova T | System and method for providing distributed intelligent assistance |
US20120221502A1 (en) | 2010-01-25 | 2012-08-30 | Andrew Peter Nelson Jerram | Apparatuses, methods and systems for a digital conversation management platform |
US20120245926A1 (en) | 2011-02-18 | 2012-09-27 | Nuance Communications, Inc. | Methods and apparatus for formatting text for clinical fact extraction |
US20120253825A1 (en) | 2006-03-03 | 2012-10-04 | At&T Intellectual Property Ii, L.P. | Relevancy recognition for contextual question answering |
US20120265528A1 (en) | 2009-06-05 | 2012-10-18 | Apple Inc. | Using Context Information To Facilitate Processing Of Commands In A Virtual Assistant |
US20120284040A1 (en) | 2011-04-26 | 2012-11-08 | Kenneth Joe Dupin | Medical Modules and Monitoring Systems |
US20120311541A1 (en) | 2011-05-31 | 2012-12-06 | International Business Machines Corporation | Interactive semi-automatic test case maintenance |
US8346563B1 (en) | 2012-04-10 | 2013-01-01 | Artificial Solutions Ltd. | System and methods for delivering advanced natural language interaction applications |
US8352266B2 (en) | 2004-10-05 | 2013-01-08 | Inago Corporation | System and methods for improving accuracy of speech recognition utilizing concept to keyword mapping |
US20130017523A1 (en) | 2011-07-15 | 2013-01-17 | International Business Machines Corporation | Utilizing failures in question and answer system responses to enhance the accuracy of question and answer systems |
US20130031476A1 (en) | 2011-07-25 | 2013-01-31 | Coin Emmett | Voice activated virtual assistant |
US20130046149A1 (en) | 2011-08-19 | 2013-02-21 | Accenture Global Services Limited | Interactive virtual care |
US8401842B1 (en) | 2008-03-11 | 2013-03-19 | Emc Corporation | Phrase matching for document classification |
CN103051669A (en) | 2011-10-17 | 2013-04-17 | 捷讯研究有限公司 | Methods and devices for creating a communications log and visualisations of communications across multiple services |
US8424840B2 (en) | 2007-11-08 | 2013-04-23 | Robert Bosch Gmbh | Valve cartridge for a solenoid valve, and associated solenoid valve |
US8433556B2 (en) | 2006-11-02 | 2013-04-30 | University Of Southern California | Semi-supervised training for statistical word alignment |
US20130152092A1 (en) | 2011-12-08 | 2013-06-13 | Osher Yadgar | Generic virtual personal assistant platform |
US8473420B2 (en) | 2009-06-26 | 2013-06-25 | Microsoft Corporation | Computational models for supporting situated interactions in multi-user scenarios |
US20130204813A1 (en) | 2012-01-20 | 2013-08-08 | Fluential, Llc | Self-learning, context aware virtual assistants, systems and methods |
US8510276B2 (en) | 2010-09-29 | 2013-08-13 | Microsoft Corporation | Comparing and selecting data cleansing service providers |
US8519963B2 (en) | 2007-01-07 | 2013-08-27 | Apple Inc. | Portable multifunction device, method, and graphical user interface for interpreting a finger gesture on a touch screen display |
US20130254139A1 (en) | 2012-03-21 | 2013-09-26 | Xiaoguang Lei | Systems and methods for building a universal intelligent assistant with learning capabilities |
US20130258040A1 (en) | 2012-04-02 | 2013-10-03 | Argela Yazilim ve Bilisim Teknolojileri San. ve Tic. A.S. | Interactive Avatars for Telecommunication Systems |
US20130262467A1 (en) | 2010-12-23 | 2013-10-03 | Nokia Corporation | Method and apparatus for providing token-based classification of device information |
US20130275875A1 (en) | 2010-01-18 | 2013-10-17 | Apple Inc. | Automatically Adapting User Interfaces for Hands-Free Interaction |
US20130283168A1 (en) | 2012-04-18 | 2013-10-24 | Next It Corporation | Conversation User Interface |
US20140029734A1 (en) | 2012-07-27 | 2014-01-30 | Lg Electronics Inc. | Electronic device and method of controlling the same |
US20140040748A1 (en) | 2011-09-30 | 2014-02-06 | Apple Inc. | Interface for a Virtual Digital Assistant |
US20140047001A1 (en) | 2012-08-10 | 2014-02-13 | Nuance Communications, Inc. | Virtual agent communication for electronic device |
US20140053102A1 (en) | 2012-08-20 | 2014-02-20 | Pantech Co., Ltd. | Terminal and method for providing user interface |
US20140074454A1 (en) | 2012-09-07 | 2014-03-13 | Next It Corporation | Conversational Virtual Healthcare Assistant |
US20140098948A1 (en) | 2009-12-22 | 2014-04-10 | Cyara Solutions Pty Ltd | System and method for automated chat testing |
US20140115456A1 (en) | 2012-09-28 | 2014-04-24 | Oracle International Corporation | System for accessing software functionality |
US8731929B2 (en) | 2002-06-03 | 2014-05-20 | Voicebox Technologies Corporation | Agent architecture for determining meanings of natural language utterances |
US20140164508A1 (en) | 2012-12-11 | 2014-06-12 | Nuance Communications, Inc. | Systems and methods for sharing information between virtual agents |
US20140164476A1 (en) | 2012-12-06 | 2014-06-12 | At&T Intellectual Property I, Lp | Apparatus and method for providing a virtual assistant |
US8756326B1 (en) | 2005-11-08 | 2014-06-17 | Rockstar Consortium Us Lp | Using interactive communication session cookies in web sessions |
US8762152B2 (en) | 1999-11-12 | 2014-06-24 | Nuance Communications, Inc. | Speech recognition system interactive agent |
US20140181741A1 (en) | 2012-12-24 | 2014-06-26 | Microsoft Corporation | Discreetly displaying contextually relevant information |
US20140201675A1 (en) | 2013-01-11 | 2014-07-17 | Samsung Electronics Co., Ltd. | Method and mobile device for providing recommended items based on context awareness |
US8819003B2 (en) | 2008-09-30 | 2014-08-26 | Yahoo! Inc. | Query refinement based on user selections |
US20140244712A1 (en) | 2013-02-25 | 2014-08-28 | Artificial Solutions Iberia SL | System and methods for virtual assistant networks |
US20140244266A1 (en) | 2013-02-22 | 2014-08-28 | Next It Corporation | Interaction with a Portion of a Content Item through a Virtual Assistant |
US20140245140A1 (en) | 2013-02-22 | 2014-08-28 | Next It Corporation | Virtual Assistant Transfer between Smart Devices |
US20140282109A1 (en) | 2013-03-15 | 2014-09-18 | GroupSystems Corporation d/b/a ThinkTank by GroupS | Context frame for sharing context information |
US20140317502A1 (en) | 2013-04-18 | 2014-10-23 | Next It Corporation | Virtual assistant focused user interfaces |
US20150066817A1 (en) | 2013-08-27 | 2015-03-05 | Persais, Llc | System and method for virtual assistants with shared capabilities |
US20150185996A1 (en) | 2013-12-31 | 2015-07-02 | Next It Corporation | Virtual assistant team identification |
US9202171B2 (en) | 2008-11-11 | 2015-12-01 | Digideal Corporation | Virtual game assistant based on artificial intelligence |
US20160012186A1 (en) | 2013-03-01 | 2016-01-14 | 3M Innovative Properties Company | Systems and methods for requesting medical information |
US20160071517A1 (en) | 2014-09-09 | 2016-03-10 | Next It Corporation | Evaluating Conversation Data based on Risk Factors |
US9424840B1 (en) * | 2012-08-31 | 2016-08-23 | Amazon Technologies, Inc. | Speech recognition platforms |
US20170277993A1 (en) | 2016-03-22 | 2017-09-28 | Next It Corporation | Virtual assistant escalation |
-
2015
- 2015-09-09 US US14/849,541 patent/US20160071517A1/en not_active Abandoned
-
2017
- 2017-12-28 US US15/857,160 patent/US10175865B2/en active Active
-
2019
- 2019-01-08 US US16/242,639 patent/US10545648B2/en active Active
Patent Citations (221)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5535120A (en) | 1990-12-31 | 1996-07-09 | Trans-Link International Corp. | Machine translation and telecommunications system using user ID data to select dictionaries |
US5278980A (en) | 1991-08-16 | 1994-01-11 | Xerox Corporation | Iterative technique for phrase query formation and an information retrieval system employing same |
US5418948A (en) | 1991-10-08 | 1995-05-23 | West Publishing Company | Concept matching of natural language queries with a database of document concepts |
US5727174A (en) | 1992-03-23 | 1998-03-10 | International Business Machines Corporation | Graphical end-user interface for intelligent assistants |
US5677835A (en) | 1992-09-04 | 1997-10-14 | Caterpillar Inc. | Integrated authoring and translation system |
US6658627B1 (en) | 1992-09-04 | 2003-12-02 | Caterpillar Inc | Integrated and authoring and translation system |
US5615112A (en) | 1993-01-29 | 1997-03-25 | Arizona Board Of Regents | Synthesized object-oriented entity-relationship (SOOER) model for coupled knowledge-base/database of image retrieval expert system (IRES) |
US6388665B1 (en) | 1994-07-08 | 2002-05-14 | Microsoft Corporation | Software platform having a real world interface with animated characters |
US5682539A (en) | 1994-09-29 | 1997-10-28 | Conrad; Donovan | Anticipated meaning natural language interface |
US20040107088A1 (en) | 1994-09-30 | 2004-06-03 | Budzinski Robert L. | Memory system for storing and retrieving experience and knowledge with natural language utilizing state representation data, word sense numbers, function codes, directed graphs and/or context memory |
US20010000356A1 (en) | 1995-07-07 | 2001-04-19 | Woods William A. | Method and apparatus for generating query responses in a computer-based document retrieval system |
US6012053A (en) | 1997-06-23 | 2000-01-04 | Lycos, Inc. | Computer system with user-controlled relevance ranking of search results |
US6112177A (en) | 1997-11-07 | 2000-08-29 | At&T Corp. | Coarticulation method for audio-visual text-to-speech synthesis |
US6396951B1 (en) | 1997-12-29 | 2002-05-28 | Xerox Corporation | Document-based query data for information retrieval |
US6175829B1 (en) | 1998-04-22 | 2001-01-16 | Nec Usa, Inc. | Method and apparatus for facilitating query reformulation |
US6144938A (en) | 1998-05-01 | 2000-11-07 | Sun Microsystems, Inc. | Voice user interface with personality |
US20070294229A1 (en) | 1998-05-28 | 2007-12-20 | Q-Phrase Llc | Chat conversation methods traversing a provisional scaffold of meanings |
US6353817B1 (en) | 1998-06-26 | 2002-03-05 | Charles M Jacobs | Multi-user system for creating and maintaining a medical-decision-making knowledge base |
US6285978B1 (en) | 1998-09-24 | 2001-09-04 | International Business Machines Corporation | System and method for estimating accuracy of an automatic natural language translation |
US20050120276A1 (en) | 1999-01-06 | 2005-06-02 | Parasoft Corporation | Modularizing a computer program for testing and debugging |
US6282507B1 (en) | 1999-01-29 | 2001-08-28 | Sony Corporation | Method and apparatus for interactive source language expression recognition and alternative hypothesis presentation and selection |
US6401061B1 (en) | 1999-05-13 | 2002-06-04 | Yuri L. Zieman | Combinatorial computational technique for transformation phrase text-phrase meaning |
US20070156677A1 (en) | 1999-07-21 | 2007-07-05 | Alberti Anemometer Llc | Database access system |
US8762152B2 (en) | 1999-11-12 | 2014-06-24 | Nuance Communications, Inc. | Speech recognition system interactive agent |
US6826540B1 (en) | 1999-12-29 | 2004-11-30 | Virtual Personalities, Inc. | Virtual human interface for conducting surveys |
US20010053968A1 (en) | 2000-01-10 | 2001-12-20 | Iaskweb, Inc. | System, method, and computer program product for responding to natural language queries |
US6829603B1 (en) | 2000-02-02 | 2004-12-07 | International Business Machines Corp. | System, method and program product for interactive natural dialog |
US20010033298A1 (en) | 2000-03-01 | 2001-10-25 | Benjamin Slotznick | Adjunct use of instant messenger software to enable communications to or between chatterbots or other software agents |
US7337158B2 (en) | 2000-03-06 | 2008-02-26 | Consona Crm Inc. | System and method for providing an intelligent multi-step dialog with a user |
US20010049688A1 (en) | 2000-03-06 | 2001-12-06 | Raya Fratkina | System and method for providing an intelligent multi-step dialog with a user |
US7539656B2 (en) | 2000-03-06 | 2009-05-26 | Consona Crm Inc. | System and method for providing an intelligent multi-step dialog with a user |
US6757362B1 (en) | 2000-03-06 | 2004-06-29 | Avaya Technology Corp. | Personal virtual assistant |
US20010044751A1 (en) | 2000-04-03 | 2001-11-22 | Pugliese Anthony V. | System and method for displaying and selling goods and services |
US20020123994A1 (en) | 2000-04-26 | 2002-09-05 | Yves Schabes | System for fulfilling an information need using extended matching techniques |
US20020008716A1 (en) | 2000-07-21 | 2002-01-24 | Colburn Robert A. | System and method for controlling expression characteristics of a virtual agent |
US20020032591A1 (en) | 2000-09-08 | 2002-03-14 | Agentai, Inc. | Service request processing performed by artificial intelligence systems in conjunctiion with human intervention |
US6999932B1 (en) | 2000-10-10 | 2006-02-14 | Intel Corporation | Language independent voice-based search system |
US6987514B1 (en) | 2000-11-09 | 2006-01-17 | Nokia Corporation | Voice avatars for wireless multiuser entertainment services |
US20080222734A1 (en) | 2000-11-13 | 2008-09-11 | Redlich Ron M | Security System with Extraction, Reconstruction and Secure Recovery and Storage of Data |
US6834120B1 (en) | 2000-11-15 | 2004-12-21 | Sri International | Method and system for estimating the accuracy of inference algorithms using the self-consistency methodology |
US7548899B1 (en) | 2000-12-08 | 2009-06-16 | Del Favero Jr John P | Method and system for information retrieval based on menu selections |
US20020129031A1 (en) | 2001-01-05 | 2002-09-12 | Lau Lee Min | Managing relationships between unique concepts in a database |
US6661418B1 (en) | 2001-01-22 | 2003-12-09 | Digital Animations Limited | Character animation system |
US7711547B2 (en) | 2001-03-16 | 2010-05-04 | Meaningful Machines, L.L.C. | Word association method and apparatus |
US20020198885A1 (en) | 2001-04-09 | 2002-12-26 | Streepy Larry V. | Method and system for interfacing with a multi-level data structure |
US7194483B1 (en) | 2001-05-07 | 2007-03-20 | Intelligenxia, Inc. | Method, system, and computer program product for concept-based multi-dimensional analysis of unstructured information |
US7536413B1 (en) | 2001-05-07 | 2009-05-19 | Ixreveal, Inc. | Concept-based categorization of unstructured objects |
US20080162498A1 (en) | 2001-06-22 | 2008-07-03 | Nosa Omoigui | System and method for knowledge retrieval, management, delivery and presentation |
US20030004908A1 (en) | 2001-06-29 | 2003-01-02 | Linthicum Steven Eric | Method and system for automated maintenance and training instruction generation and validation |
US7483829B2 (en) | 2001-07-26 | 2009-01-27 | International Business Machines Corporation | Candidate synonym support device for generating candidate synonyms that can handle abbreviations, mispellings, and the like |
US20030041307A1 (en) | 2001-08-24 | 2003-02-27 | Jong-Won Park | Electronic dictionary system with a tree structure and its output |
US20030061029A1 (en) | 2001-08-29 | 2003-03-27 | Efraim Shaket | Device for conducting expectation based mixed initiative natural language dialogs |
US20030088547A1 (en) | 2001-11-06 | 2003-05-08 | Hammond Joel K. | Method and apparatus for providing comprehensive search results in response to user queries entered over a computer network |
US20030142829A1 (en) | 2001-11-26 | 2003-07-31 | Cristiano Avigni | Systems and methods for determining sound of a moving object |
US20030126090A1 (en) | 2001-12-28 | 2003-07-03 | Fujitsu Limited | Conversation method, device, program and computer-readable recording medium on which conversation program is recorded |
US20030126089A1 (en) | 2001-12-28 | 2003-07-03 | Fujitsu Limited | Conversation method, device, program and computer-readable recording medium on which conversation program is recorded |
US7263493B1 (en) | 2002-01-11 | 2007-08-28 | P5, Inc. | Delivering electronic versions of supporting documents associated with an insurance claim |
US20110301982A1 (en) | 2002-04-19 | 2011-12-08 | Green Jr W T | Integrated medical software system with clinical decision support |
US20030212544A1 (en) | 2002-05-10 | 2003-11-13 | Alejandro Acero | System for automatically annotating training data for a natural language understanding system |
US7076430B1 (en) | 2002-05-16 | 2006-07-11 | At&T Corp. | System and method of providing conversational visual prosody for talking heads |
US20060074689A1 (en) | 2002-05-16 | 2006-04-06 | At&T Corp. | System and method of providing conversational visual prosody for talking heads |
US8731929B2 (en) | 2002-06-03 | 2014-05-20 | Voicebox Technologies Corporation | Agent architecture for determining meanings of natural language utterances |
US20100070448A1 (en) | 2002-06-24 | 2010-03-18 | Nosa Omoigui | System and method for knowledge retrieval, management, delivery and presentation |
US7058902B2 (en) | 2002-07-30 | 2006-06-06 | Microsoft Corporation | Enhanced on-object context menus |
US7739604B1 (en) | 2002-09-25 | 2010-06-15 | Apple Inc. | Method and apparatus for managing windows |
US20040141013A1 (en) | 2003-01-21 | 2004-07-22 | Microsoft Corporation | System and method for directly accessing functionality provided by an application |
US20060080107A1 (en) | 2003-02-11 | 2006-04-13 | Unveil Technologies, Inc., A Delaware Corporation | Management of conversations |
US20090228264A1 (en) | 2003-02-11 | 2009-09-10 | Microsoft Corporation | Management of conversations |
US7599831B2 (en) | 2003-03-14 | 2009-10-06 | Sonum Technologies, Inc. | Multi-stage pattern reduction for natural language processing |
US20040186705A1 (en) | 2003-03-18 | 2004-09-23 | Morgan Alexander P. | Concept word management |
US7797146B2 (en) | 2003-05-13 | 2010-09-14 | Interactive Drama, Inc. | Method and system for simulated interactive conversation |
US20050027694A1 (en) | 2003-07-31 | 2005-02-03 | Volker Sauermann | User-friendly search results display system, method, and computer program product |
US20050054381A1 (en) | 2003-09-05 | 2005-03-10 | Samsung Electronics Co., Ltd. | Proactive user interface |
US20060004826A1 (en) | 2004-05-04 | 2006-01-05 | Mark Zartler | Data disambiguation systems and methods |
US20060037076A1 (en) | 2004-05-04 | 2006-02-16 | Shantu Roy | Methods and systems for enforcing network and computer use policy |
US20080005158A1 (en) | 2004-05-04 | 2008-01-03 | Mark Zartler | Data Disambiguation Systems and Methods |
US7558792B2 (en) | 2004-06-29 | 2009-07-07 | Palo Alto Research Center Incorporated | Automatic extraction of human-readable lists from structured documents |
US20060020466A1 (en) | 2004-07-26 | 2006-01-26 | Cousineau Leo E | Ontology based medical patient evaluation method for data capture and knowledge representation |
US20060036430A1 (en) | 2004-08-12 | 2006-02-16 | Junling Hu | System and method for domain-based natural language consultation |
US20060047632A1 (en) | 2004-08-12 | 2006-03-02 | Guoming Zhang | Method using ontology and user query processing to solve inventor problems and user problems |
US20070197296A1 (en) | 2004-08-27 | 2007-08-23 | Nhn Corporation | Method and system for providing character having game item functions |
US20060074831A1 (en) | 2004-09-20 | 2006-04-06 | Hyder Andrew D | Virtual assistant |
US20060092978A1 (en) | 2004-09-30 | 2006-05-04 | Ajita John | Method and apparatus for developing a virtual assistant for a communication |
US20060067352A1 (en) | 2004-09-30 | 2006-03-30 | Ajita John | Method and apparatus for providing a virtual assistant to a communication participant |
US8352266B2 (en) | 2004-10-05 | 2013-01-08 | Inago Corporation | System and methods for improving accuracy of speech recognition utilizing concept to keyword mapping |
US20060206483A1 (en) | 2004-10-27 | 2006-09-14 | Harris Corporation | Method for domain identification of documents in a document database |
US20060161414A1 (en) | 2004-12-15 | 2006-07-20 | C.R.F. Societa Consortile Per Azioni | Event-driven model generated from an ordered natural language interface |
US7426697B2 (en) | 2005-01-18 | 2008-09-16 | Microsoft Corporation | Multi-application tabbing system |
US20060253427A1 (en) | 2005-05-04 | 2006-11-09 | Jun Wu | Suggesting and refining user input based on original user input |
US7912701B1 (en) | 2005-05-04 | 2011-03-22 | IgniteIP Capital IA Special Management LLC | Method and apparatus for semiotic correlation |
US20070130112A1 (en) | 2005-06-30 | 2007-06-07 | Intelligentek Corp. | Multimedia conceptual search system and associated search method |
US20070043687A1 (en) | 2005-08-19 | 2007-02-22 | Accenture Llp | Virtual assistant |
US9501741B2 (en) | 2005-09-08 | 2016-11-22 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US20070100790A1 (en) | 2005-09-08 | 2007-05-03 | Adam Cheyer | Method and apparatus for building an intelligent automated assistant |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US20070106670A1 (en) | 2005-11-08 | 2007-05-10 | Nortel Networks Limited | Interactive communication session cookies |
US8756326B1 (en) | 2005-11-08 | 2014-06-17 | Rockstar Consortium Us Lp | Using interactive communication session cookies in web sessions |
US20070134631A1 (en) | 2005-12-13 | 2007-06-14 | Posit Science Corporation | Progressions in HiFi assessments |
US20090030800A1 (en) | 2006-02-01 | 2009-01-29 | Dan Grois | Method and System for Searching a Data Network by Using a Virtual Assistant and for Advertising by using the same |
US20070185702A1 (en) | 2006-02-09 | 2007-08-09 | John Harney | Language independent parsing in natural language systems |
US20120253825A1 (en) | 2006-03-03 | 2012-10-04 | At&T Intellectual Property Ii, L.P. | Relevancy recognition for contextual question answering |
US20070265533A1 (en) | 2006-05-12 | 2007-11-15 | Bao Tran | Cuffless blood pressure monitoring appliance |
US7610382B1 (en) | 2006-06-30 | 2009-10-27 | Amazon Technologies, Inc. | System and method for marking content |
US20080010268A1 (en) | 2006-07-06 | 2008-01-10 | Oracle International Corporation | Document ranking with sub-query series |
US20080016040A1 (en) | 2006-07-14 | 2008-01-17 | Chacha Search Inc. | Method and system for qualifying keywords in query strings |
US20080036756A1 (en) | 2006-08-10 | 2008-02-14 | Maria Gaos | System and methods for content conversion and distribution |
US9117447B2 (en) | 2006-09-08 | 2015-08-25 | Apple Inc. | Using event alert text as input to an automated assistant |
US8942986B2 (en) | 2006-09-08 | 2015-01-27 | Apple Inc. | Determining user intent based on ontologies of domains |
US8930191B2 (en) | 2006-09-08 | 2015-01-06 | Apple Inc. | Paraphrasing of user requests and results by automated digital assistant |
US20080091406A1 (en) | 2006-10-16 | 2008-04-17 | Voicebox Technologies, Inc. | System and method for a cooperative conversational voice user interface |
US20080096533A1 (en) | 2006-10-24 | 2008-04-24 | Kallideas Spa | Virtual Assistant With Real-Time Emotions |
US8433556B2 (en) | 2006-11-02 | 2013-04-30 | University Of Southern California | Semi-supervised training for statistical word alignment |
US20080133444A1 (en) | 2006-12-05 | 2008-06-05 | Microsoft Corporation | Web-based collocation error proofing |
US8160979B1 (en) * | 2006-12-20 | 2012-04-17 | Cisco Technology, Inc. | Method and apparatus for providing a virtual service agent that receives queries, compares questions to a set of queries, and allows a user to confirm a closest match |
US20090077488A1 (en) | 2007-01-07 | 2009-03-19 | Bas Ording | Device, Method, and Graphical User Interface for Electronic Document Translation on a Touch-Screen Display |
US8519963B2 (en) | 2007-01-07 | 2013-08-27 | Apple Inc. | Portable multifunction device, method, and graphical user interface for interpreting a finger gesture on a touch screen display |
US20100005122A1 (en) | 2007-01-30 | 2010-01-07 | International Business Machines Corporation | Dynamic information systems |
US20080235604A1 (en) | 2007-03-23 | 2008-09-25 | Peter Ebert | Model-based customer engagement techniques |
US7970663B2 (en) | 2007-05-02 | 2011-06-28 | Ganz | Method of calculating an estimated market value of a character |
US20080305815A1 (en) | 2007-05-18 | 2008-12-11 | Smarttouch, Inc. | System and method for enhanced communications via small data rate communication systems |
US20090006525A1 (en) | 2007-06-26 | 2009-01-01 | Darryl Cynthia Moore | Methods, systems, and products for producing persona-based hosts |
US20090157386A1 (en) | 2007-08-11 | 2009-06-18 | Microsoft Corporation | Diagnostic evaluation of machine translators |
US20090063427A1 (en) | 2007-09-03 | 2009-03-05 | Marc Zuta | Communications System and Method |
US20090070103A1 (en) | 2007-09-07 | 2009-03-12 | Enhanced Medical Decisions, Inc. | Management and Processing of Information |
US20090089100A1 (en) | 2007-10-01 | 2009-04-02 | Valeriy Nenov | Clinical information system |
US20110004841A1 (en) | 2007-10-19 | 2011-01-06 | Tyler Gildred | Conversational interface having visual representations for interactive data gathering |
US7818183B2 (en) | 2007-10-22 | 2010-10-19 | American Well Corporation | Connecting consumers with service providers |
US20090119587A1 (en) | 2007-11-02 | 2009-05-07 | Allen James F | Interactive complex task teaching system |
US20090119095A1 (en) | 2007-11-05 | 2009-05-07 | Enhanced Medical Decisions. Inc. | Machine Learning Systems and Methods for Improved Natural Language Processing |
US8424840B2 (en) | 2007-11-08 | 2013-04-23 | Robert Bosch Gmbh | Valve cartridge for a solenoid valve, and associated solenoid valve |
US20090171923A1 (en) | 2008-01-02 | 2009-07-02 | Michael Patrick Nash | Domain-specific concept model for associating structured data that enables a natural language query |
US20140365223A1 (en) | 2008-01-15 | 2014-12-11 | Next It Corporation | Virtual Assistant Conversations |
US20140343924A1 (en) | 2008-01-15 | 2014-11-20 | Next It Corporation | Active Lab |
US20140365407A1 (en) | 2008-01-15 | 2014-12-11 | Next It Corporation | Context-Based Virtual Assistant Conversations |
US20090182702A1 (en) | 2008-01-15 | 2009-07-16 | Miller Tanya M | Active Lab |
US20090204677A1 (en) | 2008-02-11 | 2009-08-13 | Avaya Technology Llc | Context based filter method and apparatus |
US20090216691A1 (en) | 2008-02-27 | 2009-08-27 | Inteliwise Sp Z.O.O. | Systems and Methods for Generating and Implementing an Interactive Man-Machine Web Interface Based on Natural Language Processing and Avatar Virtual Agent Based Character |
US20090225041A1 (en) | 2008-03-04 | 2009-09-10 | Apple Inc. | Language input interface on a device |
US20090227223A1 (en) | 2008-03-05 | 2009-09-10 | Jenkins Nevin C | Versatile personal medical emergency communication system |
US8401842B1 (en) | 2008-03-11 | 2013-03-19 | Emc Corporation | Phrase matching for document classification |
US20090235356A1 (en) | 2008-03-14 | 2009-09-17 | Clear Blue Security, Llc | Multi virtual expert system and method for network management |
US20090248399A1 (en) | 2008-03-21 | 2009-10-01 | Lawrence Au | System and method for analyzing text using emotional intelligence factors |
US20090271205A1 (en) | 2008-04-24 | 2009-10-29 | Finn Peter G | Preferred customer service representative presentation to virtual universe clients |
US20110213642A1 (en) | 2008-05-21 | 2011-09-01 | The Delfin Project, Inc. | Management system for a conversational system |
US20120030553A1 (en) | 2008-06-13 | 2012-02-02 | Scrible, Inc. | Methods and systems for annotating web pages and managing annotations and annotated web pages |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US20100050237A1 (en) | 2008-08-19 | 2010-02-25 | Brian Ronald Bokor | Generating user and avatar specific content in a virtual world |
US20100070871A1 (en) | 2008-09-12 | 2010-03-18 | International Business Machines Corporation | Extendable Recommender Framework for Web-Based Systems |
US8819003B2 (en) | 2008-09-30 | 2014-08-26 | Yahoo! Inc. | Query refinement based on user selections |
US9202171B2 (en) | 2008-11-11 | 2015-12-01 | Digideal Corporation | Virtual game assistant based on artificial intelligence |
US20100153398A1 (en) | 2008-12-12 | 2010-06-17 | Next It Corporation | Leveraging concepts with information retrieval techniques and knowledge bases |
US20100169336A1 (en) | 2008-12-30 | 2010-07-01 | Target Brands Incorporated | Customer Search Utility |
US20120041903A1 (en) | 2009-01-08 | 2012-02-16 | Liesl Jane Beilby | Chatbots |
US20100226490A1 (en) | 2009-03-05 | 2010-09-09 | Verizon Patent And Licensing, Inc. | Virtual call center manager |
US20100235808A1 (en) | 2009-03-12 | 2010-09-16 | Deutsche Telekom Ag | Method and system for task modeling of mobile phone applications |
US20100281012A1 (en) | 2009-04-29 | 2010-11-04 | Microsoft Corporation | Automatic recommendation of vertical search engines |
US20120265528A1 (en) | 2009-06-05 | 2012-10-18 | Apple Inc. | Using Context Information To Facilitate Processing Of Commands In A Virtual Assistant |
US20140297284A1 (en) | 2009-06-05 | 2014-10-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US20100312547A1 (en) | 2009-06-05 | 2010-12-09 | Apple Inc. | Contextual voice commands |
US8473420B2 (en) | 2009-06-26 | 2013-06-25 | Microsoft Corporation | Computational models for supporting situated interactions in multi-user scenarios |
US20140310005A1 (en) | 2009-09-22 | 2014-10-16 | Next It Corporation | Virtual assistant conversations for ambiguous user input and goals |
US20110071819A1 (en) | 2009-09-22 | 2011-03-24 | Tanya Miller | Apparatus, system, and method for natural language processing |
US20140343928A1 (en) | 2009-09-22 | 2014-11-20 | Next It Corporation | Wearable-Based Virtual Agents |
US8943094B2 (en) | 2009-09-22 | 2015-01-27 | Next It Corporation | Apparatus, system, and method for natural language processing |
US20170132220A1 (en) | 2009-09-22 | 2017-05-11 | Next It Corporation | Apparatus, system, and method for natural language processing |
US20110078105A1 (en) | 2009-09-29 | 2011-03-31 | Pandorabots | Method for personalizing chat bots |
US20110119196A1 (en) | 2009-11-18 | 2011-05-19 | Nohold, Inc. | Confederated knowledge for two or more internet-accessible knowledge management systems |
US20140098948A1 (en) | 2009-12-22 | 2014-04-10 | Cyara Solutions Pty Ltd | System and method for automated chat testing |
WO2011088053A2 (en) | 2010-01-18 | 2011-07-21 | Apple Inc. | Intelligent automated assistant |
US20120022872A1 (en) | 2010-01-18 | 2012-01-26 | Apple Inc. | Automatically Adapting User Interfaces For Hands-Free Interaction |
US20120016678A1 (en) | 2010-01-18 | 2012-01-19 | Apple Inc. | Intelligent Automated Assistant |
US20130275875A1 (en) | 2010-01-18 | 2013-10-17 | Apple Inc. | Automatically Adapting User Interfaces for Hands-Free Interaction |
US8670979B2 (en) | 2010-01-18 | 2014-03-11 | Apple Inc. | Active input elicitation by intelligent automated assistant |
US20110179126A1 (en) | 2010-01-20 | 2011-07-21 | Aol Inc. | Systems And Methods For Electronic Distribution Of Messages Over Communication Channels |
US20120221502A1 (en) | 2010-01-25 | 2012-08-30 | Andrew Peter Nelson Jerram | Apparatuses, methods and systems for a digital conversation management platform |
US20110288947A1 (en) | 2010-05-20 | 2011-11-24 | Biran Dov | Method, system, and apparatus for social network based support platform using content packages and expert interaction |
US20110307245A1 (en) | 2010-06-14 | 2011-12-15 | Xerox Corporation | Word alignment method and system for improved vocabulary coverage in statistical machine translation |
US20120078891A1 (en) | 2010-09-28 | 2012-03-29 | International Business Machines Corporation | Providing answers to questions using multiple models to score candidate answers |
US8510276B2 (en) | 2010-09-29 | 2013-08-13 | Microsoft Corporation | Comparing and selecting data cleansing service providers |
US20120117005A1 (en) | 2010-10-11 | 2012-05-10 | Spivack Nova T | System and method for providing distributed intelligent assistance |
US20120110473A1 (en) | 2010-11-02 | 2012-05-03 | Erick Tseng | Avatar-Based Communications Launching System |
US20130262467A1 (en) | 2010-12-23 | 2013-10-03 | Nokia Corporation | Method and apparatus for providing token-based classification of device information |
US20120245926A1 (en) | 2011-02-18 | 2012-09-27 | Nuance Communications, Inc. | Methods and apparatus for formatting text for clinical fact extraction |
US20120284040A1 (en) | 2011-04-26 | 2012-11-08 | Kenneth Joe Dupin | Medical Modules and Monitoring Systems |
US20120311541A1 (en) | 2011-05-31 | 2012-12-06 | International Business Machines Corporation | Interactive semi-automatic test case maintenance |
US20130017523A1 (en) | 2011-07-15 | 2013-01-17 | International Business Machines Corporation | Utilizing failures in question and answer system responses to enhance the accuracy of question and answer systems |
US20130031476A1 (en) | 2011-07-25 | 2013-01-31 | Coin Emmett | Voice activated virtual assistant |
US20130046149A1 (en) | 2011-08-19 | 2013-02-21 | Accenture Global Services Limited | Interactive virtual care |
US20140040748A1 (en) | 2011-09-30 | 2014-02-06 | Apple Inc. | Interface for a Virtual Digital Assistant |
CN103051669A (en) | 2011-10-17 | 2013-04-17 | 捷讯研究有限公司 | Methods and devices for creating a communications log and visualisations of communications across multiple services |
US20130152092A1 (en) | 2011-12-08 | 2013-06-13 | Osher Yadgar | Generic virtual personal assistant platform |
US20130204813A1 (en) | 2012-01-20 | 2013-08-08 | Fluential, Llc | Self-learning, context aware virtual assistants, systems and methods |
US20130254139A1 (en) | 2012-03-21 | 2013-09-26 | Xiaoguang Lei | Systems and methods for building a universal intelligent assistant with learning capabilities |
US20130258040A1 (en) | 2012-04-02 | 2013-10-03 | Argela Yazilim ve Bilisim Teknolojileri San. ve Tic. A.S. | Interactive Avatars for Telecommunication Systems |
US8346563B1 (en) | 2012-04-10 | 2013-01-01 | Artificial Solutions Ltd. | System and methods for delivering advanced natural language interaction applications |
US20160110071A1 (en) | 2012-04-18 | 2016-04-21 | Next It Corporation | Conversation User Interface |
US20130283168A1 (en) | 2012-04-18 | 2013-10-24 | Next It Corporation | Conversation User Interface |
US20140029734A1 (en) | 2012-07-27 | 2014-01-30 | Lg Electronics Inc. | Electronic device and method of controlling the same |
US20140047001A1 (en) | 2012-08-10 | 2014-02-13 | Nuance Communications, Inc. | Virtual agent communication for electronic device |
US20140053102A1 (en) | 2012-08-20 | 2014-02-20 | Pantech Co., Ltd. | Terminal and method for providing user interface |
US9424840B1 (en) * | 2012-08-31 | 2016-08-23 | Amazon Technologies, Inc. | Speech recognition platforms |
US20140337048A1 (en) | 2012-09-07 | 2014-11-13 | Next It Corporation | Conversational Virtual Healthcare Assistant |
US20140074454A1 (en) | 2012-09-07 | 2014-03-13 | Next It Corporation | Conversational Virtual Healthcare Assistant |
US20140115456A1 (en) | 2012-09-28 | 2014-04-24 | Oracle International Corporation | System for accessing software functionality |
US20140164476A1 (en) | 2012-12-06 | 2014-06-12 | At&T Intellectual Property I, Lp | Apparatus and method for providing a virtual assistant |
US20140164508A1 (en) | 2012-12-11 | 2014-06-12 | Nuance Communications, Inc. | Systems and methods for sharing information between virtual agents |
US20140181741A1 (en) | 2012-12-24 | 2014-06-26 | Microsoft Corporation | Discreetly displaying contextually relevant information |
US20140201675A1 (en) | 2013-01-11 | 2014-07-17 | Samsung Electronics Co., Ltd. | Method and mobile device for providing recommended items based on context awareness |
US20140245140A1 (en) | 2013-02-22 | 2014-08-28 | Next It Corporation | Virtual Assistant Transfer between Smart Devices |
US20140244266A1 (en) | 2013-02-22 | 2014-08-28 | Next It Corporation | Interaction with a Portion of a Content Item through a Virtual Assistant |
US20140244712A1 (en) | 2013-02-25 | 2014-08-28 | Artificial Solutions Iberia SL | System and methods for virtual assistant networks |
US20160012186A1 (en) | 2013-03-01 | 2016-01-14 | 3M Innovative Properties Company | Systems and methods for requesting medical information |
US20140282109A1 (en) | 2013-03-15 | 2014-09-18 | GroupSystems Corporation d/b/a ThinkTank by GroupS | Context frame for sharing context information |
US20140317502A1 (en) | 2013-04-18 | 2014-10-23 | Next It Corporation | Virtual assistant focused user interfaces |
US20150066817A1 (en) | 2013-08-27 | 2015-03-05 | Persais, Llc | System and method for virtual assistants with shared capabilities |
US20150186156A1 (en) | 2013-12-31 | 2015-07-02 | Next It Corporation | Virtual assistant conversations |
US20150186154A1 (en) | 2013-12-31 | 2015-07-02 | Next It Corporation | Virtual assistant team customization |
US20150186155A1 (en) | 2013-12-31 | 2015-07-02 | Next It Corporation | Virtual assistant acquisitions and training |
US20150185996A1 (en) | 2013-12-31 | 2015-07-02 | Next It Corporation | Virtual assistant team identification |
US20160071517A1 (en) | 2014-09-09 | 2016-03-10 | Next It Corporation | Evaluating Conversation Data based on Risk Factors |
US20170277993A1 (en) | 2016-03-22 | 2017-09-28 | Next It Corporation | Virtual assistant escalation |
Non-Patent Citations (14)
Title |
---|
"5 wearable electronic phones", retrieved on Feb. 13, 2015 at <<http://limcorp.net/2009/5-wearable-electronic-phones>>, 2009, 12 pages. |
"Frost & Sullivan Commends Next IT for Leading the Virtual Agent Applications Industry in Competitive Strategy Innovation," 2014, 5 pages. |
"Meet Jenn, Your Virtual Assistant at alaskaair.com," retrieved on Apr. 13, 2015 at <<http://www.alaskaair.com/content/about-us/site-info/ask-jenn.aspx>>, 1 page. |
"Meet Julia-TAM Airlines′ most famous new hire," Case Study, 2015, 2 pages. |
"SGT STAR Wins Intelligent Assistant Award," PRWEB Online Visibility from Vocus, Sep. 24, 2014, 2 pages. |
"The Army's Robot Recruiter," Transcript from New York Public Radio, retrieved on Jan. 20, 2015 at <<http://www.onthemedia.org/story/armys-robot-recruiter-aug/transcript>>, Aug. 8, 2014, 3 pages. |
"With Alme, Alaska Airlines soars", Case Study, retrieved on Apr. 10, 2015 at <<http://www.nextit.com/media/downloads/Case-study-Alaska-Air.pdf>>, 3 pages. |
"Meet Julia—TAM Airlines′ most famous new hire," Case Study, 2015, 2 pages. |
Cassell, J., et al., "Embodied Conversational Agents," MIT Press, 2000, pp. 272 and 275. |
Guzzoni, D., et al., "Modeling Human-Agent Interaction with Active Ontologies," 2007 AAAI Spring Symposium, 8 pages. |
Krahmer, E., et al., "Problem Spotting in Human-Machine Interaction," IPO Center for Research on User-System Interaction, Sixth European Conference on Speech Communication and Technology, Budapest, Hungary, 1999, 4 pages. |
Langkilde, I., et al., "Automatic Prediction of Problematic Human-Computer Dialogues in How May I Help You?", AT&T Labs Research, 1999, 5 pages. |
Walker, M., et al., "Learning to Predict Problematic Situations in a Spoken Dialogue System: Experiments with How May I Help You?", AT&T Labs Research, NAACL 2000 Proceedings of the 1st North American chapter of the Association for Computational Linguistics conference, Seattle, Washington, Apr. 29-May 4, 2000, 8 pages. |
Wikipedia page "CALO", retrieved on Nov. 15, 2017 at <<https://en.wikipedia.org/wiki/CALO>>, 5 pages. |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11875118B2 (en) | 2017-05-10 | 2024-01-16 | Oracle International Corporation | Detection of deception within text using communicative discourse trees |
US11586827B2 (en) * | 2017-05-10 | 2023-02-21 | Oracle International Corporation | Generating desired discourse structure from an arbitrary text |
US11373632B2 (en) | 2017-05-10 | 2022-06-28 | Oracle International Corporation | Using communicative discourse trees to create a virtual persuasive dialogue |
US11386274B2 (en) * | 2017-05-10 | 2022-07-12 | Oracle International Corporation | Using communicative discourse trees to detect distributed incompetence |
US12141535B2 (en) | 2017-05-10 | 2024-11-12 | Oracle International Corporation | Techniques for maintaining rhetorical flow |
US20220284194A1 (en) * | 2017-05-10 | 2022-09-08 | Oracle International Corporation | Using communicative discourse trees to detect distributed incompetence |
US11783126B2 (en) | 2017-05-10 | 2023-10-10 | Oracle International Corporation | Enabling chatbots by detecting and supporting affective argumentation |
US12001804B2 (en) * | 2017-05-10 | 2024-06-04 | Oracle International Corporation | Using communicative discourse trees to detect distributed incompetence |
US11775771B2 (en) | 2017-05-10 | 2023-10-03 | Oracle International Corporation | Enabling rhetorical analysis via the use of communicative discourse trees |
US11960844B2 (en) | 2017-05-10 | 2024-04-16 | Oracle International Corporation | Discourse parsing using semantic and syntactic relations |
US11615145B2 (en) | 2017-05-10 | 2023-03-28 | Oracle International Corporation | Converting a document into a chatbot-accessible form via the use of communicative discourse trees |
US11694037B2 (en) | 2017-05-10 | 2023-07-04 | Oracle International Corporation | Enabling rhetorical analysis via the use of communicative discourse trees |
US11748572B2 (en) | 2017-05-10 | 2023-09-05 | Oracle International Corporation | Enabling chatbots by validating argumentation |
US11797773B2 (en) | 2017-09-28 | 2023-10-24 | Oracle International Corporation | Navigating electronic documents using domain discourse trees |
US11782985B2 (en) | 2018-05-09 | 2023-10-10 | Oracle International Corporation | Constructing imaginary discourse trees to improve answering convergent questions |
US11328016B2 (en) | 2018-05-09 | 2022-05-10 | Oracle International Corporation | Constructing imaginary discourse trees to improve answering convergent questions |
US11455494B2 (en) | 2018-05-30 | 2022-09-27 | Oracle International Corporation | Automated building of expanded datasets for training of autonomous agents |
US11868732B2 (en) * | 2018-10-10 | 2024-01-09 | Verint Americas Inc. | System for minimizing repetition in intelligent virtual assistant conversations |
US20220382990A1 (en) * | 2018-10-10 | 2022-12-01 | Verint Americas Inc. | System for minimizing repetition in intelligent virtual assistant conversations |
US11842410B2 (en) | 2019-06-06 | 2023-12-12 | Verint Americas Inc. | Automated conversation review to surface virtual assistant misunderstandings |
US11436416B2 (en) | 2019-06-06 | 2022-09-06 | Verint Americas Inc. | Automated conversation review to surface virtual assistant misunderstandings |
US11577745B2 (en) * | 2020-01-31 | 2023-02-14 | Toyota Jidosha Kabushiki Kaisha | Information processing device, information processing method, and storage medium storing information processing program |
US20210237757A1 (en) * | 2020-01-31 | 2021-08-05 | Toyota Jidosha Kabushiki Kaisha | Information processing device, information processing method, and storage medium storing information processing program |
US12001805B2 (en) * | 2022-04-25 | 2024-06-04 | Gyan Inc. | Explainable natural language understanding platform |
US11842144B1 (en) * | 2022-07-26 | 2023-12-12 | Rammer Technologies, Inc. | Summarizing conversational speech |
Also Published As
Publication number | Publication date |
---|---|
US20190138190A1 (en) | 2019-05-09 |
US20180121062A1 (en) | 2018-05-03 |
US20160071517A1 (en) | 2016-03-10 |
US10175865B2 (en) | 2019-01-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10545648B2 (en) | Evaluating conversation data based on risk factors | |
US11868732B2 (en) | System for minimizing repetition in intelligent virtual assistant conversations | |
US20210173548A1 (en) | Virtual assistant acquisitions and training | |
US11825023B2 (en) | Method and system for virtual assistant conversations | |
US11893311B2 (en) | Virtual assistant configured to automatically customize groups of actions | |
US20190057298A1 (en) | Mapping actions and objects to tasks | |
US11099867B2 (en) | Virtual assistant focused user interfaces | |
US20170277993A1 (en) | Virtual assistant escalation | |
US11093536B2 (en) | Explicit signals personalized search | |
US20220237486A1 (en) | Suggesting activities | |
US10417567B1 (en) | Learning user preferences in a conversational system | |
US11182447B2 (en) | Customized display of emotionally filtered social media content | |
US10444970B1 (en) | Contextual state-based user interface format adaptation | |
CN119226620A (en) | Recommended methods, devices, equipment and media |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
AS | Assignment |
Owner name: VERINT AMERICAS INC., GEORGIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NEXT IT CORPORATION;REEL/FRAME:049469/0615 Effective date: 20180131 Owner name: NEXT IT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BEAVER, IAN;BROWN, FRED;GOSSARD, CASEY;SIGNING DATES FROM 20160115 TO 20160118;REEL/FRAME:049476/0799 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |