US20200081939A1 - System for optimizing detection of intent[s] by automated conversational bot[s] for providing human like responses - Google Patents
System for optimizing detection of intent[s] by automated conversational bot[s] for providing human like responses Download PDFInfo
- Publication number
- US20200081939A1 US20200081939A1 US16/258,680 US201916258680A US2020081939A1 US 20200081939 A1 US20200081939 A1 US 20200081939A1 US 201916258680 A US201916258680 A US 201916258680A US 2020081939 A1 US2020081939 A1 US 2020081939A1
- Authority
- US
- United States
- Prior art keywords
- intent
- bot
- automated conversational
- graph
- dialogues
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000004044 response Effects 0.000 title claims abstract description 48
- 238000001514 detection method Methods 0.000 title claims abstract description 30
- 238000012549 training Methods 0.000 claims abstract description 34
- 230000002787 reinforcement Effects 0.000 claims abstract description 12
- 238000012795 verification Methods 0.000 claims abstract description 8
- 238000000034 method Methods 0.000 claims description 43
- 238000003058 natural language processing Methods 0.000 claims description 42
- 238000000605 extraction Methods 0.000 claims description 15
- 230000015654 memory Effects 0.000 claims description 13
- 238000013507 mapping Methods 0.000 claims description 5
- 230000009471 action Effects 0.000 description 7
- 238000013473 artificial intelligence Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 5
- 238000012545 processing Methods 0.000 description 4
- 238000012360 testing method Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000011161 development Methods 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 3
- 230000009118 appropriate response Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 230000000153 supplemental effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
-
- G06F17/21—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/289—Phrasal analysis, e.g. finite state techniques or chunking
- G06F40/295—Named entity recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/29—Geographical information databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/901—Indexing; Data structures therefor; Storage structures
- G06F16/9024—Graphs; Linked lists
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/903—Querying
- G06F16/9032—Query formulation
- G06F16/90332—Natural language query formulation or dialogue systems
-
- G06F17/27—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/004—Artificial life, i.e. computing arrangements simulating life
- G06N3/006—Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/02—Knowledge representation; Symbolic representation
- G06N5/022—Knowledge engineering; Knowledge acquisition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
Definitions
- the present subject matter described herein in general, relates to a method and system for detecting an intent by an automated conversational bot for providing human like responses. More specifically, the method and system for providing the human like responses upon feeding a call recording archive.
- a customer service representative answers the phone call of a customer and provides required information or action in a friendly manner aligned to the region and culture. It may be noted that all such calls are recorded for training and quality analysis purposes.
- AI Artificial Intelligence
- the enterprises are moving towards automated conversational Voice/Chat based bots that have capabilities to answer the queries of the customer.
- the global Bot market is expected to reach S1.23 billion by 2025.
- These AI based Voice/Chat bots are not like the legacy Interactive Voice Response (IVR) applications built using scripting languages such as Voice XML, where the customer has to go through a menu and select a specific input from a keypad in order to choose an intent.
- IVR Interactive Voice Response
- the AI based Voice/Chat bots have Natural Language Processing (NLP) capabilities that may help in detecting the intent.
- NLP Natural Language Processing
- the AI based Voice/Chat bots may detect the intent
- the intent detection by the AI based Voice/Chat bots may be limited various input variations and utterances that were configured by a developer during the bot building phase.
- the aforementioned approach for detecting the intent is limited to possible intent combinations and appropriate responses as the developer cannot think of all possible utterances during the building of the AI based Voice/Chat bots.
- a system for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user may comprise a processor and a memory coupled to the processor.
- the processor may execute a plurality of modules present in the memory.
- the plurality of modules may comprise an analyzer module, an extraction module, a builder module, and a verification module.
- the analyzer module may build an intent graph storing input dialogues, utterances, and output dialogues associated to an intent.
- the intent may indicate a context of a conversation between a caller and a call center representative.
- the extraction module may feed each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine to create a set of raw text transcripts pertaining to a category.
- NLP Natural Language Processing
- the extraction module may further determine a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts.
- each intent may be associated to at least one category.
- the extraction module may further map the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent.
- the builder module may feed training data, comprising the intent graph stored in the graph database, to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances.
- the verification module may train the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot.
- the automated conversational bot may be trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response.
- the output dialogue may be provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
- a method for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user is disclosed.
- an intent graph storing input dialogues, utterances, and output dialogues associated to an intent may be built.
- the intent may indicate a context of a conversation between a caller and a call center representative.
- the intent graph is built by feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category.
- NLP Natural Language Processing
- a plurality of intents may be determined from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts.
- each intent may be associated to at least one category.
- the input dialogues, the utterances, and the output dialogues may be mapped with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent.
- training data may be fed to an automated conversational bot by enabling a bot builder in order to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances.
- the training data may comprise the intent graph stored in the graph database.
- the automated conversational bot may be trained through reinforcement learning by providing a feedback to the automated conversational bot.
- the automated conversational bot may be trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
- the aforementioned method for optimizing detection of the intent by the automated conversational bot may be performed by a processor using programmed instructions stored in a memory of the system.
- non-transitory computer readable medium embodying a program executable in a computing device for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user characterized by feeding a call recording archive to the automated conversational bot is disclosed.
- the program may comprise a program code for building, by a processor, an intent graph storing input dialogues, utterances, and output dialogues associated to an intent, wherein the intent indicates a context of a conversation between a caller and a call center representative, and wherein the intent graph is built by feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category, determining a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts, wherein each intent is associated to at least one category, mapping the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent.
- NLP Natural Language Processing
- the program may further comprise a program code for feeding, by the processor, training data, comprising the intent graph stored in the graph database, to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances.
- the program may further comprise a program code for training, by the processor, the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot, wherein the automated conversational bot is trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
- FIG. 1 illustrates a network implementation of a system for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user, in accordance with an embodiment of the present subject matter.
- FIG. 2 illustrates the system, in accordance with an embodiment of the present subject matter.
- FIGS. 3 illustrates various components of the system, in accordance with an embodiment of the present subject matter.
- FIG. 4 illustrates a method for optimizing detection of the intent by an automated conversational bot for providing human like responses to the user, in accordance with an embodiment of the present subject matter.
- the proposed invention facilitates to simplify and accelerate development of an automated conversational bot by feeding valuable corpus of audio files comprising call recordings archive to a system in order to provide human like responses to callers.
- the system cleanses each call recording in accordance with a requirements of a type of automated conversational bot (such as AlexaTM, SlackTM GoogleTM Assistant, etc.).
- the system cleanses each call recording based on a set of filters. Examples of the set of filters may include, but not limited to, voice gender of the caller and the call center representative, language used in the call, the at least one category associated to the intent, and call duration.
- each call recording may be fed into a Natural Language Processing (NLP) engine which is then read by the NLP engine to create a set of raw text transcripts pertaining to a recording category.
- NLP Natural Language Processing
- the recording category in Bank domain may include, but not limited to, New account, Existing customer, and Lost Card.
- high-level NLP entities like Noun, Verbs, Question segment and Answer segment may be created.
- the set of raw text transcripts may then be processed to determine a plurality of intents from high-level NLP entities present in the set of raw text transcripts.
- each intent may be mapped with input dialogues, utterances, and output dialogues thereby building an intent graph pertaining to each intent and storing the intent graph in a graph database.
- the intent graph may be fed into an automated conversational bot and enable a bot builder to fill in a bot template with a set of parameters with distinct utterances of an intent and output dialogues associated to the distinct utterances. It may be noted that intent graph is fed upon building a specific Bot script to bootstrap the development of the automated conversational bot by using specific templates and user configured actions such as invoke REST API or Lambda functions.
- the automated conversational bot After feeding the intent graph, the automated conversational bot is trained by providing a feedback to the automated conversational bot. In other words, such building of the intent graph for a specific automated conversational bot assists the developer in testing the automated conversational bot using the raw text transcripts / audio files extracted from the call recordings archive.
- the aforementioned methodology may enable the developer in continuous tuning/training the automated conversational bot by feeding training data comprising the intent graph as part of the Machine learning model training purpose.
- a network implementation 100 of a system 102 for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user is disclosed.
- the system 102 builds an intent graph storing input dialogues, utterances, and output dialogues associated to an intent.
- the intent may indicate a context of a conversation between a caller and a call center representative.
- the system further 102 feeds each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category.
- NLP Natural Language Processing
- the system 102 further determines a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts.
- each intent may be associated to at least one category.
- the system 102 further maps the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent.
- the system 102 further feeds training data, comprising the intent graph stored in the graph database, to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances.
- the system 102 further trains the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot.
- the automated conversational bot may be trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
- system 102 may be implemented in a variety of computing systems, such as a laptop computer, a desktop computer, a notebook, a workstation, a mainframe computer, a server, a network server, a cloud-based computing environment. It will be understood that the system 102 may be accessed by multiple users through one or more user devices 104 - 1 , 104 - 2 . . . 104 -N, collectively referred to as user 104 or stakeholders, hereinafter, or applications residing on the user devices 104 . In one implementation, the system 102 may comprise the cloud-based computing environment in which a user may operate individual computing systems configured to execute remotely located applications.
- Examples of the user devices 104 may include, but are not limited to, a IoT device, IoT gateway, portable computer, a personal digital assistant, a handheld device, and a workstation.
- the user devices 104 are communicatively coupled to the system 102 through a network 106 .
- the network 106 may be a wireless network, a wired network or a combination thereof.
- the network 106 can be implemented as one of the different types of networks, such as intranet, local area network (LAN), wide area network (WAN), the internet, and the like.
- the network 106 may either be a dedicated network or a shared network.
- the shared network represents an association of the different types of networks that use a variety of protocols, for example, Hypertext Transfer Protocol (HTTP), Hypertext Transfer Protocol Secure (HTTPS), Transmission Control Protocol/Internet Protocol (TCP/IP), Wireless Application Protocol (WAP), and the like, to communicate with one another.
- the network 106 may include a variety of network devices, including routers, bridges, servers, computing devices, storage devices, and the like.
- the system 102 may include at least one processor 202 , an input/output (I/O) interface 204 , and a memory 206 .
- the at least one processor 202 may be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions.
- the at least one processor 202 is configured to fetch and execute computer-readable instructions stored in the memory 206 .
- the I/O interface 204 may include a variety of software and hardware interfaces, for example, a web interface, a graphical user interface, and the like.
- the I/O interface 204 may allow the system 102 to interact with the user directly or through the user devices 104 . Further, the I/O interface 204 may enable the system 102 to communicate with other computing devices, such as web servers and external data servers (not shown).
- the I/O interface 204 can facilitate multiple communications within a wide variety of networks and protocol types, including wired networks, for example, LAN, cable, etc., and wireless networks, such as WLAN, cellular, or satellite.
- the I/O interface 204 may include one or more ports for connecting a number of devices to one another or to another server.
- the memory 206 may include any computer-readable medium or computer program product known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non-volatile memory, such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
- volatile memory such as static random access memory (SRAM) and dynamic random access memory (DRAM)
- non-volatile memory such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
- ROM read only memory
- erasable programmable ROM erasable programmable ROM
- the modules 208 include routines, programs, objects, components, data structures, etc., which perform particular tasks or implement particular abstract data types.
- the modules 208 may include an analyzer module 212 , an extraction module 214 , a builder module 216 , a verification module 218 , and other modules 220 .
- the other modules 220 may include programs or coded instructions that supplement applications and functions of the system 102 .
- the modules 208 described herein may be implemented as software modules that may be executed in the cloud-based computing environment of the system 102 .
- the data 210 serves as a repository for storing data processed, received, and generated by one or more of the modules 208 .
- the data 210 may also include a graph database 222 and other data 224 .
- the other data 224 may include data generated as a result of the execution of one or more modules in the other modules 220 .
- a user may use the user device 104 to access the system 102 via the I/O interface 204 .
- the user may register them using the I/O interface 204 to use the system 102 .
- the user may access the I/O interface 204 of the system 102 .
- the system 102 may employ the analyzer module 212 , the extraction module 214 , the builder module 216 , and the verification module 218 . The detail functioning of the modules is described below with the help of figures.
- a major process in developing the automated conversational bot is to detect the intent pertaining to a query and correct human like response for the intent outcome.
- a developer of the automated conversational bot has to think of various combinations and utterances of possible voice/chat inputs from end user for detecting the intent and its associated inputs.
- the developer may embed of a limited set of possibilities of asking queries having similar intent.
- the system 102 uses a call recording archive comprising a corpus of real life conversations between callers and call center representatives.
- the call recording archive comprises a corpus of real life conversations for all possible intents.
- the system continuously trains the automated conversational bot by feeding the call recording archive that may be used to train the automated conversational bot in providing the human like responses.
- the developer may then focus on core intent actions instead of thinking about the Natural language part of the automated conversational bot.
- the analyzer module 212 builds an intent graph storing input dialogues, utterances, and output dialogues associated to an intent based on the call recording archive 300 fed into the automated conversational bot.
- the intent indicates a context of a conversation between a caller and a call center representative.
- the call recording archive 300 may be accessed from a distinct set of data sources.
- the distinct set of data sources may include, but not limited, Network Attached Storage (NAS), Storage Area Network (SAN), Object Storage, and File Servers.
- NAS Network Attached Storage
- SAN Storage Area Network
- Object Storage Object Storage
- File Servers the call recording archive 300 from the aforementioned data sources may be accessed through one or more adapters for each of the above data sources.
- NFS Client to access NFS shares may be accessed through one or more adapters for each of the above data sources.
- NFS Client to access NFS shares
- REST Clients to access Object Storage
- SCP/FTP Clients to access SCP/FTP based File servers.
- each audio file containing the call recording is stored in a standard audio format such as WAV or MP3.
- the extraction module may use decoders for each file format to retrieved the conversation present in each audio file.
- each call recording present in the call recording archive 300 may be cleansed based on one or more filters.
- the one or more filters may include, but not limited to, voice gender of the caller and the call center representative, language used in the call, the at least one category associated to the intent, and call duration. It may be noted that each call recording may be cleansed to select an appropriate set of audio files as per the configuration of an automated conversational bot.
- each call recording includes metadata for each call recorded and stores the data in following ways
- the extraction module 214 extracts the metadata, stored as above, from each call recording by following methodologies respectively.
- the metadata may be used against the filters to select the appropriate set of audio files as per the configuration of the automated conversational bot.
- the extraction module 214 fed each audio file, selected, to a Natural Language Processing (NLP) engine 304 in order to create a set of raw text transcripts pertaining to a category.
- the category in ‘Banking’ domain may include, but not limited to, New account, Existing customer, and Lost Card.
- the extraction module 214 creates the set of raw text transcripts from the appropriate set of audio files by using a speech to text engine 302 and then provided to the NLP engine 304 for further processing.
- the set of raw text transcripts is stored in a transcript store 306 .
- the extraction module 214 further determines a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts.
- the extraction module 214 determines one or more intents from the set of raw text transcripts pertaining to a category, as follows. It may be understood that the category is a ‘Banking’
- each intent is associated to at least one category.
- the extraction module 214 maps the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent.
- the intent graph may be built by using a conceptual graph concept and stored in the graph database 222 .
- the conceptual graph concept is used to implement Question and Answer systems for Fuzzy logic based AI model.
- the input dialogues, the utterances, and the output dialogues may be mapped with each intent by using a WordNet Ontology.
- the analyzer module 212 further comprises an intent analyzer 308 which analyzes that what action needs to be taken for an intent that has been detected.
- the builder module 216 feeds training data, comprising the intent graph stored in the graph database 222 , to the automated conversational bot.
- the training data may be fed by enabling a bot builder to fill a specific bot template stored in a template store 310 .
- the builder module 216 further comprises an intent action store 312 .
- the intent action store 312 comprises mapping of action[s] that needs to be performed against the intent that has been detected by the intent analyzer 308 .
- the builder module 216 further comprises a Voice Assistant Dialog Engine 314 /a Bot Dialog Engine 316 .
- both the Voice Assistant Dialog Engine 314 /the Bot Dialog Engine 316 contain generic templates for the development of the automated conversational bot. Since each bot platform have their own format, the Voice Assistant Dialog Engine 314 /the Bot Dialog Engine 316 are configured to convert the generic template to the Bot platform specific template.
- each automated conversational bot has its own bot template that may be filled with a set of parameters.
- the set of parameters indicates distinct utterances of an intent and output dialogues associated to the distinct utterances.
- the bot template is mentioned as follows.
- respective bot template processor- 1 , 2 , 3 . . . , N of the builder module 216 queries the graph database 222 to retrieve possible strings including the output dialogues for the intent.
- the strings may also be queried using the “Potential Answer” query offered by the graph database 222 , which in turn fetch all the possible answers matching the question related to intent.
- the automated conversational bot (such as AlexaTM, SlackTM, GoogleTM Assistant) is deployed on a test environment.
- the set of raw transcripts extracted is fed as the training data to the automated conversational bot and the output dialogue is validated against an expected response by a validator 318 . If the response has anomalies, the expected response is fed again to train the automated conversational model.
- the verification module 218 enables a trainer 320 to train the automated conversational bot through a reinforcement learning method.
- the reinforcement learning method includes providing a feedback to the automated conversational bot by validating an output dialogue, against an input dialogue received from the caller, with an expected response. Over a period of time, the reinforcement learning makes the automated conversational bot to gradually learn from the feedback and enables the developer in continuous tuning/training the automated conversational bot by the feedback as part of the Machine learning model training purpose.
- the system 102 optimizes detection of the intent and thereby provides human like responses.
- the system 102 tests and trains the automated conversational bot using the set of raw text transcripts and audio from the call recording archives thereby attaining human like natural language processing including regional slangs and phrases.
- a method 400 for detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user is shown, in accordance with an embodiment of the present subject matter.
- the method 400 may be described in the general context of computer executable instructions.
- computer executable instructions can include routines, programs, objects, components, data structures, procedures, modules, functions, etc., that perform particular functions or implement particular abstract data types.
- the method 400 may also be practiced in a distributed computing environment where functions are performed by remote processing devices that are linked through a communications network.
- computer executable instructions may be located in both local and remote computer storage media, including memory storage devices.
- the order in which the method 400 is described is not intended to be construed as a limitation, and any number of the described method blocks can be combined in any order to implement the method 400 or alternate methods. Additionally, individual blocks may be deleted from the method 400 without departing from the spirit and scope of the subject matter described herein. Furthermore, the method can be implemented in any suitable hardware, software, firmware, or combination thereof. However, for ease of explanation, in the embodiments described below, the method 400 may be considered to be implemented as described in the system 102 .
- an intent graph storing input dialogues, utterances, and output dialogues associated to an intent may be built.
- the intent indicates a context of a conversation between a caller and a call center representative.
- the intent graph is built by feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category, determining a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts, wherein each intent is associated to at least one category, and mapping the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent.
- the intent graph may be built by the analyzer module 212 .
- training data comprising the intent graph stored in the graph database, may be fed to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances.
- the training data may be fed by the builder module 216 .
- the automated conversational bot may be trained through reinforcement learning by providing a feedback to the automated conversational bot.
- the automated conversational bot may be trained by validating an output dialogue against, an input dialogue received from the caller, with an expected response.
- the output dialogue may be provided by the automated conversational bot based on the bot template and the training data.
- the automated conversational bot may be trained by the verification module 218 .
- Some embodiments enable a system and a method to assist bot developer in designing the possible input utterances for intent detection and intent inputs.
- Some embodiments enable a system and a method to assist in forming human like responses for all possible intent responses including emotions and sentiments.
- Some embodiments enable a system and a method to test and train the bot using a set of raw transcripts and audio from voice archives thereby attaining human like natural language processing including regional slangs and phrases.
- Some embodiments enable a system and a method to reuse of existing contact center call recording archives instead of engaging in research to obtain input combinations thereby saving time in developing the bot with all possible utterances for intent detections.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Remote Sensing (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Machine Translation (AREA)
Abstract
Disclosed is a system for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user. An analyzer module builds an intent graph storing input dialogues, utterances, and output dialogues associated to an intent. A builder module fed training data, comprising the intent graph stored in the graph database to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances. A verification module trains the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot. In one aspect, the automated conversational bot may be trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response.
Description
- The present application claims benefit from Indian Complete Patent Application No. 201811034198 filed on 11 Sep. 2018 the entirety of which is hereby incorporated by reference.
- The present subject matter described herein, in general, relates to a method and system for detecting an intent by an automated conversational bot for providing human like responses. More specifically, the method and system for providing the human like responses upon feeding a call recording archive.
- For the last couple of decades, enterprises are using telephone based contact centers to provide support to their customers. A customer service representative answers the phone call of a customer and provides required information or action in a friendly manner aligned to the region and culture. It may be noted that all such calls are recorded for training and quality analysis purposes.
- In the last few years, due to the advent of Artificial Intelligence (AI) using Machine Learning, the enterprises are moving towards automated conversational Voice/Chat based bots that have capabilities to answer the queries of the customer. According to a recent report by Grand View Research, the global Bot market is expected to reach S1.23 billion by 2025. These AI based Voice/Chat bots are not like the legacy Interactive Voice Response (IVR) applications built using scripting languages such as Voice XML, where the customer has to go through a menu and select a specific input from a keypad in order to choose an intent. Instead, the AI based Voice/Chat bots have Natural Language Processing (NLP) capabilities that may help in detecting the intent.
- Though the AI based Voice/Chat bots may detect the intent, the intent detection by the AI based Voice/Chat bots may be limited various input variations and utterances that were configured by a developer during the bot building phase. In other words, the aforementioned approach for detecting the intent is limited to possible intent combinations and appropriate responses as the developer cannot think of all possible utterances during the building of the AI based Voice/Chat bots.
- Before the present systems and methods, are described, it is to be understood that this application is not limited to the particular systems, and methodologies described, as there can be multiple possible embodiments which are not expressly illustrated in the present disclosure. It is also to be understood that the terminology used in the description is for the purpose of describing the particular versions or embodiments only, and is not intended to limit the scope of the present application. This summary is provided to introduce concepts related to systems and methods for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user and the concepts are further described below in the detailed description. This summary is not intended to identify essential features of the claimed subject matter nor is it intended for use in limiting the scope of the claimed subject matter.
- In one implementation, a system for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user is disclosed. The system may comprise a processor and a memory coupled to the processor. The processor may execute a plurality of modules present in the memory. The plurality of modules may comprise an analyzer module, an extraction module, a builder module, and a verification module. The analyzer module may build an intent graph storing input dialogues, utterances, and output dialogues associated to an intent. In one aspect, the intent may indicate a context of a conversation between a caller and a call center representative. In order to build the intent graph, the extraction module may feed each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine to create a set of raw text transcripts pertaining to a category. The extraction module may further determine a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts. In one aspect, each intent may be associated to at least one category. The extraction module may further map the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent. The builder module may feed training data, comprising the intent graph stored in the graph database, to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances. The verification module may train the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot. In one aspect, the automated conversational bot may be trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response. The output dialogue may be provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
- In another implementation, a method for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user is disclosed. In order to optimize detection of the intent, initially, an intent graph storing input dialogues, utterances, and output dialogues associated to an intent may be built. In one aspect, the intent may indicate a context of a conversation between a caller and a call center representative. In one aspect, the intent graph is built by feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category. Upon feeding each audio file, a plurality of intents may be determined from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts. In one aspect, each intent may be associated to at least one category. Subsequent to the determination of the plurality of intents, the input dialogues, the utterances, and the output dialogues may be mapped with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent. Post building the intent graph, training data may be fed to an automated conversational bot by enabling a bot builder in order to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances. In one aspect, the training data may comprise the intent graph stored in the graph database. Upon feeding the training data, the automated conversational bot may be trained through reinforcement learning by providing a feedback to the automated conversational bot. In one aspect, the automated conversational bot may be trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive. In one aspect, the aforementioned method for optimizing detection of the intent by the automated conversational bot may be performed by a processor using programmed instructions stored in a memory of the system.
- In yet another implementation, non-transitory computer readable medium embodying a program executable in a computing device for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user characterized by feeding a call recording archive to the automated conversational bot is disclosed. The program may comprise a program code for building, by a processor, an intent graph storing input dialogues, utterances, and output dialogues associated to an intent, wherein the intent indicates a context of a conversation between a caller and a call center representative, and wherein the intent graph is built by feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category, determining a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts, wherein each intent is associated to at least one category, mapping the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent. The program may further comprise a program code for feeding, by the processor, training data, comprising the intent graph stored in the graph database, to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances. The program may further comprise a program code for training, by the processor, the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot, wherein the automated conversational bot is trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
- The foregoing detailed description of embodiments is better understood when read in conjunction with the appended drawings. For the purpose of illustrating the disclosure, example constructions of the disclosure are shown in the present document; however, the disclosure is not limited to the specific methods and apparatus disclosed in the document and the drawings.
- The detailed description is given with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same numbers are used throughout the drawings to refer like features and components.
-
FIG. 1 illustrates a network implementation of a system for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user, in accordance with an embodiment of the present subject matter. -
FIG. 2 illustrates the system, in accordance with an embodiment of the present subject matter. -
FIGS. 3 illustrates various components of the system, in accordance with an embodiment of the present subject matter. -
FIG. 4 illustrates a method for optimizing detection of the intent by an automated conversational bot for providing human like responses to the user, in accordance with an embodiment of the present subject matter. - Some embodiments of this disclosure, illustrating all its features, will now be discussed in detail. The words “comprising,” “having,” “containing,” and “including,” and other forms thereof, are intended to be equivalent in meaning and be open ended in that an item or items following any one of these words is not meant to be an exhaustive listing of such item or items, or meant to be limited to only the listed item or items. It must also be noted that as used herein and in the appended claims, the singular forms “a,” “an,” and “the” include plural references unless the context clearly dictates otherwise. Although any systems and methods similar or equivalent to those described herein can be used in the practice, the exemplary, systems and methods are now described. The disclosed embodiments are merely exemplary of the disclosure, which may be embodied in various forms.
- Various modifications to the embodiment will be readily apparent to those skilled in the art and the generic principles herein may be applied to other embodiments. However, one of ordinary skill in the art will readily recognize that the present disclosure is not intended to be limited to the embodiments illustrated, but is to be accorded the widest scope consistent with the principles and features described herein.
- The proposed invention facilitates to simplify and accelerate development of an automated conversational bot by feeding valuable corpus of audio files comprising call recordings archive to a system in order to provide human like responses to callers. Upon feeding the call recordings archive, the system cleanses each call recording in accordance with a requirements of a type of automated conversational bot (such as Alexa™, Slack™ Google™ Assistant, etc.). In one aspect, the system cleanses each call recording based on a set of filters. Examples of the set of filters may include, but not limited to, voice gender of the caller and the call center representative, language used in the call, the at least one category associated to the intent, and call duration.
- Upon cleansing, each call recording may be fed into a Natural Language Processing (NLP) engine which is then read by the NLP engine to create a set of raw text transcripts pertaining to a recording category. For example, the recording category in Bank domain may include, but not limited to, New account, Existing customer, and Lost Card. Using the set of raw text transcripts, high-level NLP entities like Noun, Verbs, Question segment and Answer segment may be created. The set of raw text transcripts may then be processed to determine a plurality of intents from high-level NLP entities present in the set of raw text transcripts. Upon determination of the plurality of intents, each intent may be mapped with input dialogues, utterances, and output dialogues thereby building an intent graph pertaining to each intent and storing the intent graph in a graph database.
- Once the intent graph is built, the intent graph may be fed into an automated conversational bot and enable a bot builder to fill in a bot template with a set of parameters with distinct utterances of an intent and output dialogues associated to the distinct utterances. It may be noted that intent graph is fed upon building a specific Bot script to bootstrap the development of the automated conversational bot by using specific templates and user configured actions such as invoke REST API or Lambda functions.
- After feeding the intent graph, the automated conversational bot is trained by providing a feedback to the automated conversational bot. In other words, such building of the intent graph for a specific automated conversational bot assists the developer in testing the automated conversational bot using the raw text transcripts / audio files extracted from the call recordings archive. The aforementioned methodology may enable the developer in continuous tuning/training the automated conversational bot by feeding training data comprising the intent graph as part of the Machine learning model training purpose.
- While aspects of described system and method for optimizing detection of the intent by the automated conversational bot for providing human like responses may be implemented in any number of different computing systems, environments, and/or configurations, the embodiments are described in the context of the following exemplary system.
- Referring now to
FIG. 1 , anetwork implementation 100 of asystem 102 for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user is disclosed. Thesystem 102 builds an intent graph storing input dialogues, utterances, and output dialogues associated to an intent. In one aspect, the intent may indicate a context of a conversation between a caller and a call center representative. The system further 102 feeds each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category. Thesystem 102 further determines a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts. In one aspect, each intent may be associated to at least one category. Thesystem 102 further maps the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent. Thesystem 102 further feeds training data, comprising the intent graph stored in the graph database, to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances. Thesystem 102 further trains the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot. In one aspect, the automated conversational bot may be trained by validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive. - Although the present disclosure is explained considering that the
system 102 is implemented on a server, it may be understood that thesystem 102 may be implemented in a variety of computing systems, such as a laptop computer, a desktop computer, a notebook, a workstation, a mainframe computer, a server, a network server, a cloud-based computing environment. It will be understood that thesystem 102 may be accessed by multiple users through one or more user devices 104-1, 104-2 . . . 104-N, collectively referred to asuser 104 or stakeholders, hereinafter, or applications residing on theuser devices 104. In one implementation, thesystem 102 may comprise the cloud-based computing environment in which a user may operate individual computing systems configured to execute remotely located applications. Examples of theuser devices 104 may include, but are not limited to, a IoT device, IoT gateway, portable computer, a personal digital assistant, a handheld device, and a workstation. Theuser devices 104 are communicatively coupled to thesystem 102 through anetwork 106. - In one implementation, the
network 106 may be a wireless network, a wired network or a combination thereof. Thenetwork 106 can be implemented as one of the different types of networks, such as intranet, local area network (LAN), wide area network (WAN), the internet, and the like. Thenetwork 106 may either be a dedicated network or a shared network. The shared network represents an association of the different types of networks that use a variety of protocols, for example, Hypertext Transfer Protocol (HTTP), Hypertext Transfer Protocol Secure (HTTPS), Transmission Control Protocol/Internet Protocol (TCP/IP), Wireless Application Protocol (WAP), and the like, to communicate with one another. Further thenetwork 106 may include a variety of network devices, including routers, bridges, servers, computing devices, storage devices, and the like. - Referring now to
FIG. 2 , thesystem 102 is illustrated in accordance with an embodiment of the present subject matter. In one embodiment, thesystem 102 may include at least oneprocessor 202, an input/output (I/O)interface 204, and amemory 206. The at least oneprocessor 202 may be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions. Among other capabilities, the at least oneprocessor 202 is configured to fetch and execute computer-readable instructions stored in thememory 206. - The I/
O interface 204 may include a variety of software and hardware interfaces, for example, a web interface, a graphical user interface, and the like. The I/O interface 204 may allow thesystem 102 to interact with the user directly or through theuser devices 104. Further, the I/O interface 204 may enable thesystem 102 to communicate with other computing devices, such as web servers and external data servers (not shown). The I/O interface 204 can facilitate multiple communications within a wide variety of networks and protocol types, including wired networks, for example, LAN, cable, etc., and wireless networks, such as WLAN, cellular, or satellite. The I/O interface 204 may include one or more ports for connecting a number of devices to one another or to another server. - The
memory 206 may include any computer-readable medium or computer program product known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non-volatile memory, such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes. Thememory 206 may includemodules 208 anddata 210. - The
modules 208 include routines, programs, objects, components, data structures, etc., which perform particular tasks or implement particular abstract data types. In one implementation, themodules 208 may include ananalyzer module 212, anextraction module 214, abuilder module 216, averification module 218, andother modules 220. Theother modules 220 may include programs or coded instructions that supplement applications and functions of thesystem 102. Themodules 208 described herein may be implemented as software modules that may be executed in the cloud-based computing environment of thesystem 102. - The
data 210, amongst other things, serves as a repository for storing data processed, received, and generated by one or more of themodules 208. Thedata 210 may also include agraph database 222 andother data 224. Theother data 224 may include data generated as a result of the execution of one or more modules in theother modules 220. - As there are various challenges observed in the existing art, the challenges necessitate the need to build the
system 102 for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user. In order to enable the automated conversational bot to provide human like responses to a user, at first, a user may use theuser device 104 to access thesystem 102 via the I/O interface 204. The user may register them using the I/O interface 204 to use thesystem 102. In one aspect, the user may access the I/O interface 204 of thesystem 102. Thesystem 102 may employ theanalyzer module 212, theextraction module 214, thebuilder module 216, and theverification module 218. The detail functioning of the modules is described below with the help of figures. - It may be noted that a major process in developing the automated conversational bot is to detect the intent pertaining to a query and correct human like response for the intent outcome. As part of this process, a developer of the automated conversational bot has to think of various combinations and utterances of possible voice/chat inputs from end user for detecting the intent and its associated inputs. To an extent, the developer may embed of a limited set of possibilities of asking queries having similar intent. However, it may not be feasible for the developer to think all possible intent combinations and appropriate responses and embed the same into the automated conversational bot to provide the human like responses.
- To overcome the aforementioned limitation, the
system 102 uses a call recording archive comprising a corpus of real life conversations between callers and call center representatives. The call recording archive comprises a corpus of real life conversations for all possible intents. With the call recording archive, the system continuously trains the automated conversational bot by feeding the call recording archive that may be used to train the automated conversational bot in providing the human like responses. As a result of this, the developer may then focus on core intent actions instead of thinking about the Natural language part of the automated conversational bot. - Further referring to
FIGS. 2 and 3 . To facilitate the above, theanalyzer module 212 builds an intent graph storing input dialogues, utterances, and output dialogues associated to an intent based on thecall recording archive 300 fed into the automated conversational bot. In one aspect, the intent indicates a context of a conversation between a caller and a call center representative. - In order to feed the
call recording archive 300, thecall recording archive 300 may be accessed from a distinct set of data sources. The distinct set of data sources may include, but not limited, Network Attached Storage (NAS), Storage Area Network (SAN), Object Storage, and File Servers. It may be noted that thecall recording archive 300 from the aforementioned data sources may be accessed through one or more adapters for each of the above data sources. For example, NFS Client to access NFS shares, File System drivers to mount SAN block devices, REST Clients to access Object Storage such as AWS S3, and SCP/FTP Clients to access SCP/FTP based File servers. It may further be noted that each audio file containing the call recording is stored in a standard audio format such as WAV or MP3. The extraction module may use decoders for each file format to retrieved the conversation present in each audio file. - Upon decoding, each call recording present in the
call recording archive 300 may be cleansed based on one or more filters. The one or more filters may include, but not limited to, voice gender of the caller and the call center representative, language used in the call, the at least one category associated to the intent, and call duration. It may be noted that each call recording may be cleansed to select an appropriate set of audio files as per the configuration of an automated conversational bot. In one embodiment, each call recording includes metadata for each call recorded and stores the data in following ways -
- 1. As part of the Audio file (MP3/WAV headers)
- 2. Additional metadata file (JSON/XML)
- 3. Providing an API
- The
extraction module 214 extracts the metadata, stored as above, from each call recording by following methodologies respectively. -
- 1. MP3/WAV file decoders to parse the metadata in audio file.
- 2. JSON/XML parsers to fetch the metadata from supplemental file for a given audio file.
- 3. API Clients to parse the data from HTTP header or Body.
- Once the metadata is extracted, the metadata may be used against the filters to select the appropriate set of audio files as per the configuration of the automated conversational bot. Subsequent to the extraction of the metadata, the
extraction module 214 fed each audio file, selected, to a Natural Language Processing (NLP)engine 304 in order to create a set of raw text transcripts pertaining to a category. In one example, the category in ‘Banking’ domain may include, but not limited to, New account, Existing customer, and Lost Card. It may be noted that theextraction module 214 creates the set of raw text transcripts from the appropriate set of audio files by using a speech totext engine 302 and then provided to theNLP engine 304 for further processing. Upon processing by theNLP engine 304, the set of raw text transcripts is stored in atranscript store 306. - Subsequently, the
extraction module 214 further determines a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts. In one example, theextraction module 214 determines one or more intents from the set of raw text transcripts pertaining to a category, as follows. It may be understood that the category is a ‘Banking’ - Case 1: Customer—Hi, I would like to know the balance of my checking
- Bot Response—Sure. Your balance as of today is XXX dollars
- Case 2: Customer—How much can I withdraw today?
- Bot Response—Let me Check. As of today the maximum cash you can withdraw is XXX Dollars
- The intent in the above cases (1) and (2) is “Account Balance”.
- Case 3: Customer—My ATM pin is not working. I am very frustrated
- Bot Response—Sorry to hear that. Let me fix that for you
- Case 4: Customer—I want to close my account since your ATM is not working
- Bot Response—Thanks for the feedback. Please provide an opportunity to correct it for you
- The intent in the above cases (3) and (4) is “ATM Complaint Intent”. From the above examples, it must be noted that each intent is associated to at least one category.
- After the determination of the plurality of intents, the
extraction module 214 maps the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent. In one aspect, the intent graph may be built by using a conceptual graph concept and stored in thegraph database 222. The conceptual graph concept is used to implement Question and Answer systems for Fuzzy logic based AI model. In one embodiment, the input dialogues, the utterances, and the output dialogues may be mapped with each intent by using a WordNet Ontology. - In one embodiment, the
analyzer module 212 further comprises anintent analyzer 308 which analyzes that what action needs to be taken for an intent that has been detected. - Once the intent graph is built, the
builder module 216 feeds training data, comprising the intent graph stored in thegraph database 222, to the automated conversational bot. The training data may be fed by enabling a bot builder to fill a specific bot template stored in atemplate store 310. Thebuilder module 216 further comprises anintent action store 312. Theintent action store 312 comprises mapping of action[s] that needs to be performed against the intent that has been detected by theintent analyzer 308. Thebuilder module 216 further comprises a VoiceAssistant Dialog Engine 314/aBot Dialog Engine 316. In one aspect, both the VoiceAssistant Dialog Engine 314/theBot Dialog Engine 316 contain generic templates for the development of the automated conversational bot. Since each bot platform have their own format, the VoiceAssistant Dialog Engine 314/theBot Dialog Engine 316 are configured to convert the generic template to the Bot platform specific template. - Further it may be understood that each automated conversational bot has its own bot template that may be filled with a set of parameters. In one aspect, the set of parameters indicates distinct utterances of an intent and output dialogues associated to the distinct utterances. In an example of the bot template is mentioned as follows.
-
“resource”: { “name”: “<#Name>”, “version”: “<#Version>”, “intents”: [ { “name”: “<#IntentName>”, “version”: “<#Version>”, “fulfillmentActivity”: { “type”: “ReturnIntent” }, “sampleUtterances”: [ “#<Utterance1>”, “#<Utterance2>” ] - In order to fill the set of parameters in the bot template, as aforementioned, respective bot template processor-1, 2, 3 . . . , N of the
builder module 216 queries thegraph database 222 to retrieve possible strings including the output dialogues for the intent. In addition, the strings may also be queried using the “Potential Answer” query offered by thegraph database 222, which in turn fetch all the possible answers matching the question related to intent. - Subsequent to the feeding of the training data, the automated conversational bot (such as Alexa™, Slack™, Google™ Assistant) is deployed on a test environment. The set of raw transcripts extracted is fed as the training data to the automated conversational bot and the output dialogue is validated against an expected response by a
validator 318. If the response has anomalies, the expected response is fed again to train the automated conversational model. In other words, theverification module 218 enables atrainer 320 to train the automated conversational bot through a reinforcement learning method. The reinforcement learning method includes providing a feedback to the automated conversational bot by validating an output dialogue, against an input dialogue received from the caller, with an expected response. Over a period of time, the reinforcement learning makes the automated conversational bot to gradually learn from the feedback and enables the developer in continuous tuning/training the automated conversational bot by the feedback as part of the Machine learning model training purpose. - This, in this manner, the
system 102 optimizes detection of the intent and thereby provides human like responses. In other words, thesystem 102 tests and trains the automated conversational bot using the set of raw text transcripts and audio from the call recording archives thereby attaining human like natural language processing including regional slangs and phrases. - Referring now to
FIG. 4 , amethod 400 for detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user is shown, in accordance with an embodiment of the present subject matter. Themethod 400 may be described in the general context of computer executable instructions. Generally, computer executable instructions can include routines, programs, objects, components, data structures, procedures, modules, functions, etc., that perform particular functions or implement particular abstract data types. Themethod 400 may also be practiced in a distributed computing environment where functions are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, computer executable instructions may be located in both local and remote computer storage media, including memory storage devices. - The order in which the
method 400 is described is not intended to be construed as a limitation, and any number of the described method blocks can be combined in any order to implement themethod 400 or alternate methods. Additionally, individual blocks may be deleted from themethod 400 without departing from the spirit and scope of the subject matter described herein. Furthermore, the method can be implemented in any suitable hardware, software, firmware, or combination thereof. However, for ease of explanation, in the embodiments described below, themethod 400 may be considered to be implemented as described in thesystem 102. - At
block 402, an intent graph storing input dialogues, utterances, and output dialogues associated to an intent may be built. In one aspect, the intent indicates a context of a conversation between a caller and a call center representative. In one aspect, the intent graph is built by feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category, determining a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts, wherein each intent is associated to at least one category, and mapping the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent. In one implementation, the intent graph may be built by theanalyzer module 212. - At
block 404, training data, comprising the intent graph stored in the graph database, may be fed to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances. In one implementation, the training data may be fed by thebuilder module 216. - At
block 406, the automated conversational bot may be trained through reinforcement learning by providing a feedback to the automated conversational bot. In one aspect, the automated conversational bot may be trained by validating an output dialogue against, an input dialogue received from the caller, with an expected response. The output dialogue may be provided by the automated conversational bot based on the bot template and the training data. In one implementation, the automated conversational bot may be trained by theverification module 218. - Exemplary embodiments discussed above may provide certain advantages. Though not required to practice aspects of the disclosure, these advantages may include those provided by the following features.
- Some embodiments enable a system and a method to assist bot developer in designing the possible input utterances for intent detection and intent inputs.
- Some embodiments enable a system and a method to assist in forming human like responses for all possible intent responses including emotions and sentiments.
- Some embodiments enable a system and a method to test and train the bot using a set of raw transcripts and audio from voice archives thereby attaining human like natural language processing including regional slangs and phrases.
- Some embodiments enable a system and a method to reuse of existing contact center call recording archives instead of engaging in research to obtain input combinations thereby saving time in developing the bot with all possible utterances for intent detections.
- Although implementations for methods and systems for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user sources have been described in language specific to structural features and/or methods, it is to be understood that the appended claims are not necessarily limited to the specific features or methods described. Rather, the specific features and methods are disclosed as examples of implementations for optimizing detection of the intent for providing human like responses to the user.
Claims (11)
1. A method for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user characterized by feeding a call recording archive to the automated conversational bot, the method comprising:
building, by a processor, an intent graph storing input dialogues, utterances, and output dialogues associated to an intent, wherein the intent indicates a context of a conversation between a caller and a call center representative, and wherein the intent graph is built by,
feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category,
determining a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts, wherein each intent is associated to at least one category, and
mapping the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent;
feeding, by the processor, training data, comprising the intent graph stored in the graph database, to an automated conversational bot thereby enabling a bot builder to fill a bot template pertaining to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances; and
training, by the processor the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot, wherein the automated conversational bot is trained by,
validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data
thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
2. The method as claimed in claim 1 , wherein each call recording, present in the call recording archive, is fed to the NLP engine upon cleansing each audio file based on one or more filters, wherein the one or more filters comprises voice gender of the caller and the call center representative, language used in the call, the at least one category associated to the intent, and call duration.
3. The method as claimed in claim 1 , wherein the intent graph is built by using a conceptual graph concept.
4. The method as claimed in claim 1 , wherein the one or more NLP entities comprises noun, verbs, Question segment and Answer segment.
5. The method as claimed in claim 1 , wherein the intent graph pertaining to each intent is stored in an intent graph database and wherein the set of parameters is filled in the bot template upon querying the intent graph database by the bot builder.
6. A system for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user characterized by feeding a call recording archive to the automated conversational bot, the system comprising:
a processor and
a memory coupled to the processor wherein the processor is capable of executing a plurality of modules stored in the memory and wherein the plurality of modules comprising:
an analyzer module for building an intent graph storing input dialogues, utterances, and output dialogues associated to an intent, wherein the intent indicates a context of a conversation between a caller and a call center representative, and wherein the intent graph is built by enabling an extraction module to
feed each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category,
determine a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts, wherein each intent is associated to at least one category, and
map the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent;
a builder module for feeding training data, comprising the intent graph stored in the graph database to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances; and
a verification module for training the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot, wherein the automated conversational bot is trained by,
validating an output dialogue against, an input dialogue received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data,
thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
7. The system as claimed in claim 6 , wherein each call recording, present in the call recording archive, is fed to the NLP engine upon cleansing each audio file based on one or more filters, wherein the one or more filters comprises voice gender of the caller and the call center representative, language used in the call, the at least one category associated to the intent, and call duration.
8. The system as claimed in claim 6 , wherein the intent graph is built by using a conceptual graph concept.
9. The system as claimed in claim 6 , wherein the one or more NLP entities comprises noun, verbs, Question segment and Answer segment.
10. The system as claimed in claim 6 , wherein the intent graph pertaining to each intent is stored in an intent graph database and wherein the set of parameters is filled in the bot template upon querying the intent graph database by the bot builder.
11. A non-transitory computer readable medium embodying a program executable in a computing device for optimizing detection of an intent, pertaining to a query, by an automated conversational bot for providing human like responses to a user characterized by feeding a call recording archive to the automated conversational bot, the program comprising a program code:
a program code for building an intent graph storing input dialogues, utterances, and output dialogues associated to an intent, wherein the intent indicates a context of a conversation between a caller and a call center representative, and wherein the intent graph is built by,
feeding each audio file indicating a call recording, present in a call recording archive, to a Natural Language Processing (NLP) engine in order to create a set of raw text transcripts pertaining to a category,
determining a plurality of intents from the set of raw text transcripts upon identifying one or more NLP entities from words present in the set of raw text transcripts, wherein each intent is associated to at least one category, and
mapping the input dialogues, the utterances, and the output dialogues with each intent, of the plurality of intents thereby building the intent graph pertaining to each intent;
a program code for feeding training data, comprising the intent graph stored in the graph database to an automated conversational bot by enabling a bot builder to fill a bot template associated to each intent with a set of parameters indicating distinct utterances of an intent and output dialogues associated to the distinct utterances; and
a program code for training the automated conversational bot through reinforcement learning by providing a feedback to the automated conversational bot, wherein the automated conversational bot is trained by,
validating an output dialogue against an input dialogue, received from the caller, with an expected response, wherein the output dialogue is provided by the automated conversational bot based on the bot template and the training data,
thereby optimizing detection of the intent of a query by the automated conversational bot for providing human like responses to the user based on the call recording archive.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IN201811034198 | 2018-09-11 | ||
IN201811034198 | 2018-09-11 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20200081939A1 true US20200081939A1 (en) | 2020-03-12 |
Family
ID=69720841
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/258,680 Abandoned US20200081939A1 (en) | 2018-09-11 | 2019-01-28 | System for optimizing detection of intent[s] by automated conversational bot[s] for providing human like responses |
Country Status (1)
Country | Link |
---|---|
US (1) | US20200081939A1 (en) |
Cited By (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10747957B2 (en) * | 2018-11-13 | 2020-08-18 | Asapp, Inc. | Processing communications using a prototype classifier |
CN111611360A (en) * | 2020-03-31 | 2020-09-01 | 北京文思海辉金信软件有限公司 | Multi-round conversation strategy generation method and device, computer equipment and storage medium |
US10818293B1 (en) | 2020-07-14 | 2020-10-27 | Drift.com, Inc. | Selecting a response in a multi-turn interaction between a user and a conversational bot |
CN111859985A (en) * | 2020-07-23 | 2020-10-30 | 平安普惠企业管理有限公司 | AI customer service model testing method, device, electronic equipment and storage medium |
CN111881251A (en) * | 2020-07-23 | 2020-11-03 | 平安科技(深圳)有限公司 | AI telephone sales test method, device, electronic equipment and storage medium |
US10841251B1 (en) * | 2020-02-11 | 2020-11-17 | Moveworks, Inc. | Multi-domain chatbot |
US20200380076A1 (en) * | 2019-05-30 | 2020-12-03 | Microsoft Technology Licensing, Llc | Contextual feedback to a natural understanding system in a chat bot using a knowledge model |
US11057320B2 (en) * | 2019-06-27 | 2021-07-06 | Walmart Apollo, Llc | Operation for multiple chat bots operation in organization |
US11074484B2 (en) * | 2019-01-31 | 2021-07-27 | International Business Machines Corporation | Self-improving transferring in bot conversation |
CN113360618A (en) * | 2021-06-07 | 2021-09-07 | 暨南大学 | Intelligent robot dialogue method and system based on offline reinforcement learning |
US11113475B2 (en) * | 2019-04-15 | 2021-09-07 | Accenture Global Solutions Limited | Chatbot generator platform |
US11122165B1 (en) * | 2020-05-29 | 2021-09-14 | Verizon Patent And Licensing Inc. | Systems and methods for customer service agent-guided chat session digital assistant |
US11184298B2 (en) * | 2019-08-28 | 2021-11-23 | International Business Machines Corporation | Methods and systems for improving chatbot intent training by correlating user feedback provided subsequent to a failed response to an initial user intent |
CN113806499A (en) * | 2020-12-30 | 2021-12-17 | 京东科技控股股份有限公司 | Telephone work training method and device, electronic equipment and storage medium |
US11206190B1 (en) | 2021-02-01 | 2021-12-21 | International Business Machines Corporation | Using an artificial intelligence based system to guide user dialogs in designing computing system architectures |
US11228681B1 (en) * | 2020-07-29 | 2022-01-18 | Five9, Inc. | Systems for summarizing contact center calls and methods of using same |
US20220103689A1 (en) * | 2019-12-31 | 2022-03-31 | InContact Inc. | Enhancing agent's efficiency in a contact center by using a multi-agent to multi-contact routing orchestration |
US11386259B2 (en) | 2018-04-27 | 2022-07-12 | Asapp, Inc. | Removing personal information from text using multiple levels of redaction |
US11392773B1 (en) * | 2019-01-31 | 2022-07-19 | Amazon Technologies, Inc. | Goal-oriented conversational training data generation |
US11423232B2 (en) * | 2020-03-17 | 2022-08-23 | NFL Enterprises LLC | Systems and methods for deploying computerized conversational agents |
US11477140B2 (en) * | 2019-05-30 | 2022-10-18 | Microsoft Technology Licensing, Llc | Contextual feedback to a natural understanding system in a chat bot |
US11575624B2 (en) | 2019-05-30 | 2023-02-07 | Microsoft Technology Licensing, Llc | Contextual feedback, with expiration indicator, to a natural understanding system in a chat bot |
US11615422B2 (en) | 2016-07-08 | 2023-03-28 | Asapp, Inc. | Automatically suggesting completions of text |
US11706168B2 (en) * | 2017-09-27 | 2023-07-18 | Salesforce, Inc. | Triggering event identification and application dialog validation |
US11710476B2 (en) * | 2020-04-27 | 2023-07-25 | Robert Bosch Gmbh | System and method for automatic testing of conversational assistance |
US11734089B2 (en) | 2021-10-11 | 2023-08-22 | Fmr Llc | Dynamic option reselection in virtual assistant communication sessions |
US11743105B2 (en) * | 2019-06-03 | 2023-08-29 | Hewlett Packard Enterprise Development Lp | Extracting and tagging text about networking entities from human readable textual data sources and using tagged text to build graph of nodes including networking entities |
US11763097B1 (en) | 2022-08-02 | 2023-09-19 | Fmr Llc | Intelligent dialogue recovery for virtual assistant communication sessions |
WO2023178979A1 (en) * | 2022-03-23 | 2023-09-28 | 康键信息技术(深圳)有限公司 | Question labeling method and apparatus, electronic device and storage medium |
US11966705B2 (en) * | 2019-11-13 | 2024-04-23 | Fmr Llc | Systems and methods for bot dialog delegation |
US20240135110A1 (en) * | 2021-05-24 | 2024-04-25 | Narrative Science Inc. | Applied Artificial Intelligence Technology for Natural Language Generation Using a Graph Data Structure and Configurable Chooser Code |
Citations (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020198714A1 (en) * | 2001-06-26 | 2002-12-26 | Guojun Zhou | Statistical spoken dialog system |
US20040006471A1 (en) * | 2001-07-03 | 2004-01-08 | Leo Chiu | Method and apparatus for preprocessing text-to-speech files in a voice XML application distribution system using industry specific, social and regional expression rules |
US20040153649A1 (en) * | 1995-07-27 | 2004-08-05 | Rhoads Geoffrey B. | Digital authentication with digital and analog documents |
US20050028085A1 (en) * | 2001-05-04 | 2005-02-03 | Irwin James S. | Dynamic generation of voice application information from a web server |
US20080201365A1 (en) * | 2007-02-15 | 2008-08-21 | John Edward Petri | Method for multicontext xml fragment reuse and validation in a content management system |
US7447636B1 (en) * | 2005-05-12 | 2008-11-04 | Verizon Corporate Services Group Inc. | System and methods for using transcripts to train an automated directory assistance service |
US20120089394A1 (en) * | 2010-10-06 | 2012-04-12 | Virtuoz Sa | Visual Display of Semantic Information |
US20130332162A1 (en) * | 2012-06-08 | 2013-12-12 | Apple Inc. | Systems and Methods for Recognizing Textual Identifiers Within a Plurality of Words |
US20140288932A1 (en) * | 2011-01-05 | 2014-09-25 | Interactions Corporation | Automated Speech Recognition Proxy System for Natural Language Understanding |
US9092802B1 (en) * | 2011-08-15 | 2015-07-28 | Ramakrishna Akella | Statistical machine learning and business process models systems and methods |
US20150242975A1 (en) * | 2014-02-24 | 2015-08-27 | Mindojo Ltd. | Self-construction of content in adaptive e-learning datagraph structures |
US9190055B1 (en) * | 2013-03-14 | 2015-11-17 | Amazon Technologies, Inc. | Named entity recognition with personalized models |
US9275042B2 (en) * | 2010-03-26 | 2016-03-01 | Virtuoz Sa | Semantic clustering and user interfaces |
US20160283463A1 (en) * | 2015-03-26 | 2016-09-29 | Tata Consultancy Services Limited | Context based conversation system |
US20160364476A1 (en) * | 2015-06-11 | 2016-12-15 | Nuance Communications, Inc. | Systems and methods for learning semantic patterns from textual data |
US20170330077A1 (en) * | 2016-05-13 | 2017-11-16 | Microsoft Technology Licensing, Llc | Deep learning of bots through examples and experience |
US20170359393A1 (en) * | 2016-06-14 | 2017-12-14 | Wipro Limited | System and Method for Building Contextual Highlights for Conferencing Systems |
US20170359463A1 (en) * | 2016-06-13 | 2017-12-14 | Google Inc. | Dynamic initiation of automated call |
US20170357638A1 (en) * | 2016-06-14 | 2017-12-14 | Microsoft Technology Licensing, Llc | Computer messaging bot creation |
US20180011868A1 (en) * | 2016-07-08 | 2018-01-11 | International Business Machines Corporation | Automatically Reorganize Folder/File Visualizations Based on Natural Language-Derived Intent |
US20180048682A1 (en) * | 2016-08-12 | 2018-02-15 | International Business Machines Corporation | Notification bot for topics of interest on voice communication devices |
US9922650B1 (en) * | 2013-12-20 | 2018-03-20 | Amazon Technologies, Inc. | Intent-specific automatic speech recognition result generation |
US20180165691A1 (en) * | 2016-12-09 | 2018-06-14 | Nuance Communications, Inc. | Learning and automating agent actions |
US20180210874A1 (en) * | 2017-01-25 | 2018-07-26 | Google Llc | Automatic suggested responses to images received in messages using language model |
US20180214061A1 (en) * | 2014-08-22 | 2018-08-02 | Sri International | Systems for speech-based assessment of a patient's state-of-mind |
US20190043106A1 (en) * | 2017-08-01 | 2019-02-07 | Facebook, Inc. | Training a chatbot for a digital advertisement to simulate common conversations associated with similar digital advertisements |
US10339216B2 (en) * | 2013-07-26 | 2019-07-02 | Nuance Communications, Inc. | Method and apparatus for selecting among competing models in a tool for building natural language understanding models |
US20190221206A1 (en) * | 2019-03-27 | 2019-07-18 | Intel Corporation | Spoken keyword detection based utterance-level wake on intent system |
US20190294676A1 (en) * | 2018-03-23 | 2019-09-26 | Servicenow, Inc. | Written-modality prosody subsystem in a natural language understanding (nlu) framework |
US20190354598A1 (en) * | 2018-05-18 | 2019-11-21 | Open Text Corporation | Data Processing System for Automatic Presetting of Controls in an Evaluation Operator Interface |
US20190370629A1 (en) * | 2018-05-31 | 2019-12-05 | International Business Machines Corporation | Building a gossip group of domain-specific chatbots |
US20200007493A1 (en) * | 2018-06-28 | 2020-01-02 | Paypal, Inc. | Mid-tier messaging system |
US20200074052A1 (en) * | 2018-08-28 | 2020-03-05 | International Business Machines Corporation | Intelligent user identification |
US10691897B1 (en) * | 2019-08-29 | 2020-06-23 | Accenture Global Solutions Limited | Artificial intelligence based virtual agent trainer |
-
2019
- 2019-01-28 US US16/258,680 patent/US20200081939A1/en not_active Abandoned
Patent Citations (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040153649A1 (en) * | 1995-07-27 | 2004-08-05 | Rhoads Geoffrey B. | Digital authentication with digital and analog documents |
US20050028085A1 (en) * | 2001-05-04 | 2005-02-03 | Irwin James S. | Dynamic generation of voice application information from a web server |
US20020198714A1 (en) * | 2001-06-26 | 2002-12-26 | Guojun Zhou | Statistical spoken dialog system |
US20040006471A1 (en) * | 2001-07-03 | 2004-01-08 | Leo Chiu | Method and apparatus for preprocessing text-to-speech files in a voice XML application distribution system using industry specific, social and regional expression rules |
US7447636B1 (en) * | 2005-05-12 | 2008-11-04 | Verizon Corporate Services Group Inc. | System and methods for using transcripts to train an automated directory assistance service |
US20080201365A1 (en) * | 2007-02-15 | 2008-08-21 | John Edward Petri | Method for multicontext xml fragment reuse and validation in a content management system |
US9275042B2 (en) * | 2010-03-26 | 2016-03-01 | Virtuoz Sa | Semantic clustering and user interfaces |
US20120089394A1 (en) * | 2010-10-06 | 2012-04-12 | Virtuoz Sa | Visual Display of Semantic Information |
US20140288932A1 (en) * | 2011-01-05 | 2014-09-25 | Interactions Corporation | Automated Speech Recognition Proxy System for Natural Language Understanding |
US9092802B1 (en) * | 2011-08-15 | 2015-07-28 | Ramakrishna Akella | Statistical machine learning and business process models systems and methods |
US20130332162A1 (en) * | 2012-06-08 | 2013-12-12 | Apple Inc. | Systems and Methods for Recognizing Textual Identifiers Within a Plurality of Words |
US9190055B1 (en) * | 2013-03-14 | 2015-11-17 | Amazon Technologies, Inc. | Named entity recognition with personalized models |
US10339216B2 (en) * | 2013-07-26 | 2019-07-02 | Nuance Communications, Inc. | Method and apparatus for selecting among competing models in a tool for building natural language understanding models |
US9922650B1 (en) * | 2013-12-20 | 2018-03-20 | Amazon Technologies, Inc. | Intent-specific automatic speech recognition result generation |
US20150242975A1 (en) * | 2014-02-24 | 2015-08-27 | Mindojo Ltd. | Self-construction of content in adaptive e-learning datagraph structures |
US20180214061A1 (en) * | 2014-08-22 | 2018-08-02 | Sri International | Systems for speech-based assessment of a patient's state-of-mind |
US20160283463A1 (en) * | 2015-03-26 | 2016-09-29 | Tata Consultancy Services Limited | Context based conversation system |
US20160364476A1 (en) * | 2015-06-11 | 2016-12-15 | Nuance Communications, Inc. | Systems and methods for learning semantic patterns from textual data |
US20170330077A1 (en) * | 2016-05-13 | 2017-11-16 | Microsoft Technology Licensing, Llc | Deep learning of bots through examples and experience |
US20170359463A1 (en) * | 2016-06-13 | 2017-12-14 | Google Inc. | Dynamic initiation of automated call |
US20170357638A1 (en) * | 2016-06-14 | 2017-12-14 | Microsoft Technology Licensing, Llc | Computer messaging bot creation |
US20170359393A1 (en) * | 2016-06-14 | 2017-12-14 | Wipro Limited | System and Method for Building Contextual Highlights for Conferencing Systems |
US20180011868A1 (en) * | 2016-07-08 | 2018-01-11 | International Business Machines Corporation | Automatically Reorganize Folder/File Visualizations Based on Natural Language-Derived Intent |
US20180048682A1 (en) * | 2016-08-12 | 2018-02-15 | International Business Machines Corporation | Notification bot for topics of interest on voice communication devices |
US20180165691A1 (en) * | 2016-12-09 | 2018-06-14 | Nuance Communications, Inc. | Learning and automating agent actions |
US20180210874A1 (en) * | 2017-01-25 | 2018-07-26 | Google Llc | Automatic suggested responses to images received in messages using language model |
US20190043106A1 (en) * | 2017-08-01 | 2019-02-07 | Facebook, Inc. | Training a chatbot for a digital advertisement to simulate common conversations associated with similar digital advertisements |
US20190294676A1 (en) * | 2018-03-23 | 2019-09-26 | Servicenow, Inc. | Written-modality prosody subsystem in a natural language understanding (nlu) framework |
US20190354598A1 (en) * | 2018-05-18 | 2019-11-21 | Open Text Corporation | Data Processing System for Automatic Presetting of Controls in an Evaluation Operator Interface |
US20190370629A1 (en) * | 2018-05-31 | 2019-12-05 | International Business Machines Corporation | Building a gossip group of domain-specific chatbots |
US20200007493A1 (en) * | 2018-06-28 | 2020-01-02 | Paypal, Inc. | Mid-tier messaging system |
US20200074052A1 (en) * | 2018-08-28 | 2020-03-05 | International Business Machines Corporation | Intelligent user identification |
US20190221206A1 (en) * | 2019-03-27 | 2019-07-18 | Intel Corporation | Spoken keyword detection based utterance-level wake on intent system |
US10691897B1 (en) * | 2019-08-29 | 2020-06-23 | Accenture Global Solutions Limited | Artificial intelligence based virtual agent trainer |
Cited By (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US12039545B2 (en) | 2016-07-08 | 2024-07-16 | Asapp, Inc. | Third-party service for suggesting a response to a received message |
US11615422B2 (en) | 2016-07-08 | 2023-03-28 | Asapp, Inc. | Automatically suggesting completions of text |
US11706168B2 (en) * | 2017-09-27 | 2023-07-18 | Salesforce, Inc. | Triggering event identification and application dialog validation |
US11386259B2 (en) | 2018-04-27 | 2022-07-12 | Asapp, Inc. | Removing personal information from text using multiple levels of redaction |
US10747957B2 (en) * | 2018-11-13 | 2020-08-18 | Asapp, Inc. | Processing communications using a prototype classifier |
US11074484B2 (en) * | 2019-01-31 | 2021-07-27 | International Business Machines Corporation | Self-improving transferring in bot conversation |
US11392773B1 (en) * | 2019-01-31 | 2022-07-19 | Amazon Technologies, Inc. | Goal-oriented conversational training data generation |
US11113475B2 (en) * | 2019-04-15 | 2021-09-07 | Accenture Global Solutions Limited | Chatbot generator platform |
US11575624B2 (en) | 2019-05-30 | 2023-02-07 | Microsoft Technology Licensing, Llc | Contextual feedback, with expiration indicator, to a natural understanding system in a chat bot |
US20200380076A1 (en) * | 2019-05-30 | 2020-12-03 | Microsoft Technology Licensing, Llc | Contextual feedback to a natural understanding system in a chat bot using a knowledge model |
US11477140B2 (en) * | 2019-05-30 | 2022-10-18 | Microsoft Technology Licensing, Llc | Contextual feedback to a natural understanding system in a chat bot |
US11743105B2 (en) * | 2019-06-03 | 2023-08-29 | Hewlett Packard Enterprise Development Lp | Extracting and tagging text about networking entities from human readable textual data sources and using tagged text to build graph of nodes including networking entities |
US11057320B2 (en) * | 2019-06-27 | 2021-07-06 | Walmart Apollo, Llc | Operation for multiple chat bots operation in organization |
US11184298B2 (en) * | 2019-08-28 | 2021-11-23 | International Business Machines Corporation | Methods and systems for improving chatbot intent training by correlating user feedback provided subsequent to a failed response to an initial user intent |
US11966705B2 (en) * | 2019-11-13 | 2024-04-23 | Fmr Llc | Systems and methods for bot dialog delegation |
US11659092B2 (en) * | 2019-12-31 | 2023-05-23 | InContact Inc. | Enhancing agent's efficiency in a contact center by using a multi-agent to multi-contact routing orchestration |
US20220103689A1 (en) * | 2019-12-31 | 2022-03-31 | InContact Inc. | Enhancing agent's efficiency in a contact center by using a multi-agent to multi-contact routing orchestration |
US11303753B2 (en) * | 2019-12-31 | 2022-04-12 | InContact Inc. | Enhancing agent's efficiency in a contact center by using a multi-agent to multi-contact routing orchestration |
US10841251B1 (en) * | 2020-02-11 | 2020-11-17 | Moveworks, Inc. | Multi-domain chatbot |
US11423232B2 (en) * | 2020-03-17 | 2022-08-23 | NFL Enterprises LLC | Systems and methods for deploying computerized conversational agents |
CN111611360A (en) * | 2020-03-31 | 2020-09-01 | 北京文思海辉金信软件有限公司 | Multi-round conversation strategy generation method and device, computer equipment and storage medium |
US11710476B2 (en) * | 2020-04-27 | 2023-07-25 | Robert Bosch Gmbh | System and method for automatic testing of conversational assistance |
US11647117B2 (en) | 2020-05-29 | 2023-05-09 | Verizon Patent And Licensing Inc. | Systems and methods for customer service agent-guided chat session digital assistant |
US11122165B1 (en) * | 2020-05-29 | 2021-09-14 | Verizon Patent And Licensing Inc. | Systems and methods for customer service agent-guided chat session digital assistant |
US10818293B1 (en) | 2020-07-14 | 2020-10-27 | Drift.com, Inc. | Selecting a response in a multi-turn interaction between a user and a conversational bot |
CN111881251A (en) * | 2020-07-23 | 2020-11-03 | 平安科技(深圳)有限公司 | AI telephone sales test method, device, electronic equipment and storage medium |
CN111859985A (en) * | 2020-07-23 | 2020-10-30 | 平安普惠企业管理有限公司 | AI customer service model testing method, device, electronic equipment and storage medium |
US11228681B1 (en) * | 2020-07-29 | 2022-01-18 | Five9, Inc. | Systems for summarizing contact center calls and methods of using same |
CN113806499A (en) * | 2020-12-30 | 2021-12-17 | 京东科技控股股份有限公司 | Telephone work training method and device, electronic equipment and storage medium |
US11206190B1 (en) | 2021-02-01 | 2021-12-21 | International Business Machines Corporation | Using an artificial intelligence based system to guide user dialogs in designing computing system architectures |
US20240135110A1 (en) * | 2021-05-24 | 2024-04-25 | Narrative Science Inc. | Applied Artificial Intelligence Technology for Natural Language Generation Using a Graph Data Structure and Configurable Chooser Code |
US12248461B2 (en) * | 2021-05-24 | 2025-03-11 | Salesforce, Inc. | Applied artificial intelligence technology for natural language generation using a graph data structure and configurable chooser code |
CN113360618A (en) * | 2021-06-07 | 2021-09-07 | 暨南大学 | Intelligent robot dialogue method and system based on offline reinforcement learning |
US11734089B2 (en) | 2021-10-11 | 2023-08-22 | Fmr Llc | Dynamic option reselection in virtual assistant communication sessions |
WO2023178979A1 (en) * | 2022-03-23 | 2023-09-28 | 康键信息技术(深圳)有限公司 | Question labeling method and apparatus, electronic device and storage medium |
US11763097B1 (en) | 2022-08-02 | 2023-09-19 | Fmr Llc | Intelligent dialogue recovery for virtual assistant communication sessions |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20200081939A1 (en) | System for optimizing detection of intent[s] by automated conversational bot[s] for providing human like responses | |
US10679631B2 (en) | Generating chat bots from web API specifications | |
US10431205B2 (en) | Dialog device with dialog support generated using a mixture of language models combined using a recurrent neural network | |
US7647228B2 (en) | Method and apparatus for speech processing incorporating user intent | |
CN113228606A (en) | Semantic CRM copies from mobile communication sessions | |
US9424253B2 (en) | Domain specific natural language normalization | |
CN113039537A (en) | Semantic artificial intelligence agent | |
CN105592343A (en) | Display Apparatus And Method For Question And Answer | |
JP2008512789A (en) | Machine learning | |
WO2004072926A2 (en) | Management of conversations | |
US20180255180A1 (en) | Bridge for Non-Voice Communications User Interface to Voice-Enabled Interactive Voice Response System | |
Kopparapu | Non-linguistic analysis of call center conversations | |
JP6625772B2 (en) | Search method and electronic device using the same | |
US20230244878A1 (en) | Extracting conversational relationships based on speaker prediction and trigger word prediction | |
US12130847B2 (en) | Methods and systems for ambiguity resolution in conversations managed by a virtual assistant server | |
US20230274093A1 (en) | Method and system for performance evaluation using an ai model | |
CN113168639A (en) | Semantic CRM mobile communication session | |
CN115033677A (en) | Event processing method, device, equipment and medium based on conversation robot | |
KR102147619B1 (en) | Method for managing phone calls and artificial intelligence assistant system for performing the same | |
KR102610360B1 (en) | Method for providing labeling for spoken voices, and apparatus implementing the same method | |
US20200204679A1 (en) | System for processing voice responses using a natural language processing engine | |
Chung et al. | A question detection algorithm for text analysis | |
CN114462376A (en) | RPA and AI-based court trial record generation method, device, equipment and medium | |
CN113744712A (en) | Intelligent outbound voice splicing method, device, equipment, medium and program product | |
Kanniappan et al. | Ai-based automated approach for trend data generation and competitor benchmark to enhance voice ai services |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HCL TECHNOLOGIES LIMITED, INDIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SUBRAMANIAM, SENTHIL KUMAR;REEL/FRAME:048234/0863 Effective date: 20190124 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |