CN107919123B - Multi-voice assistant control method, device and computer readable storage medium - Google Patents
Multi-voice assistant control method, device and computer readable storage medium Download PDFInfo
- Publication number
- CN107919123B CN107919123B CN201711287860.8A CN201711287860A CN107919123B CN 107919123 B CN107919123 B CN 107919123B CN 201711287860 A CN201711287860 A CN 201711287860A CN 107919123 B CN107919123 B CN 107919123B
- Authority
- CN
- China
- Prior art keywords
- voice
- voice assistant
- assistant
- target
- task
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Program initiating; Program switching, e.g. by interrupt
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The disclosure relates to a multi-voice assistant control method, a device and a computer readable storage medium, which are used for solving the technical problems that when the voice assistant is switched by the existing electronic equipment, the voice assistant needs to be manually enabled, and the switching mode is not flexible enough. The multi-voice assistant control method is applied to electronic equipment integrating a plurality of voice assistants, and comprises the following steps: when a trigger event for waking up the voice assistant is detected, determining a target voice assistant from a plurality of voice assistants according to the trigger event, wherein the waking functions of the plurality of voice assistants are all in an enabling state; the target voice assistant is awakened.
Description
Technical Field
The present disclosure relates to the field of communications technologies, and in particular, to a multi-voice assistant control method, apparatus, and computer-readable storage medium.
Background
With the development of speech recognition technology and the increasing demand of people for speech interaction, an intelligent application of speech assistant is produced. In the related art, the voice assistant may implement an intelligent dialog or an instant question and answer with the user to help the user solve a problem. However, the conventional electronic device usually only has one voice assistant built therein, even if a plurality of voice assistants are built therein, only one of the voice assistants is in a wakenable state at the same time, the plurality of voice assistants cannot be turned on simultaneously, and the voice assistant cannot be wakened up if the voice assistant is not turned on.
Disclosure of Invention
The present disclosure provides a multi-voice assistant control method, an apparatus and a computer-readable storage medium, which are used to solve the technical problem that when an existing electronic device switches a voice assistant, the voice assistant needs to be manually enabled, and the switching manner is not flexible enough.
In order to achieve the above object, a first aspect of the embodiments of the present disclosure provides a multi-voice assistant control method applied to an electronic device integrating multiple voice assistants, the method including:
when a trigger event for waking up the voice assistant is detected, determining a target voice assistant from a plurality of voice assistants according to the trigger event, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
waking the target voice assistant.
Optionally, before the determining a target voice assistant from the plurality of voice assistants according to the trigger event, the method includes:
collecting input first voice information;
judging whether the first voice message comprises a keyword which is consistent with a preset awakening word or not;
and if the first voice message comprises a keyword which is consistent with the preset awakening word, determining that the trigger event occurs.
Optionally, the triggering event includes collecting second voice information for waking up a voice assistant, and the determining a target voice assistant from the plurality of voice assistants according to the triggering event includes:
determining whether keywords associated with the identity of any of the voice assistants are included in the second voice message;
and if the second voice message comprises the keywords associated with the identification of the first voice assistant, determining that the first voice assistant is the target voice assistant.
Optionally, before the determining a target voice assistant from the plurality of voice assistants according to the trigger event, the method includes:
acquiring a first voice task input by a user through a second voice assistant awakened by the electronic equipment at present;
determining that the triggering event occurs when the first voice task instructs the second voice assistant to wake up the first voice assistant.
Optionally, the method further comprises:
acquiring a second voice task input by a user through a first voice assistant which is awakened currently by the electronic equipment, wherein the second voice task comprises target task information indicating the execution of the second voice assistant;
calling an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
presenting the response information by the first voice assistant.
In a second aspect of the embodiments of the present disclosure, a multi-voice assistant control apparatus is provided, where the apparatus is applied to an electronic device integrating multiple voice assistants, and includes:
the voice assistant determination module is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake-up module configured to wake up the target voice assistant.
Optionally, the apparatus comprises:
a voice information collection module configured to collect input first voice information before the target voice assistant is determined from the plurality of voice assistants according to the trigger event;
the keyword judgment module is configured to judge whether the first voice message comprises a keyword which is consistent with a preset awakening word;
a trigger event determination module configured to determine that the trigger event occurs if the first voice message includes a keyword consistent with the preset wake-up word.
Optionally, the triggering event includes acquiring second voice information for waking up a voice assistant, and the voice assistant determining module includes:
a keyword judgment sub-module configured to judge whether a keyword associated with any of the voice assistant identifiers is included in the second voice message;
and the voice assistant determining sub-module is configured to determine that the first voice assistant is the target voice assistant if the second voice message comprises a keyword associated with the identifier of the first voice assistant.
Optionally, the apparatus comprises:
a first voice task obtaining module configured to obtain a first voice task input by a user through a second voice assistant which is already woken up currently by the electronic equipment before the target voice assistant is determined from the plurality of voice assistants according to the trigger event;
a trigger event determination module configured to determine that the trigger event occurs when the first voice task instructs the second voice assistant to wake up the first voice assistant.
Optionally, the apparatus further comprises:
the second voice task acquisition module is configured to acquire a second voice task input by a user through a first voice assistant which is awakened currently by the electronic equipment, and the second voice task comprises target task information indicating execution of the second voice assistant;
a response information acquisition module configured to call an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
a response information presentation module configured to present the response information by the first voice assistant.
In a third aspect of the embodiments of the present disclosure, a multi-voice assistant control apparatus is provided, where the apparatus is applied to an electronic device integrating multiple voice assistants, and the apparatus includes:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to:
when a trigger event for waking up a voice assistant is detected, determining a target voice assistant from a plurality of voice assistants according to the trigger event, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
waking the target voice assistant.
In a fourth aspect of the embodiments of the present disclosure, a computer-readable storage medium is provided, on which computer program instructions are stored, which when executed by a processor, implement the steps of the multi-voice assistant control method provided in the first aspect of the embodiments of the present disclosure.
The technical scheme provided by the embodiment of the disclosure can have the following beneficial effects:
by the technical scheme, for the electronic equipment integrated with the voice assistants, when a trigger event for waking up the voice assistants is detected, the target voice assistant can be determined from the voice assistants of which the wake-up functions are all in the enabled state according to the trigger event, and the target voice assistant is woken up. Therefore, each voice assistant awakening function is kept in an enabled state, different voice assistants can be awakened directly through different trigger events, and the technical problem that when the voice assistants are switched by the existing electronic equipment, the voice assistants need to be manually enabled, and the switching mode is not flexible enough is solved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and together with the description, serve to explain the principles of the disclosure.
FIG. 1 is a flow diagram illustrating a multi-voice assistant control method in accordance with an exemplary embodiment.
FIG. 2 is a flowchart illustrating a multi-voice assistant control method according to an example embodiment.
FIG. 3 is a flowchart illustrating a multi-voice assistant control method according to an example embodiment.
FIG. 4 is a flow diagram illustrating a multi-voice assistant control method in accordance with an exemplary embodiment.
FIG. 5 is a flow diagram illustrating a multi-voice assistant control method in accordance with an exemplary embodiment.
FIG. 6 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 7 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 8 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 9 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 10 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
FIG. 11 is a block diagram illustrating a multi-voice assistant control apparatus according to an example embodiment.
Detailed Description
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
FIG. 1 is a flow chart illustrating a multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, where the method includes the following steps:
The triggering event includes voice information for waking up the voice assistant, a key operation or a touch screen operation, where the touch screen operation may be a user performing a click operation or a sliding operation in a designated direction based on a designated area of a touch screen of the electronic device, for example, double-clicking a menu key or drawing a preset wake-up gesture path, and the electronic device determines an operation object, that is, a target voice assistant, according to a preset corresponding relationship between specific content of the triggering event and the voice assistant. After determining the target voice assistant, the electronic device may wake up the target voice assistant by turning off the screen or in a customized manner.
It should be noted that, a plurality of voice assistants in the electronic device are all in an enabled state, and can be awakened accordingly in case of a trigger event, and the plurality of voice assistants can coexist simultaneously.
By adopting the method, each voice assistant awakening function keeps the enabled state, different voice assistants can be awakened directly through different trigger events, and the technical problem that the voice assistants are required to be manually enabled when the existing electronic equipment switches the voice assistants, and the switching mode is not flexible enough is solved.
FIG. 2 is a flowchart illustrating a multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, as shown in FIG. 2, the method comprising the steps of:
In one possible implementation, the triggering event may be the acquisition of voice information by the electronic device to wake up a voice assistant. For example, the electronic device may collect voice information from the surrounding environment in real time through a recording function, input the collected voice information as first voice information into the system for recognition, and perform step 202 on the recognition result.
In step 202, the recognition result of the first voice message obtained in step 201 may be determined to determine whether the result includes a keyword consistent with the preset wake-up word. For example, the keyword of the preset wakeup word may be a name of each voice assistant, or may be a word or a sentence customized by the user according to actual needs. For example, the electronic device includes three voice assistants, wherein the name of voice assistant a is small a, the name of voice assistant B is small B, the name of voice assistant C is small C, and if the recognition result of the first voice message includes a keyword of small a, small B, or small C, the determination result in step 202 is yes. If the preset awakening words customized by the user are words such as "hello" and "out bar", the determination result in the step 202 may be determined to be "yes" when the recognition result of the first voice message includes the words.
It should be noted that the preset wake-up word customized by the user may or may not specifically specify the voice assistant. For example, if a preset corresponding relationship exists between a preset wake-up word and a voice assistant, for example, the preset wake-up word "wake-up voice assistant" corresponds to the voice assistant a, "hello" corresponds to the voice assistant B, "out-of-the-bar" corresponds to the voice assistant C, it indicates that the preset wake-up word of the user can specifically specify which voice assistant to wake up, and the electronic device can determine the target voice assistant by performing keyword judgment once; if the preset corresponding relation does not exist between the preset awakening words and the voice assistant, the preset awakening words 'awakening the voice assistant' and 'hello' defined by the user do not specifically specify which voice assistant to awaken, and the system can directly awaken the default voice assistant or prompt the user to select which voice assistant to awaken again.
In the case that the determination result in step 202 is "yes", it is determined that a trigger event occurs, and the target voice assistant determined according to the content of the trigger event is awakened.
By adopting the method, whether the acquired voice information comprises the keyword which is consistent with the preset awakening word or not can be judged, and the target voice assistant is awakened correspondingly according to the judgment result. In addition, the plurality of voice assistant awakening functions are kept in the enabled state, different voice assistants can be directly awakened through different trigger events, preset awakening words can be defined by a user, the use method of the voice assistants is enriched, the single role of the voice assistants is avoided, and the technical problem that when the voice assistants are switched by the existing electronic equipment, the voice assistants need to be manually enabled, and the switching mode is not flexible is solved.
FIG. 3 is a flowchart illustrating a multi-voice assistant control method according to an exemplary embodiment, applied to an electronic device integrating multiple voice assistants, as shown in FIG. 3, the method comprising the steps of:
step 303 wakes up the target voice assistant.
In another possible implementation, when the first voice message collected and input by the electronic device includes a keyword matching a preset wake-up word, but does not specifically specify which voice assistant to wake up, the system may prompt the user to select which voice assistant to wake up again. In this case, the electronic device may collect second voice information for waking up the voice assistant from the surrounding environment through the recording function, input the collected second voice information into the system for recognition, and perform step 301 on the recognition result.
It may be determined whether a keyword associated with an identifier of any of the voice assistants is included in the recognition result of the second voice information through step 301. The identification of the voice assistant may include the name or number of the voice assistant, and may also be a user-defined keyword, where there is a preset correspondence between the keyword and the voice assistant role. For example, the name of the first voice assistant is small a, the number is 1, the user-defined keyword is "intern", and if the collected second voice information includes a word associated with the identifier of the first voice assistant, it may be determined that the first voice assistant is the target voice assistant, and the electronic device performs an operation of waking up the first voice assistant.
By adopting the method, the target voice assistant can be determined again by collecting and judging the keywords in the second voice message under the condition that the specific voice assistant can not be awakened according to the first voice message, so that the technical problem that the existing electronic equipment needs to manually enable the voice assistant when switching the voice assistant and the switching mode is not flexible is solved.
FIG. 4 is a flowchart illustrating a multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, as shown in FIG. 4, the method comprising the steps of:
at step 403, the target voice assistant is awakened.
In the embodiment of the disclosure, the second voice assistant is a woken voice assistant, when the second voice assistant acquires a first voice task input by a user and the task instructs the second voice assistant to wake up the first voice assistant, it is determined that a trigger event occurs, and according to the trigger event, it may be determined that the first voice assistant is a target voice assistant, and a wake-up operation is performed on the first voice assistant. Exemplarily, the voice assistant B is in an awakened state, when a voice task acquired by the voice assistant B includes a preset keyword for awakening the voice assistant a, it is determined that a trigger event occurs, and it is determined that the voice assistant a is the target voice assistant, at this time, the electronic device awakens the voice assistant a according to the voice task, and the voice assistant a performs subsequent interaction with the user.
By adopting the method, when one voice assistant is in an awakening state, other voice assistants can be awakened in a chained manner by releasing the voice task, so that the problem that a user awakens the voice assistant in a single role is avoided, the interestingness of man-machine interaction is improved, and the technical problem that the voice assistant needs to be manually enabled when the existing electronic equipment switches the voice assistants and the switching manner is not flexible enough is solved.
FIG. 5 is a flowchart illustrating a multi-voice assistant control method applied to an electronic device integrating multiple voice assistants according to an exemplary embodiment, as shown in FIG. 5, the method comprising the steps of:
In another possible implementation manner, the voice assistant is awakened through a preset awakening word, the first voice assistant is determined to be the target voice assistant, and after the first voice assistant is awakened, a second voice task input by a user can be acquired through the first voice assistant.
For example, the first voice assistant is small a, the second voice assistant is small B, and in the case that the small voice assistant a is awakened, the second voice task input by the user is "small a, ask how small B hides the Excel cell". At this time, the voice assistant widget A recognizes that the voice task needs to call an application interface of the second voice assistant widget B so as to transmit the query of the user to the second voice assistant widget B, and the processing result of the second voice assistant widget B is displayed on the human-computer interaction interface through the first voice assistant widget A.
By adopting the method, based on the first voice assistant system, other voice assistants can be called to serve the user according to the voice task, and the technical problem that the existing electronic equipment needs to manually enable the voice assistant when switching the voice assistant and the switching mode is not flexible is solved.
Fig. 6 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 6, the apparatus includes:
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake module 620 configured to wake the target voice assistant.
By adopting the device, each voice assistant awakening function keeps the enabled state, and different voice assistants can be directly awakened through the voice assistant determining module 610 according to different trigger events, so that the technical problem that the voice assistants are required to be manually enabled when the existing electronic equipment switches the voice assistants, and the switching mode is not flexible enough is solved.
Fig. 7 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 7, the apparatus includes:
a voice information collection module 630 configured to collect the input first voice information before determining the target voice assistant from the plurality of voice assistants according to the triggering event;
a keyword judgment module 640 configured to judge whether the first voice message includes a keyword consistent with a preset wake-up word;
a trigger event determination module 650 configured to determine that a trigger event occurs if the first voice message includes a keyword corresponding to a preset wakeup word;
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake module 620 configured to wake the target voice assistant.
Optionally, the triggering event includes collecting a second voice message for waking up the voice assistant, and referring to fig. 8, the voice assistant determining module 610 may include:
a keyword determination sub-module 611 configured to determine whether the second voice message includes a keyword associated with an identifier of any voice assistant;
a voice assistant determination sub-module 612 configured to determine that the first voice assistant is the target voice assistant if the second voice message includes a keyword associated with the identifier of the first voice assistant.
By adopting the device, whether the acquired voice information comprises the keyword which is consistent with the preset awakening word or not can be judged, and the target voice assistant is awakened correspondingly according to the judgment result. In addition, the plurality of voice assistant awakening functions are kept in the enabled state, different voice assistants can be directly awakened through different trigger events, preset awakening words can be defined by a user, the use method of the voice assistants is enriched, the single role of the voice assistants is avoided, and the technical problem that when the voice assistants are switched by the existing electronic equipment, the voice assistants need to be manually enabled, and the switching mode is not flexible is solved.
Fig. 9 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 9, the apparatus includes:
a first voice task obtaining module 660 configured to obtain a first voice task input by a user through a second voice assistant that is currently awakened by the electronic device before a target voice assistant is determined from the plurality of voice assistants according to a trigger event;
a trigger event determination module 650 configured to determine that a trigger event occurs when the first voice task instructs the second voice assistant to wake up the first voice assistant;
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake module 620 configured to wake the target voice assistant.
By adopting the device, when one voice assistant is in an awakening state, other voice assistants can be awakened in a chained mode of issuing voice tasks, the condition that a user awakens the voice assistant to have a single interaction role is avoided, the interestingness of man-machine interaction is improved, and the technical problem that the voice assistant is required to be manually enabled when the existing electronic equipment switches the voice assistants, and the switching mode is not flexible enough is solved.
Fig. 10 is a block diagram of a multi-voice assistant control apparatus 600 applied to an electronic device integrating a plurality of voice assistants according to an exemplary embodiment, and referring to fig. 10, the apparatus includes:
the voice assistant determination module 610 is configured to determine a target voice assistant from a plurality of voice assistants according to a trigger event when the trigger event for waking up the voice assistant is detected, wherein the wake-up functions of the plurality of voice assistants are all in an enabled state;
a voice assistant wake-up module 620 configured to wake up the target voice assistant;
a second voice task obtaining module 670 configured to obtain, by the first voice assistant that is currently woken up by the electronic device, a second voice task input by the user, where the second voice task includes target task information indicating that the second voice assistant executes;
a response information obtaining module 680 configured to call an application interface of the second voice assistant through the target task information based on the first voice assistant to obtain response information of the second voice assistant for the target task information;
a response information presentation module 690 configured to present the response information via the first voice assistant.
By adopting the device, based on the first voice assistant system, other voice assistants can be called to serve the user according to the voice task, and the technical problem that the existing electronic equipment needs to manually enable the voice assistant when switching the voice assistant and the switching mode is not flexible is solved.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
The present disclosure also provides a computer-readable storage medium having stored thereon computer program instructions which, when executed by a processor, implement the steps of the multi-voice assistant control method provided by the present disclosure.
FIG. 11 is a block diagram illustrating a multi-voice assistant control apparatus 1100 according to an example embodiment. For example, the apparatus 1100 may be a smartphone, tablet device, personal digital assistant, or the like.
Referring to fig. 11, apparatus 1100 may include one or more of the following components: processing component 1101, memory 1102, power component 1103, multimedia component 1104, audio component 1105, input/output (I/O) interface 1106, sensor component 1107, and communications component 1108.
The processing component 1101 generally controls the overall operation of the device 1100, such as operations associated with display, touch operations, and recording operations. The processing component 1101 may include one or more processors 1109 to execute instructions to perform all or a portion of the steps of the multi-voice assistant control method described above. Further, the processing component 1101 can include one or more modules that facilitate interaction between the processing component 1101 and other components. For example, the processing component 1101 can include a multimedia module to facilitate interaction between the multimedia component 1104 and the processing component 1101.
The memory 1102 is configured to store various types of data to support operation at the device 1100. Examples of such data include instructions for any application or method operating on device 1100, contact data, phonebook data, messages, pictures, videos, and so forth. The memory 1102 may be implemented by any type or combination of volatile or non-volatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks.
The power components 1103 provide power to the various components of the device 1100. The power components 1103 may include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for the apparatus 1100.
The multimedia component 1104 includes a screen that provides an output interface between the device 1100 and the user. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive an input signal from a user. The touch panel includes one or more touch sensors to sense touch, slide, and gestures on the touch panel. The touch sensor may not only sense the boundary of a touch or slide action, but also detect the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 1104 includes a front facing camera and/or a rear facing camera. The front camera and/or the rear camera may receive external multimedia data when the device 1100 is in an operating mode, such as a shooting mode or a video mode. Each front camera and rear camera may be a fixed optical lens system or have a focal length and optical zoom capability.
The I/O interface 1106 provides an interface between the processing component 1101 and peripheral interface modules, which may be keyboards, click wheels, buttons, etc. These buttons may include, but are not limited to: a home button, a volume button, a start button, and a lock button.
The communication component 1108 is configured to facilitate communications between the apparatus 1100 and other devices in a wired or wireless manner. The apparatus 1100 may access a Wireless network based on a communication standard, such as Wi-Fi (Wireless-Fidelity), 2G or 3G, or a combination thereof. In an exemplary embodiment, the communication component 1108 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 1108 also includes a Near Field Communication (NFC) module to facilitate short-range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, Ultra Wideband (UWB) technology, Bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the apparatus 1100 may be implemented by one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Digital Signal Processing Devices (DSPDs), Programmable Logic Devices (PLDs), Field Programmable Gate Arrays (FPGAs), controllers, micro-controllers, microprocessors or other electronic components for performing the multi-voice assistant control method described above.
In an exemplary embodiment, a non-transitory computer-readable storage medium comprising instructions, such as memory 1102 comprising instructions, executable by a processor of apparatus 1100 to perform the multi-voice assistant control method described above is also provided. For example, the non-transitory computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It will be understood that the present disclosure is not limited to the precise arrangements described above and shown in the drawings and that various modifications and changes may be made without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.
Claims (10)
1. A multi-voice assistant control method applied to an electronic device integrating a plurality of voice assistants, the method comprising:
when a trigger event for awakening the voice assistant is detected, determining a first voice assistant as a target voice assistant according to a preset corresponding relation between the trigger event and the voice assistant, wherein the first voice assistant is at least one voice assistant in a plurality of voice assistants of which awakening functions are all in an enabling state;
waking the target voice assistant;
acquiring a second voice task input by a user through the first voice assistant which is awakened currently by the electronic equipment, wherein the second voice task comprises target task information indicating the execution of the second voice assistant;
calling an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
presenting the response information by the first voice assistant.
2. The method of claim 1, wherein prior to the determining the first voice assistant as the target voice assistant based on the predetermined correspondence between the triggering event and the voice assistant, comprising:
collecting input first voice information;
judging whether the first voice message comprises a keyword which is consistent with a preset awakening word or not;
and if the first voice message comprises a keyword which is consistent with the preset awakening word, determining that the trigger event occurs.
3. The method of claim 1, wherein the triggering event includes collecting second voice information for waking up a voice assistant, and wherein determining the first voice assistant as the target voice assistant according to a preset correspondence between the triggering event and the voice assistant comprises:
determining whether keywords associated with the identification of any of the voice assistants are included in the second voice message;
and if the second voice message comprises the keywords associated with the identification of the first voice assistant, determining that the first voice assistant is the target voice assistant.
4. The method of claim 1, wherein prior to the determining the first voice assistant as the target voice assistant based on the predetermined correspondence between the triggering event and the voice assistant, comprising:
acquiring a first voice task input by a user through a third voice assistant awakened by the electronic equipment at present;
determining that the triggering event occurs when the first voice task instructs the third voice assistant to wake up the first voice assistant.
5. A multi-voice assistant control apparatus applied to an electronic device integrating a plurality of voice assistants, the apparatus comprising:
the voice assistant determining module is configured to determine a first voice assistant as a target voice assistant according to a preset corresponding relation between a trigger event and the voice assistant when the trigger event for waking up the voice assistant is detected, wherein the first voice assistant is at least one of a plurality of voice assistants of which the wake-up functions are all in an enabling state;
a voice assistant wake-up module configured to wake up the target voice assistant;
the device further comprises:
a second voice task obtaining module configured to obtain, by the first voice assistant that is currently woken up by the electronic device, a second voice task input by a user, where the second voice task includes target task information indicating that the second voice assistant executes;
a response information acquisition module configured to call an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
a response information presentation module configured to present the response information by the first voice assistant.
6. The apparatus of claim 5, comprising:
a voice information collecting module configured to collect the input first voice information before the first voice assistant is determined as the target voice assistant according to the triggering event;
the keyword judgment module is configured to judge whether the first voice message comprises a keyword which is consistent with a preset awakening word;
a trigger event determination module configured to determine that the trigger event occurs if the first voice message includes a keyword consistent with the preset wake-up word.
7. The apparatus of claim 5, wherein the triggering event comprises a collection of second voice information for waking a voice assistant, and wherein the voice assistant determination module comprises:
a keyword judgment sub-module configured to judge whether a keyword associated with any of the voice assistant identifiers is included in the second voice message;
and the voice assistant determining sub-module is configured to determine that the first voice assistant is the target voice assistant if the second voice message comprises a keyword associated with the identifier of the first voice assistant.
8. The apparatus of claim 5, comprising:
a first voice task obtaining module configured to obtain a first voice task input by a user through a third voice assistant which is already woken up currently by the electronic equipment before the first voice assistant is determined as a target voice assistant according to the trigger event;
a trigger event determination module configured to determine that the trigger event occurs when the first voice task instructs the third voice assistant to wake up the first voice assistant.
9. A multi-voice assistant control apparatus for use with an electronic device that integrates multiple voice assistants, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to:
when a trigger event for awakening the voice assistant is detected, determining a first voice assistant as a target voice assistant according to a preset corresponding relation between the trigger event and the voice assistant, wherein the first voice assistant is at least one voice assistant in a plurality of voice assistants of which awakening functions are all in an enabling state;
waking the target voice assistant;
acquiring a second voice task input by a user through the first voice assistant which is awakened currently by the electronic equipment, wherein the second voice task comprises target task information indicating the execution of the second voice assistant;
calling an application interface of the second voice assistant through the target task information based on the first voice assistant to acquire response information of the second voice assistant for the target task information;
presenting the response information by the first voice assistant.
10. A computer-readable storage medium, on which computer program instructions are stored, which program instructions, when executed by a processor, carry out the steps of the method according to any one of claims 1 to 4.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711287860.8A CN107919123B (en) | 2017-12-07 | 2017-12-07 | Multi-voice assistant control method, device and computer readable storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711287860.8A CN107919123B (en) | 2017-12-07 | 2017-12-07 | Multi-voice assistant control method, device and computer readable storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107919123A CN107919123A (en) | 2018-04-17 |
CN107919123B true CN107919123B (en) | 2022-06-03 |
Family
ID=61898470
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711287860.8A Active CN107919123B (en) | 2017-12-07 | 2017-12-07 | Multi-voice assistant control method, device and computer readable storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107919123B (en) |
Families Citing this family (194)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US20120311585A1 (en) | 2011-06-03 | 2012-12-06 | Apple Inc. | Organizing task items that represent tasks to perform |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10417037B2 (en) | 2012-05-15 | 2019-09-17 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
AU2014214676A1 (en) | 2013-02-07 | 2015-08-27 | Apple Inc. | Voice trigger for a digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
KR101922663B1 (en) | 2013-06-09 | 2018-11-28 | 애플 인크. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
DE112014003653B4 (en) | 2013-08-06 | 2024-04-18 | Apple Inc. | Automatically activate intelligent responses based on activities from remote devices |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10152299B2 (en) | 2015-03-06 | 2018-12-11 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10460227B2 (en) | 2015-05-15 | 2019-10-29 | Apple Inc. | Virtual assistant in a communication session |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10200824B2 (en) | 2015-05-27 | 2019-02-05 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US20160378747A1 (en) | 2015-06-29 | 2016-12-29 | Apple Inc. | Virtual assistant for media playback |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10740384B2 (en) | 2015-09-08 | 2020-08-11 | Apple Inc. | Intelligent automated assistant for media search and playback |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10331312B2 (en) | 2015-09-08 | 2019-06-25 | Apple Inc. | Intelligent automated assistant in a media environment |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10956666B2 (en) | 2015-11-09 | 2021-03-23 | Apple Inc. | Unconventional virtual assistant interactions |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US9826306B2 (en) | 2016-02-22 | 2017-11-21 | Sonos, Inc. | Default playback device designation |
US10095470B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Audio response playback |
US10264030B2 (en) | 2016-02-22 | 2019-04-16 | Sonos, Inc. | Networked microphone device control |
US10509626B2 (en) | 2016-02-22 | 2019-12-17 | Sonos, Inc | Handling of loss of pairing between networked devices |
US9947316B2 (en) | 2016-02-22 | 2018-04-17 | Sonos, Inc. | Voice control of a media playback system |
US9965247B2 (en) | 2016-02-22 | 2018-05-08 | Sonos, Inc. | Voice controlled media playback system based on user profile |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US12223282B2 (en) | 2016-06-09 | 2025-02-11 | Apple Inc. | Intelligent automated assistant in a home environment |
US9978390B2 (en) | 2016-06-09 | 2018-05-22 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
US12197817B2 (en) | 2016-06-11 | 2025-01-14 | Apple Inc. | Intelligent device arbitration and control |
US10152969B2 (en) | 2016-07-15 | 2018-12-11 | Sonos, Inc. | Voice detection by multiple devices |
US10134399B2 (en) | 2016-07-15 | 2018-11-20 | Sonos, Inc. | Contextualization of voice inputs |
US10115400B2 (en) | 2016-08-05 | 2018-10-30 | Sonos, Inc. | Multiple voice services |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US9942678B1 (en) | 2016-09-27 | 2018-04-10 | Sonos, Inc. | Audio playback settings for voice interaction |
US9743204B1 (en) | 2016-09-30 | 2017-08-22 | Sonos, Inc. | Multi-orientation playback device microphones |
US10181323B2 (en) | 2016-10-19 | 2019-01-15 | Sonos, Inc. | Arbitration-based voice recognition |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
US11183181B2 (en) | 2017-03-27 | 2021-11-23 | Sonos, Inc. | Systems and methods of multiple voice services |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | User interface for correcting recognition errors |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK180048B1 (en) | 2017-05-11 | 2020-02-04 | Apple Inc. | MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | Low-latency intelligent automated assistant |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770411A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | MULTI-MODAL INTERFACES |
US20180336892A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Detecting a trigger of a digital assistant |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US10475449B2 (en) | 2017-08-07 | 2019-11-12 | Sonos, Inc. | Wake-word detection suppression |
US10048930B1 (en) | 2017-09-08 | 2018-08-14 | Sonos, Inc. | Dynamic computation of system response volume |
US10446165B2 (en) | 2017-09-27 | 2019-10-15 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US10482868B2 (en) | 2017-09-28 | 2019-11-19 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US10051366B1 (en) | 2017-09-28 | 2018-08-14 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US10621981B2 (en) | 2017-09-28 | 2020-04-14 | Sonos, Inc. | Tone interference cancellation |
US10466962B2 (en) | 2017-09-29 | 2019-11-05 | Sonos, Inc. | Media playback system with voice assistance |
US10880650B2 (en) | 2017-12-10 | 2020-12-29 | Sonos, Inc. | Network microphone devices with automatic do not disturb actuation capabilities |
US10818290B2 (en) | 2017-12-11 | 2020-10-27 | Sonos, Inc. | Home graph |
WO2019152722A1 (en) | 2018-01-31 | 2019-08-08 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
CN108712566B (en) * | 2018-04-27 | 2020-10-30 | 维沃移动通信有限公司 | Voice assistant awakening method and mobile terminal |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US11175880B2 (en) | 2018-05-10 | 2021-11-16 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US10847178B2 (en) | 2018-05-18 | 2020-11-24 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection |
US10959029B2 (en) | 2018-05-25 | 2021-03-23 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
CN111954868A (en) * | 2018-05-31 | 2020-11-17 | 江苏树实科技有限公司 | Multi-voice assistant control method, device, system and computer readable storage medium |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US10681460B2 (en) | 2018-06-28 | 2020-06-09 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
CN110868648A (en) * | 2018-08-27 | 2020-03-06 | 杭州海康威视数字技术股份有限公司 | Intelligent voice realization method of indoor intercom device and indoor intercom device |
US11076035B2 (en) | 2018-08-28 | 2021-07-27 | Sonos, Inc. | Do not disturb feature for audio notifications |
US10461710B1 (en) | 2018-08-28 | 2019-10-29 | Sonos, Inc. | Media playback system with maximum volume setting |
US10878811B2 (en) | 2018-09-14 | 2020-12-29 | Sonos, Inc. | Networked devices, systems, and methods for intelligently deactivating wake-word engines |
US10587430B1 (en) | 2018-09-14 | 2020-03-10 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US11024331B2 (en) | 2018-09-21 | 2021-06-01 | Sonos, Inc. | Voice detection optimization using sound metadata |
US10811015B2 (en) | 2018-09-25 | 2020-10-20 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11100923B2 (en) | 2018-09-28 | 2021-08-24 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US10692518B2 (en) | 2018-09-29 | 2020-06-23 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
CN109147786A (en) * | 2018-09-30 | 2019-01-04 | 联想(北京)有限公司 | A kind of information processing method and electronic equipment |
KR20200044175A (en) | 2018-10-05 | 2020-04-29 | 삼성전자주식회사 | Electronic apparatus and assistant service providing method thereof |
US11899519B2 (en) | 2018-10-23 | 2024-02-13 | Sonos, Inc. | Multiple stage network microphone device with reduced power consumption and processing load |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
CN109325337A (en) * | 2018-11-05 | 2019-02-12 | 北京小米移动软件有限公司 | Unlocking method and device |
EP3654249A1 (en) | 2018-11-15 | 2020-05-20 | Snips | Dilated convolutions and gating for efficient keyword spotting |
CN109658927A (en) * | 2018-11-30 | 2019-04-19 | 北京小米移动软件有限公司 | Wake-up processing method, device and the management equipment of smart machine |
US11183183B2 (en) | 2018-12-07 | 2021-11-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11132989B2 (en) | 2018-12-13 | 2021-09-28 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US10602268B1 (en) | 2018-12-20 | 2020-03-24 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11315556B2 (en) | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
US10867604B2 (en) | 2019-02-08 | 2020-12-15 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
CN110310641B (en) * | 2019-02-26 | 2022-08-26 | 杭州蓦然认知科技有限公司 | Method and device for voice assistant |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
CN110164433A (en) * | 2019-04-03 | 2019-08-23 | 美国乐歌有限公司 | A kind of speech control system and method for lifting column |
US11120794B2 (en) | 2019-05-03 | 2021-09-14 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
DK201970509A1 (en) | 2019-05-06 | 2021-01-15 | Apple Inc | Spoken notifications |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
US11189279B2 (en) * | 2019-05-22 | 2021-11-30 | Microsoft Technology Licensing, Llc | Activation management for multiple voice assistants |
DK180129B1 (en) | 2019-05-31 | 2020-06-02 | Apple Inc. | USER ACTIVITY SHORTCUT SUGGESTIONS |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
DK201970511A1 (en) | 2019-05-31 | 2021-02-15 | Apple Inc | Voice identification in digital assistant systems |
US11227599B2 (en) | 2019-06-01 | 2022-01-18 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11361756B2 (en) | 2019-06-12 | 2022-06-14 | Sonos, Inc. | Conditional wake word eventing based on environment |
US10586540B1 (en) | 2019-06-12 | 2020-03-10 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US11200894B2 (en) | 2019-06-12 | 2021-12-14 | Sonos, Inc. | Network microphone device with command keyword eventing |
CN110310636A (en) * | 2019-06-24 | 2019-10-08 | 歌尔股份有限公司 | Interaction control method, device, equipment and audio frequency apparatus |
CN110335601A (en) * | 2019-07-10 | 2019-10-15 | 三星电子(中国)研发中心 | Voice assistant equipment and its voice awakening method |
US11138969B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US10871943B1 (en) | 2019-07-31 | 2020-12-22 | Sonos, Inc. | Noise classification for event detection |
US11138975B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
CN112466300B (en) * | 2019-09-09 | 2024-06-18 | 百度在线网络技术(北京)有限公司 | Interaction method, electronic device, intelligent device and readable storage medium |
CN110718219B (en) | 2019-09-12 | 2022-07-22 | 百度在线网络技术(北京)有限公司 | Voice processing method, device, equipment and computer storage medium |
CN110718218B (en) * | 2019-09-12 | 2022-08-23 | 百度在线网络技术(北京)有限公司 | Voice processing method, device, equipment and computer storage medium |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
CN110659361B (en) * | 2019-10-11 | 2023-01-17 | 卢卡(北京)智能科技有限公司 | Conversation method, device, equipment and medium |
US11189286B2 (en) | 2019-10-22 | 2021-11-30 | Sonos, Inc. | VAS toggle based on device orientation |
KR20210064594A (en) * | 2019-11-26 | 2021-06-03 | 삼성전자주식회사 | Electronic apparatus and control method thereof |
CN111091814A (en) * | 2019-12-13 | 2020-05-01 | 晶晨半导体(深圳)有限公司 | Method for constructing multi-voice assistant |
US11200900B2 (en) | 2019-12-20 | 2021-12-14 | Sonos, Inc. | Offline voice control |
CN111161714B (en) * | 2019-12-25 | 2023-07-21 | 联想(北京)有限公司 | Voice information processing method, electronic equipment and storage medium |
CN110933500B (en) * | 2019-12-30 | 2022-07-29 | 深圳Tcl新技术有限公司 | Voice triggering method, device, equipment and computer storage medium |
CN111096680B (en) * | 2019-12-31 | 2022-02-01 | 广东美的厨房电器制造有限公司 | Cooking equipment, electronic equipment, voice server, voice control method and device |
CN113129886A (en) * | 2019-12-31 | 2021-07-16 | 深圳市茁壮网络股份有限公司 | Switching method and system of voice recognition function |
US11562740B2 (en) | 2020-01-07 | 2023-01-24 | Sonos, Inc. | Voice verification for media playback |
CN111261160B (en) * | 2020-01-20 | 2023-09-19 | 联想(北京)有限公司 | Signal processing method and device |
KR20210094251A (en) * | 2020-01-21 | 2021-07-29 | 삼성전자주식회사 | Display apparatus and controlling method thereof |
US11556307B2 (en) | 2020-01-31 | 2023-01-17 | Sonos, Inc. | Local voice data processing |
US11308958B2 (en) | 2020-02-07 | 2022-04-19 | Sonos, Inc. | Localized wakeword verification |
US11183193B1 (en) | 2020-05-11 | 2021-11-23 | Apple Inc. | Digital assistant hardware abstraction |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11727919B2 (en) | 2020-05-20 | 2023-08-15 | Sonos, Inc. | Memory allocation for keyword spotting engines |
US11482224B2 (en) | 2020-05-20 | 2022-10-25 | Sonos, Inc. | Command keywords with input detection windowing |
US11308962B2 (en) | 2020-05-20 | 2022-04-19 | Sonos, Inc. | Input detection windowing |
CN111667827B (en) * | 2020-05-28 | 2023-10-17 | 北京小米松果电子有限公司 | Voice control method and device for application program and storage medium |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
US11698771B2 (en) | 2020-08-25 | 2023-07-11 | Sonos, Inc. | Vocal guidance engines for playback devices |
WO2022061293A1 (en) | 2020-09-21 | 2022-03-24 | VIDAA USA, Inc. | Display apparatus and signal transmission method for display apparatus |
CN112165640B (en) * | 2020-09-21 | 2023-04-14 | Vidaa美国公司 | Display device |
CN112291436B (en) * | 2020-10-23 | 2022-03-01 | 杭州蓦然认知科技有限公司 | Method and device for scheduling calling subscriber |
CN112261232B (en) * | 2020-10-23 | 2022-03-01 | 杭州蓦然认知科技有限公司 | Method for processing voice instruction of calling party by voice assistant and voice assistant |
CN112291431B (en) * | 2020-10-23 | 2021-10-01 | 北京蓦然认知科技有限公司 | Method and device for synchronously controlling call |
US11984123B2 (en) | 2020-11-12 | 2024-05-14 | Sonos, Inc. | Network device interaction by range |
CN112527236A (en) * | 2020-12-23 | 2021-03-19 | 北京梧桐车联科技有限责任公司 | Voice mode control method, device and storage medium |
US11551700B2 (en) | 2021-01-25 | 2023-01-10 | Sonos, Inc. | Systems and methods for power-efficient keyword detection |
CN112817556A (en) * | 2021-02-04 | 2021-05-18 | 青岛海信传媒网络技术有限公司 | Switching method of voice scheme on display equipment, display equipment and control device |
WO2022160911A1 (en) * | 2021-01-29 | 2022-08-04 | 青岛海信传媒网络技术有限公司 | Switching method for voice solution on display device, display device, and control apparatus |
US11657818B2 (en) * | 2021-03-10 | 2023-05-23 | GM Global Technology Operations LLC | Multi-assistant control |
CN115346523B (en) * | 2021-05-14 | 2024-10-22 | 联发科技(新加坡)私人有限公司 | Voice interaction method and device of voice equipment |
CN113496703A (en) * | 2021-07-23 | 2021-10-12 | 北京百度网讯科技有限公司 | Method, device and program product for controlling program in voice mode |
CN114115620B (en) * | 2021-10-27 | 2023-10-24 | 青岛海尔科技有限公司 | Prompt box response method and device, storage medium and electronic device |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104219388A (en) * | 2014-08-28 | 2014-12-17 | 小米科技有限责任公司 | Voice control method and device |
US20150186156A1 (en) * | 2013-12-31 | 2015-07-02 | Next It Corporation | Virtual assistant conversations |
CN105408919A (en) * | 2013-07-03 | 2016-03-16 | 埃森哲环球服务有限公司 | Query response device |
CN106462617A (en) * | 2014-06-30 | 2017-02-22 | 苹果公司 | Intelligent automated assistant for tv user interactions |
CN106469040A (en) * | 2015-08-19 | 2017-03-01 | 华为终端(东莞)有限公司 | Communication means, server and equipment |
CN106898352A (en) * | 2017-02-27 | 2017-06-27 | 联想(北京)有限公司 | Sound control method and electronic equipment |
CN107122179A (en) * | 2017-03-31 | 2017-09-01 | 阿里巴巴集团控股有限公司 | The function control method and device of voice |
US20170269975A1 (en) * | 2016-03-17 | 2017-09-21 | Nuance Communications, Inc. | Session processing interaction between two or more virtual assistants |
CN107393548A (en) * | 2017-07-05 | 2017-11-24 | 青岛海信电器股份有限公司 | The processing method and processing device of the voice messaging of multiple voice assistant equipment collections |
CN110741347A (en) * | 2017-10-03 | 2020-01-31 | 谷歌有限责任公司 | Coordination of multiple digital assistants in the vehicle environment |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4377361B2 (en) * | 2005-09-12 | 2009-12-02 | 株式会社ナビタイムジャパン | Navigation method and navigation system using a plurality of portable terminals |
CN101916166B (en) * | 2010-08-19 | 2013-12-04 | 中兴通讯股份有限公司 | Method for starting application program and mobile terminal |
US9172747B2 (en) * | 2013-02-25 | 2015-10-27 | Artificial Solutions Iberia SL | System and methods for virtual assistant networks |
US10311482B2 (en) * | 2013-11-11 | 2019-06-04 | At&T Intellectual Property I, Lp | Method and apparatus for adjusting a digital assistant persona |
EP2881898A1 (en) * | 2013-12-09 | 2015-06-10 | Accenture Global Services Limited | Virtual assistant interactivity platform |
CN106648289A (en) * | 2016-09-14 | 2017-05-10 | 依偎科技(南昌)有限公司 | Method for switching application program, switching device and terminal |
-
2017
- 2017-12-07 CN CN201711287860.8A patent/CN107919123B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105408919A (en) * | 2013-07-03 | 2016-03-16 | 埃森哲环球服务有限公司 | Query response device |
US20150186156A1 (en) * | 2013-12-31 | 2015-07-02 | Next It Corporation | Virtual assistant conversations |
CN106462617A (en) * | 2014-06-30 | 2017-02-22 | 苹果公司 | Intelligent automated assistant for tv user interactions |
CN104219388A (en) * | 2014-08-28 | 2014-12-17 | 小米科技有限责任公司 | Voice control method and device |
CN106469040A (en) * | 2015-08-19 | 2017-03-01 | 华为终端(东莞)有限公司 | Communication means, server and equipment |
US20170269975A1 (en) * | 2016-03-17 | 2017-09-21 | Nuance Communications, Inc. | Session processing interaction between two or more virtual assistants |
CN106898352A (en) * | 2017-02-27 | 2017-06-27 | 联想(北京)有限公司 | Sound control method and electronic equipment |
CN107122179A (en) * | 2017-03-31 | 2017-09-01 | 阿里巴巴集团控股有限公司 | The function control method and device of voice |
CN107393548A (en) * | 2017-07-05 | 2017-11-24 | 青岛海信电器股份有限公司 | The processing method and processing device of the voice messaging of multiple voice assistant equipment collections |
CN110741347A (en) * | 2017-10-03 | 2020-01-31 | 谷歌有限责任公司 | Coordination of multiple digital assistants in the vehicle environment |
Also Published As
Publication number | Publication date |
---|---|
CN107919123A (en) | 2018-04-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107919123B (en) | Multi-voice assistant control method, device and computer readable storage medium | |
CN110554815B (en) | Icon awakening method, electronic device and storage medium | |
CN108804010B (en) | Terminal control method, device and computer readable storage medium | |
CN104503691B (en) | Apparatus control method, device and intelligent terminal | |
CN106970754B (en) | Screen capture processing method and device | |
EP3113549B1 (en) | Method and device for waking up mcu chip | |
CN112037787B (en) | Wake-up control method, device and computer readable storage medium | |
KR101837333B1 (en) | Method and apparatus for awakening electronic device | |
US10241637B2 (en) | Method for reducing power consumption in electronic devices | |
EP3185160B1 (en) | Screen unlocking method and apparatus, terminal | |
US20180159809A1 (en) | Method and device for message reading | |
EP3112979B1 (en) | Method and device for waking up mcu | |
CN111063354B (en) | Man-machine interaction method and device | |
CN105677460A (en) | Application processing method and apparatus | |
EP3933570A1 (en) | Method and apparatus for controlling a voice assistant, and computer-readable storage medium | |
EP4184506A1 (en) | Audio processing | |
CN106357934B (en) | Screen locking control method and device | |
CN107506246A (en) | Information displaying method, device and computer-readable recording medium | |
CN109521923B (en) | Floating window control method and device and storage medium | |
CN106775377B (en) | Gesture recognition device, equipment and control method of gesture recognition device | |
CN113361305B (en) | Fingerprint acquisition method, electronic device and computer readable storage medium | |
CN109522058A (en) | Awakening method, device, terminal and storage medium | |
CN108874450B (en) | Method and device for waking up voice assistant | |
CN108766427B (en) | Voice control method and device | |
CN105786561B (en) | Method and device for calling process |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |