Intelligent voice interaction method, device and system

文档序号:9760 发布日期:2021-09-17 浏览:41次 中文

1. An intelligent voice interaction method is applied to a portable intelligent device, and comprises the following steps:

under the condition of being connected with a target base, acquiring identification information of the target base, and determining a target application scene corresponding to the target base according to the identification information;

receiving a voice signal and acquiring a processing instruction corresponding to the voice signal;

and executing corresponding target operation based on the processing instruction and the target application scene.

2. The method of claim 1, wherein the identification information comprises at least one of:

a device identification of the target base;

a designated field predefined in a communication protocol of the target base and the portable smart device.

3. The method of claim 1,

the target application scene comprises a vehicle-mounted scene and a family scene.

4. The method of claim 3, wherein when the target application scene is determined to be the in-vehicle scene according to the identification information, the executing the corresponding target operation based on the processing instruction and the target application scene comprises:

judging the instruction type of the processing instruction in the vehicle-mounted scene;

and executing the target operation corresponding to the instruction type.

5. The method of claim 4, wherein the instruction types include a first type and a second type; wherein, judging the instruction type of the processing instruction in the vehicle-mounted scene comprises:

judging whether an execution object of the processing instruction is a preset execution object or not;

if so, determining the instruction type of the processing instruction to be the first type;

if not, determining that the instruction type of the processing instruction is the second type.

6. The method of claim 5, wherein the first type of instruction is executed by an in-vehicle central control system and the second type of instruction is executed by the portable smart device;

wherein executing the target operation corresponding to the instruction type comprises:

when the instruction type is the first type, sending the processing instruction to the vehicle-mounted central control system, and executing an operation corresponding to the processing instruction by the vehicle-mounted central control system;

and executing the operation corresponding to the processing instruction when the instruction type is the second type.

7. The method of claim 3, wherein when the target application scenario is determined to be the home scenario according to the identification information, the performing a corresponding target operation based on the processing instruction and the target application scenario comprises:

determining the intelligent household appliances matched with the processing instruction in the family scene;

and sending the processing instruction to the intelligent household appliance, and executing the operation corresponding to the processing instruction by the intelligent household appliance.

8. An interactive device of intelligent voice, comprising:

the scene determining unit is used for acquiring identification information of a target base under the condition of being connected with the target base and determining a target application scene corresponding to the target base according to the identification information;

the instruction acquisition unit is used for receiving a voice signal and acquiring a processing instruction corresponding to the voice signal;

and the instruction execution unit executes corresponding target operation based on the processing instruction and the target application scene.

9. An electronic device comprising a processor and a memory electrically connected to the processor, the memory storing a program or instructions which, when executed by the processor, implement the method of any one of claims 1 to 7.

10. The intelligent voice interaction system is characterized by comprising a portable intelligent device and a target base, wherein:

under the condition that the portable intelligent equipment is connected with the target base, the portable intelligent equipment acquires identification information of the target base and determines a target application scene corresponding to the target base according to the identification information;

the portable intelligent equipment receives a voice signal and acquires a processing instruction corresponding to the voice signal;

and the portable intelligent equipment executes corresponding target operation based on the processing instruction and the target application scene.

Background

Currently, in different application scenarios, a user can perform intelligent voice interaction with a device. For example, in an on-vehicle scenario, a user may perform intelligent voice interaction with an on-vehicle central control system in a vehicle to control the vehicle to perform a corresponding operation. For another example, in a home scenario, a user may perform an intelligent voice interaction with an intelligent appliance to control the intelligent appliance to perform a corresponding operation.

However, with the demand of the users for diversified lives, more and more users want to realize intelligent voice interaction under different scenes by means of the same device, and an effective scheme for realizing the purpose is still lacking at present.

Disclosure of Invention

The embodiment of the application provides an intelligent voice interaction method, device and system, which are used for solving the problem that intelligent voice interaction under different scenes cannot be realized through the same equipment at present.

In order to solve the above technical problem, the embodiment of the present application is implemented as follows:

in a first aspect, an interaction method of intelligent voice is provided, which is applied to a portable intelligent device, and includes:

under the condition of being connected with a target base, acquiring identification information of the target base, and determining a target application scene corresponding to the target base according to the identification information;

receiving a voice signal and acquiring a processing instruction corresponding to the voice signal;

and executing corresponding target operation based on the processing instruction and the target application scene.

In a second aspect, an interactive device for smart voice is provided, which includes:

the scene determining unit is used for acquiring identification information of a target base under the condition of being connected with the target base and determining a target application scene corresponding to the target base according to the identification information;

the instruction acquisition unit is used for receiving a voice signal and acquiring a processing instruction corresponding to the voice signal;

and the instruction execution unit executes corresponding target operation based on the processing instruction and the target application scene.

In a third aspect, an electronic device is provided, which includes a processor and a memory electrically connected to the processor, where the memory stores a program or instructions, and the program or instructions, when executed by the processor, implement the method according to the first aspect.

In a fourth aspect, a readable storage medium is proposed, on which a program or instructions are stored, which when executed by a processor implement the method according to the first aspect.

In a fifth aspect, an interactive system of smart voice is provided, the system includes a portable smart device and a target base, wherein:

under the condition that the portable intelligent equipment is connected with the target base, the portable intelligent equipment acquires identification information of the target base and determines a target application scene corresponding to the target base according to the identification information;

the portable intelligent equipment receives a voice signal and acquires a processing instruction corresponding to the voice signal;

and the portable intelligent equipment executes corresponding target operation based on the processing instruction and the target application scene.

The embodiment of the application adopts at least one technical scheme which can achieve the following beneficial effects:

the embodiment of the application can set corresponding base identification information aiming at different application scenes in advance, when a user carries a portable intelligent device and the portable intelligent device is connected with a certain target base, the portable intelligent device can acquire the identification information of the target base, the corresponding target application scene is determined according to the identification information, and when a voice signal is received and a processing instruction corresponding to the voice signal is acquired, corresponding operation can be executed according to the processing instruction and the target application scene. Therefore, the user can carry the portable intelligent device to any application scene, and in any scene, the portable intelligent device can identify the scene through the identification information of the target base and carry out intelligent voice interaction in the scene, so that the purpose of carrying out intelligent voice interaction in different scenes through the same device can be achieved. In addition, because the embodiment of the application can share the same set of software and hardware equipment to realize intelligent voice interaction in different application scenes, the consistency of intelligent voice interaction experience in different scenes can be ensured.

Drawings

In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings needed to be used in the description of the embodiments or the prior art will be briefly introduced below, it is obvious that the drawings in the following description are only some embodiments described in the present application, and for those skilled in the art, other drawings can be obtained according to the drawings without any creative effort.

FIG. 1 is a flow diagram illustrating an interaction method for smart voice according to an embodiment of the present application;

FIG. 2 is a flow diagram illustrating a method for interaction with smart voice in accordance with an embodiment of the present application;

FIG. 3 is a schematic structural diagram of an electronic device according to an embodiment of the present application;

FIG. 4 is a schematic structural diagram of an intelligent voice interaction device according to an embodiment of the present application;

fig. 5 is a schematic structural diagram of an interactive system for smart voice according to an embodiment of the present application.

Detailed Description

In order to make those skilled in the art better understand the technical solutions in the present application, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.

The intelligent voice interaction method provided by the embodiment of the application can be applied to portable intelligent equipment, the portable intelligent equipment at least has a voice receiving and sending function, a voice recognition function and a function of being connected with the base and performing data interaction with the base, and the base can also supply power to the portable intelligent equipment. Optionally, the portable intelligent device may further have a function of interacting with the server over a network, or may further have a function of recognizing and analyzing a voice signal.

In a possible implementation manner, the portable smart device may specifically be composed of the following parts: the system comprises a Text To Speech (TTS) broadcasting system, a pickup microphone, a network communication module and a base connector. Wherein, TTS reports system is used for outwards sending pronunciation, and pickup microphone is used for receiving external pronunciation, and network communication module is used for carrying out the communication with the server, and the base connector can be the base connector of formula of magnetism (be convenient for portable smart machine to inhale the mode of formula through magnetism and be connected with the base), like collapsible magnetism inhale the plug, also can be the base connector who uses other connected modes. Optionally, the portable intelligent device may further include an interactive information display screen, where the display screen may display interactive information between the portable intelligent device and the base and/or the server, and in addition, the display screen may also display attribute information and/or function information of the portable intelligent device, and a user may set attributes, functions, and the like of the portable intelligent device through the display screen.

The technical solutions provided by the embodiments of the present application are described in detail below with reference to the accompanying drawings.

Fig. 1 is a flowchart illustrating an interaction method of smart voice according to an embodiment of the present application. The method can be applied to the portable intelligent device and specifically comprises the following steps.

S102: and under the condition of being connected with the target base, acquiring identification information of the target base, and determining a target application scene corresponding to the target base according to the identification information.

In this embodiment, the application scenarios of the intelligent voice interaction may include multiple types, different bases may have different identification information, and a correspondence relationship may be established in advance for different application scenarios and different identification information. When a user needs to perform intelligent voice interaction in a certain application scene, the portable intelligent device can be carried to the application scene, meanwhile, a base can be placed in the application scene, and the identification information of the base and the application scene have a preset corresponding relation.

For a certain application scenario, after the user carries the portable intelligent device to the application scenario, the user may connect the portable intelligent device with a base (for convenience of distinguishing, hereinafter, represented by a target base) placed in the scenario in advance under the condition that intelligent voice interaction is required. Under the condition that the portable intelligent equipment is connected with the target base, the portable intelligent equipment can perform data interaction with the target base. Through data interaction, the portable intelligent device can acquire the identification information of the target base. After the identification information of the target base is acquired, the portable intelligent device may further determine an application scene corresponding to the target base (for convenience of distinguishing, hereinafter, represented by a target application scene) according to the identification information and a preset corresponding relationship between the identification information and the application scene, where the target application scene is an application scene of current intelligent voice interaction.

In an implementation manner, the identification information of the target base may specifically be an equipment identification of the target base. The portable intelligent device can identify and obtain the device identification of the target base under the condition of carrying out data interaction with the target base.

In another implementation, the identification information of the target base may be a specific field predefined in a communication protocol between the target base and the portable intelligent device. Specifically, some fields (i.e., designated fields) may be predefined in the communication protocol of the target base and the portable intelligent device, for example, a part of sub-segments may be added as designated fields in the communication protocol, where different fields may be defined for different application scenarios. Therefore, after the portable intelligent device is connected with the target base, when the portable intelligent device is communicated with the target base, the specified field in the communication protocol can be obtained, and the corresponding target application scene is further determined according to the specified field.

It should be understood that in other implementations, the identification information of the target base may also be other identification information as long as different application scenarios can be distinguished, which is not illustrated here.

S104: and receiving the voice signal and acquiring a processing instruction corresponding to the voice signal.

Under the condition that the target application scene is identified, when a voice signal (such as voice sent by a user) from the outside is received, the portable intelligent device can acquire a processing instruction corresponding to the voice signal.

In this embodiment, when acquiring the processing instruction, the portable intelligent device may be implemented by using a server or locally. When the processing instruction is obtained by the server, firstly, after receiving the voice signal, the portable device can perform voice recognition on the voice signal to obtain corresponding text information. For example, audio information may be converted into textual information by ASR. After the text information is obtained, the text information can be sent to the server. The portable intelligent device can be connected to the network when being connected with the base, and can transmit the text message through the network when transmitting the text message to the server. The portable intelligent device is connected to the network, specifically, the portable intelligent device may be connected to the network through a USB protocol, or the portable intelligent device may be connected to the network by a user when the portable intelligent device is connected to the base for the first time, and the portable intelligent device may be automatically connected to the network when the portable intelligent device is connected to the base for the subsequent time. Of course, the connection to the network may be achieved in other ways, which are not illustrated here.

After receiving the text information, the server side can perform semantic analysis on the text information to obtain an analysis result, and the analysis result can represent the intention of the voice signal. After the analysis result is obtained, the server can further obtain a processing instruction according to the analysis result. The specific implementation manner of semantic parsing performed by the server may refer to a method recorded in the prior art, and is not described in detail here. After the server side obtains the processing instruction, the processing instruction can be returned to the portable intelligent device.

Under the condition that the portable intelligent device locally obtains a processing instruction, specifically, firstly, after receiving a voice signal, the portable intelligent device can perform voice recognition on the voice signal to obtain corresponding text information; secondly, performing semantic analysis based on the text information to obtain an analysis result; and finally, further obtaining a processing instruction according to the analysis result.

It should be noted that the difference between the above-mentioned obtaining of the processing instruction by the server and obtaining of the processing instruction locally is that the execution subject of the semantic parsing operation is different, where when the processing instruction is obtained by the server, the execution subject of the semantic parsing is the server, and when the processing instruction is obtained locally, the execution subject is the portable intelligent device. In practical application, if the portable intelligent device has the capability of semantic analysis, the semantic analysis is preferably performed locally to obtain a processing instruction, and text information does not need to be uploaded to a server side for semantic analysis. On the contrary, if the portable intelligent device does not have the semantic analysis capability, the text information needs to be uploaded to the server side for semantic analysis.

S106: and executing corresponding target operation based on the processing instruction and the target application scene.

After the portable intelligent device obtains the processing instruction, the portable intelligent device may execute a corresponding target operation based on the processing instruction and the target application scenario determined in S102.

In this embodiment, the target application scene determined by the S102 according to the identification information of the target base may be a vehicle-mounted scene or a home scene. When the corresponding target operation is performed based on the processing instruction and the target application scene, different target operations may be performed for the in-vehicle scene and the home scene. The following will be separately explained.

In the case that the target application scenario is an in-vehicle scenario, it is considered that the voice received by the portable smart device may be that the portable smart device is intended to perform some operation (such as a scenario of human-computer interaction with the portable smart device), or that the user is intended to perform some operation in a vehicle, in situations where the vehicle is required to perform some operation, the portable smart device is typically unable to directly interact with components in the vehicle, but can interact with an on-board central control system in the vehicle, which can directly control components in the vehicle, and therefore, the processing instructions in the vehicle-mounted scene can be divided into different instruction types, and when the portable intelligent device executes corresponding target operation based on the processing instructions and the target application scene, the instruction type of the processing instruction in the vehicle-mounted scene can be judged first, and then the target operation corresponding to the determined instruction type is executed. The instruction types of the processing instructions at least comprise a first type and a second type, the first type of instructions represent instructions which can be executed by an on-board central control system in the vehicle, and the second type of instructions represent instructions which are executed by the portable intelligent device.

In determining the instruction type of the processing instruction, in one possible implementation, it is considered that the execution object of the instruction is usually involved in the processing instruction, and therefore, the type of the instruction may be determined by the execution object involved in the processing instruction. Specifically, it may be determined whether an execution object of the processing instruction is a preset execution object; if so, determining the instruction type of the processing instruction as a first type; if not, determining that the instruction type of the processing instruction is the second type. The preset execution object may be a name of any component in the vehicle, where the component may be software in the vehicle, such as a music player, or may also be hardware in the vehicle, such as an air conditioner.

Taking an example that the preset execution object includes an air conditioner, if the processing instruction is "turn on the air conditioner", it may be determined that the execution object of the processing instruction is the preset execution object, and the instruction type of the processing instruction is the first type. If the processing instruction is 'weather after broadcasting for one hour', it may be determined that the execution object of the processing instruction is not a preset execution object, and the instruction type of the processing instruction is a second type.

After the instruction type of the processing instruction is determined, if the judgment result is the first type, the portable intelligent device can send the processing instruction to the vehicle-mounted central control system. And after receiving the processing instruction, the vehicle-mounted central control system can execute the operation corresponding to the processing instruction. For example, if the processing instruction is "open song XX in music player", the in-vehicle central control system may control the music player in the vehicle to play song "XX". If the judgment result is the second type, the portable intelligent device can execute the operation corresponding to the processing instruction. For example, if the processing instruction is "broadcast weather after one hour", the operation may be that the portable intelligent device performs TTS broadcast.

Optionally, in a case that the instruction type of the processing instruction is the first type, considering that a multi-person conversation scenario among the user, the portable intelligent device and the vehicle-mounted central control system may be involved, the vehicle-mounted central control system may further return an execution result to the portable intelligent device after executing an operation corresponding to the processing instruction, so as to achieve the purpose of multi-person conversation.

In this way, in the vehicle-mounted scene, by judging the type of the processing instruction, executing the corresponding operation by the vehicle-mounted central control system in the case of the first type, and executing the corresponding operation by the portable intelligent device in the case of the second type, the intelligent voice interaction in the vehicle-mounted scene can be realized.

In the case that the target application scene is a home scene, considering that the voice received by the portable intelligent device generally intends to the intelligent household appliance to execute a certain operation, and the portable intelligent device generally can directly interact with the intelligent household appliance, therefore, the instruction type does not need to be divided into the processing instructions in the home scene, when the portable intelligent device executes the corresponding target operation based on the processing instructions and the target application scene, the portable intelligent device can determine the intelligent household appliance matched with the processing instructions in the home scene, and then sends the processing instructions to the intelligent household appliance, and the intelligent household appliance executes the operation corresponding to the processing instructions.

In this embodiment, the portable intelligent device may be bound to the intelligent home appliance in the home environment in advance. For example, when the portable intelligent device is connected with the base under the home environment for the first time, the user can bind the portable intelligent device with the intelligent household appliances in the home environment through the mobile phone APP. In addition, related instructions of the intelligent household appliance can be stored in the portable intelligent device in advance. Therefore, when the intelligent household appliance matched with the processing instruction in the home environment is determined, the processing instruction can be matched with the pre-stored instruction of the intelligent household appliance, and the intelligent household appliance matched with the processing instruction can be determined. Of course, in other implementation manners, the intelligent appliance matching the processing instruction may be determined by other methods, which are not illustrated here.

After the matched intelligent household appliances are determined, the processing instruction can be directly sent to the intelligent household appliances, and the intelligent household appliances can execute the operation corresponding to the processing instruction after receiving the processing instruction. For example, if the processing instruction is "turn on the air conditioner, 20 degrees", the portable intelligent device sends the instruction to the intelligent air conditioner, and the intelligent air conditioner automatically turns on and adjusts the temperature to 20 degrees.

Optionally, after the intelligent appliance executes the operation corresponding to the processing instruction, the execution result may also be returned to the portable intelligent device, so as to confirm whether the execution is successful. If the execution is successful, the user can be prompted by voice, and if the execution is failed, the user can perform voice interaction again.

In order to facilitate understanding of the technical solutions provided in the embodiments of the present application, a vehicle-mounted scene and a home scene are taken as examples and described below. Please refer to fig. 2.

Fig. 2 is a flowchart illustrating an interaction method of smart voice according to an embodiment of the present application. The execution subject of the embodiment shown in fig. 2 may be a portable smart device, and specifically includes the following steps.

S201: and acquiring the identification information of the target base under the condition of connecting with the target base.

S202: and determining a target application scene corresponding to the target base according to the identification information.

S203: and receiving the voice signal and acquiring a processing instruction corresponding to the voice signal.

If the target application scene determined in S202 is a vehicle-mounted scene, S204 may be executed; if the target application scenario is a home scenario, S207 may be executed.

S204: and judging the instruction type of the processing instruction in the vehicle-mounted scene.

If the instruction type is the first type, executing S205; if the instruction type is the second type, S206 is executed. The first type of instruction is executed by the vehicle-mounted central control system, and the second type of instruction is executed by the portable intelligent device.

S205: and sending the processing instruction to the vehicle-mounted central control system, and executing the operation corresponding to the processing instruction by the vehicle-mounted central control system.

S206: and executing the operation corresponding to the processing instruction.

S207: and determining the intelligent household appliances matched with the processing instruction in the family scene.

S208: and sending the processing instruction to the intelligent household appliance, and executing the operation corresponding to the processing instruction by the intelligent household appliance.

The embodiment of the application can set corresponding base identification information aiming at different application scenes in advance, when a user carries a portable intelligent device and the portable intelligent device is connected with a certain target base, the portable intelligent device can acquire the identification information of the target base, the corresponding target application scene is determined according to the identification information, and when a voice signal is received and a processing instruction corresponding to the voice signal is acquired, corresponding operation can be executed according to the processing instruction and the target application scene. Therefore, the user can carry the portable intelligent device to any application scene, and in any scene, the portable intelligent device can identify the scene through the identification information of the target base and carry out intelligent voice interaction in the scene, so that the purpose of carrying out intelligent voice interaction in different scenes through the same device can be achieved. In addition, because the embodiment of the application can share the same set of software and hardware equipment to realize intelligent voice interaction in different application scenes, the consistency of intelligent voice interaction experience in different scenes can be ensured.

The foregoing description of specific embodiments of the present application has been presented. Other embodiments are within the scope of the following claims. In some cases, the actions or steps recited in the claims may be performed in a different order than in the embodiments and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing may also be possible or may be advantageous.

Fig. 3 is a schematic structural diagram of an electronic device according to an embodiment of the present application. Referring to fig. 3, at a hardware level, the electronic device includes a processor, and optionally further includes an internal bus, a network interface, and a memory. The Memory may include a Memory, such as a Random-Access Memory (RAM), and may further include a non-volatile Memory, such as at least 1 disk Memory. Of course, the electronic device may also include hardware required for other services.

The processor, the network interface, and the memory may be connected to each other via an internal bus, which may be an ISA (Industry Standard Architecture) bus, a PCI (Peripheral Component Interconnect) bus, an EISA (Extended Industry Standard Architecture) bus, or the like. The bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, only one double-headed arrow is shown in FIG. 3, but this does not indicate only one bus or one type of bus.

And the memory is used for storing programs. In particular, the program may include program code comprising computer operating instructions. The memory may include both memory and non-volatile storage and provides instructions and data to the processor.

The processor reads the corresponding computer program from the nonvolatile memory to the memory and then runs the computer program to form the intelligent voice interaction device on the logic level. The processor is used for executing the program stored in the memory and is specifically used for executing the following operations:

under the condition of being connected with a target base, acquiring identification information of the target base, and determining a target application scene corresponding to the target base according to the identification information;

receiving a voice signal and acquiring a processing instruction corresponding to the voice signal;

and executing corresponding target operation based on the processing instruction and the target application scene.

The method executed by the intelligent voice interaction device disclosed in the embodiment of fig. 3 of the present application can be applied to or implemented by a processor. The processor may be an integrated circuit chip having signal processing capabilities. In implementation, the steps of the above method may be performed by integrated logic circuits of hardware in a processor or instructions in the form of software. The Processor may be a general-purpose Processor, including a Central Processing Unit (CPU), a Network Processor (NP), and the like; but also Digital Signal Processors (DSPs), Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs) or other Programmable logic devices, discrete Gate or transistor logic devices, discrete hardware components. The various methods, steps, and logic blocks disclosed in the embodiments of the present application may be implemented or performed. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the method disclosed in connection with the embodiments of the present application may be directly implemented by a hardware decoding processor, or implemented by a combination of hardware and software modules in the decoding processor. The software module may be located in ram, flash memory, rom, prom, or eprom, registers, etc. storage media as is well known in the art. The storage medium is located in a memory, and a processor reads information in the memory and completes the steps of the method in combination with hardware of the processor.

The electronic device may further execute the method shown in fig. 1 and fig. 2, and implement the functions of the intelligent voice interaction apparatus in the embodiment shown in fig. 1 and fig. 2, which are not described herein again in this embodiment of the present application.

Of course, besides the software implementation, the electronic device of the present application does not exclude other implementations, such as a logic device or a combination of software and hardware, and the like, that is, the execution subject of the following processing flow is not limited to each logic unit, and may also be hardware or a logic device.

Embodiments of the present application also provide a computer-readable storage medium storing one or more programs, where the one or more programs include instructions, which when executed by a portable electronic device including a plurality of application programs, enable the portable electronic device to perform the method of the embodiment shown in fig. 1, and are specifically configured to:

under the condition of being connected with a target base, acquiring identification information of the target base, and determining a target application scene corresponding to the target base according to the identification information;

receiving a voice signal and acquiring a processing instruction corresponding to the voice signal;

and executing corresponding target operation based on the processing instruction and the target application scene.

Fig. 4 is a schematic structural diagram of an intelligent voice interaction device 40 according to an embodiment of the present application. Referring to fig. 4, in a software implementation, the smart voice interaction device 40 may include: a scene determination unit 41, an instruction acquisition unit 42, and an instruction execution unit 43, wherein:

the scene determining unit 41 is configured to, in a case where the target base is connected to the target base, obtain identification information of the target base, and determine a target application scene corresponding to the target base according to the identification information;

an instruction acquisition unit 42 that receives a voice signal and acquires a processing instruction corresponding to the voice signal;

and an instruction execution unit 43, configured to execute a corresponding target operation based on the processing instruction and the target application scenario.

Optionally, the identification information includes at least one of:

a device identification of the target base;

a designated field predefined in a communication protocol of the target base and the portable smart device.

Optionally, the target application scenario includes an in-vehicle scenario and a home scenario.

Optionally, when it is determined that the target application scene is the in-vehicle scene according to the identification information, the instruction executing unit 43 executes a corresponding target operation based on the processing instruction and the target application scene, including:

judging the instruction type of the processing instruction in the vehicle-mounted scene;

and executing the target operation corresponding to the instruction type.

Optionally, the instruction types include a first type and a second type; the step of determining the instruction type of the processing instruction in the vehicle-mounted scene by the instruction execution unit 43 includes:

judging whether an execution object of the processing instruction is a preset execution object or not;

if so, determining the instruction type of the processing instruction to be the first type;

if not, determining that the instruction type of the processing instruction is the second type.

Optionally, the first type of instruction is executed by an in-vehicle central control system, and the second type of instruction is executed by the portable intelligent device;

wherein the instruction execution unit 43 executes a target operation corresponding to the instruction type, including:

when the instruction type is the first type, sending the processing instruction to the vehicle-mounted central control system, and executing an operation corresponding to the processing instruction by the vehicle-mounted central control system;

and executing the operation corresponding to the processing instruction when the instruction type is the second type.

Optionally, when it is determined that the target application scenario is the home scenario according to the identification information, the instruction executing unit 43 executes a corresponding target operation based on the processing instruction and the target application scenario, including:

determining the intelligent household appliances matched with the processing instruction in the family scene;

and sending the processing instruction to the intelligent household appliance, and executing the operation corresponding to the processing instruction by the intelligent household appliance.

The intelligent voice interaction apparatus 40 provided in this embodiment of the present application may also execute the method in fig. 1 and fig. 2, and implement the functions of the intelligent voice interaction apparatus in the embodiment shown in fig. 1 and fig. 2, which are not described herein again.

Fig. 5 is a schematic structural diagram of an interactive system 50 for smart voice according to an embodiment of the present application. The interactive system 50 comprises a portable smart device 51 and a target base 52, wherein:

under the condition that the portable intelligent device 51 is connected with the target base 52, the portable intelligent device 51 acquires identification information of the target base 52, and determines a target application scene corresponding to the target base 52 according to the identification information;

the portable intelligent device 51 receives a voice signal and acquires a processing instruction corresponding to the voice signal;

the portable intelligent device 51 executes corresponding target operation based on the processing instruction and the target application scenario.

In this embodiment, the portable intelligent device 51 may implement the functions implemented by the portable intelligent device in the embodiment shown in fig. 1 and fig. 2, and the target base 52 may implement the functions implemented by the target base in the embodiment shown in fig. 1 and fig. 2, which are not described again here.

In short, the above description is only a preferred embodiment of the present application, and is not intended to limit the scope of the present application. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.

The systems, devices, modules or units illustrated in the above embodiments may be implemented by a computer chip or an entity, or by a product with certain functions. One typical implementation device is a computer. In particular, the computer may be, for example, a personal computer, a laptop computer, a cellular telephone, a camera phone, a smartphone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or a combination of any of these devices.

Computer-readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, a computer readable medium does not include a transitory computer readable medium such as a modulated data signal and a carrier wave.

It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.

The embodiments in the present application are described in a progressive manner, and the same and similar parts among the embodiments can be referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, for the system embodiment, since it is substantially similar to the method embodiment, the description is simple, and for the relevant points, reference may be made to the partial description of the method embodiment.

完整详细技术资料下载
上一篇:石墨接头机器人自动装卡簧、装栓机
下一篇:一种端到端实时语音合成方法

网友询问留言

已有0条留言

还没有人留言评论。精彩留言会获得点赞!

精彩留言,会给你点赞!