CN103929532A - Information processing method and electronic equipment - Google Patents
Information processing method and electronic equipment Download PDFInfo
- Publication number
- CN103929532A CN103929532A CN201410101297.0A CN201410101297A CN103929532A CN 103929532 A CN103929532 A CN 103929532A CN 201410101297 A CN201410101297 A CN 201410101297A CN 103929532 A CN103929532 A CN 103929532A
- Authority
- CN
- China
- Prior art keywords
- voice data
- instruction
- recognition result
- preset condition
- voice
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 12
- 238000003672 processing method Methods 0.000 title claims abstract description 12
- 238000004891 communication Methods 0.000 claims abstract description 96
- 238000000034 method Methods 0.000 claims abstract description 40
- 230000004044 response Effects 0.000 claims description 29
- 238000012545 processing Methods 0.000 claims description 20
- 238000004088 simulation Methods 0.000 claims description 14
- 238000001914 filtration Methods 0.000 claims description 6
- 238000000605 extraction Methods 0.000 claims description 4
- 230000003993 interaction Effects 0.000 description 15
- 230000008569 process Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 238000013518 transcription Methods 0.000 description 4
- 230000035897 transcription Effects 0.000 description 4
- 230000001960 triggered effect Effects 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000013499 data model Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
Landscapes
- Telephonic Communication Services (AREA)
- Telephone Function (AREA)
Abstract
The invention provides an information processing method and electronic equipment. The method is applied to the first electronic equipment. The first electronic equipment comprises a sound collection unit. The method comprises the steps of obtaining a first instruction; responding to the first instruction to remind a user to respond to a communication request from the second electronic equipment; collecting voice data through the sound collection unit; responding to the communication request through the voice data; identifying the voice data to obtain a first identification result; generating a second instruction under the condition that the first identification result meets a preset condition; responding to the second instruction to enable the first electronic equipment to respond to the communication request; and sending the voice data to the second electronic equipment.
Description
Technical Field
The present invention relates to communications technologies, and in particular, to an information processing method and an electronic device.
Background
In the process of implementing the technical solution of the embodiment of the present application, the inventor of the present application finds at least the following technical problems in the prior art:
currently, when a user responds to a communication request initiated by another user, for example, when the communication request is a call answering scene, the user usually responds to the call answering by a user operation such as a key operation, for example, ringing a mobile phone, and when the user is reminded to answer the call, the user can press a physical key on the mobile phone or a virtual key on a screen of the mobile phone to respond to the call answering. However, this response method is not suitable for the user, and a scenario of the user's requirement is that the user is driving a car, and in this scenario, it is inconvenient for the user to respond to the answer by pressing a button, how to respond to the scenario? In the related art, there is no effective solution to this problem.
Disclosure of Invention
In view of the above, embodiments of the present invention are directed to an information processing method and an electronic device, which can directly trigger a response to a communication request without a user operation.
The technical scheme of the embodiment of the invention is realized as follows:
the information processing method of the embodiment of the invention is applied to first electronic equipment, wherein the first electronic equipment comprises a sound acquisition unit, and the method comprises the following steps:
acquiring a first instruction;
responding to the first instruction to remind a user to respond to a communication request from a second electronic device;
collecting voice data through the sound collection unit; the voice data is used for responding to the communication request;
recognizing the voice data to obtain a first recognition result;
if the first recognition result meets a preset condition, generating a second instruction;
responding to the second instruction to enable the first electronic equipment to respond to the communication request;
and sending the voice data to the second electronic equipment.
Preferably, the method further comprises:
extracting user habit expression mode information and storing the information in a voice database; wherein,
recognizing the voice data to obtain a first recognition result;
if the first recognition result meets a preset condition, generating a second instruction, specifically comprising:
judging whether the first recognition result exists in the voice database;
and if so, the first identification result meets the preset condition, and the second instruction is generated.
Preferably, the first and second substrates are, among others,
recognizing the voice data to obtain a first recognition result;
if the first recognition result meets a preset condition, generating a second instruction, specifically comprising:
performing voiceprint recognition on the voice data to obtain a voiceprint recognition result;
and judging whether the voiceprint recognition result meets a second preset condition, if so, judging that the first recognition result meets the preset condition and generating the second instruction.
Preferably, the method further comprises:
if the first instruction is used for requesting a user to respond in a mode of including audio, mixed data containing the voice data and the audio is obtained when the voice data is collected through the sound collection unit;
and determining incoming call audio, and filtering the incoming call audio from the mixed data to obtain the voice data.
Preferably, if the first result meets a preset condition, generating a second instruction specifically includes:
and analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is successful, determining that the first recognition result meets the preset condition, and generating the second instruction for receiving the communication request.
Preferably, the method further comprises: and analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is unsuccessful, determining that the first recognition result does not accord with the preset condition, and generating a third instruction for rejecting the communication request.
Preferably, the preset condition includes a signal-to-noise ratio, and/or an identification content.
An electronic device according to an embodiment of the present invention includes:
a first obtaining unit configured to obtain a first instruction;
the first response unit is used for responding to the first instruction so as to remind a user of responding to a communication request from the second electronic equipment;
the voice acquisition unit is used for acquiring voice data; the voice data is used for responding to the communication request;
the first processing unit is used for identifying the voice data to obtain a first identification result; if the first recognition result meets a preset condition, generating a second instruction;
a second response unit, configured to respond to the second instruction, so as to enable the first electronic device to respond to the communication request;
and the sending unit is used for sending the voice data to the second electronic equipment.
Preferably, the electronic device further includes:
the extraction unit is used for extracting the habit expression mode information of the user and storing the habit expression mode information in a voice database;
the first processing unit is further configured to determine whether the first recognition result exists in the voice database; and if so, the first identification result meets the preset condition, and the second instruction is generated.
Preferably, the first processing unit is further configured to perform voiceprint recognition on the voice data to obtain a voiceprint recognition result; and judging whether the voiceprint recognition result meets a second preset condition, if so, judging that the first recognition result meets the preset condition and generating the second instruction.
Preferably, the electronic device further includes:
the second processing unit is used for acquiring mixed data containing the voice data and the audio when the voice data is acquired by the sound acquisition unit if the first instruction is used for requesting a user to respond in a mode of including the audio; and determining incoming call audio, and filtering the incoming call audio from the mixed data to obtain the voice data.
Preferably, the first processing unit is further configured to analyze the first recognition result, match the first recognition result with data in the voice data simulation library, determine that the first recognition result meets the preset condition when matching is successful, and generate the second instruction for accepting the communication request.
Preferably, the electronic device further includes:
and the third processing unit is used for analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is unsuccessful, determining that the first recognition result does not accord with the preset condition and generating a third instruction for rejecting the communication request.
Preferably, the preset condition includes a signal-to-noise ratio, and/or an identification content.
The information processing method of the embodiment of the invention is applied to first electronic equipment, wherein the first electronic equipment comprises a sound acquisition unit, and the method comprises the following steps: acquiring a first instruction; responding to the first instruction to remind a user to respond to a communication request from a second electronic device; collecting voice data through the sound collection unit; the voice data is used for responding to the communication request; recognizing the voice data to obtain a first recognition result; if the first recognition result meets a preset condition, generating a second instruction; responding to the second instruction to enable the first electronic equipment to respond to the communication request; and sending the voice data to the second electronic equipment.
By adopting the embodiment of the invention, the voice data can be collected by the sound collection unit; the voice data is used for responding to the communication request; recognizing the voice data to obtain a first recognition result; if the first recognition result meets a preset condition, generating a second instruction; responding to the second instruction to enable the first electronic equipment to respond to the communication request; and sending the voice data to the second electronic equipment, so that the voice data is directly sent to respond to the communication request and the communication connection is directly established without any manual operation of a user, such as key operation, for triggering response.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the provided drawings without creative efforts.
FIG. 1 is a schematic flow chart of an implementation of an embodiment of the method of the present invention;
FIG. 2 is a schematic flow chart of an implementation of the embodiment of the method of the present invention;
FIG. 3 is a schematic flow chart of an implementation of the embodiment of the method of the present invention;
FIG. 4 is a schematic flow chart of an implementation of the embodiment of the method of the present invention;
FIG. 5 is a schematic diagram of a component structure of an embodiment of an electronic device according to the invention;
FIG. 6 is a schematic diagram of a component structure of an embodiment of an electronic device according to the invention;
fig. 7 is a schematic structural diagram of an electronic device according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the technical solutions in the embodiments of the present invention will be described clearly and completely with reference to the accompanying drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention. In the present application, the embodiments and features of the embodiments may be arbitrarily combined with each other without conflict. The steps illustrated in the flow charts of the figures may be performed in a computer system such as a set of computer-executable instructions. Also, while a logical order is shown in the flow diagrams, in some cases, the steps shown or described may be performed in an order different than here.
The following describes the embodiments in further detail with reference to the accompanying drawings.
The first embodiment of the method comprises the following steps:
an information processing method according to an embodiment of the present invention is applied to a first electronic device, where the first electronic device includes a sound collection unit, and as shown in fig. 1, the method includes:
step 101, obtaining a first instruction.
And 102, responding to the first instruction to remind a user to respond to a communication request from the second electronic equipment.
Step 103, acquiring voice data through the voice acquisition unit; the voice data is used to respond to the communication request.
And step 104, recognizing the voice data to obtain a first recognition result.
And 105, if the first recognition result meets a preset condition, generating a second instruction.
And step 106, responding to the second instruction, so that the first electronic equipment responds to the communication request.
And step 107, sending the voice data to the second electronic equipment.
By adopting the embodiment of the invention, the voice data can be collected by the voice collecting unit in step 103; the voice data is used for responding to the communication request; the voice data can be identified through 104-106 to obtain a first identification result; if the first recognition result meets a preset condition, generating a second instruction; responding to the second instruction to enable the first electronic equipment to respond to the communication request; the voice data can be sent to the second electronic device through step 107, so that no manual operation of the user, such as key operation, is needed to trigger the response, but the voice data is directly sent to respond to the communication request, and the communication connection is directly established.
For example, taking a mobile phone to answer a call as an example, in the prior art, when a ring tone or a vibration prompt is obtained and an incoming call needs to be answered, a physical key or a virtual key for answering the call on the mobile phone is pressed to trigger a response to the ring tone or the vibration prompt, and then voice data is sent out, such as "hello, i is one", to establish a communication connection, and start a call between users. When the embodiment of the invention is adopted to obtain the ring tone or the vibration prompt and answer the incoming call, the response to the ring tone or the vibration prompt triggered by the user operation is omitted, and the voice data is directly sent to establish the communication connection to start the communication between the users. Therefore, the voice interaction process is optimized, and the voice interaction efficiency is improved.
The second method embodiment:
an information processing method according to an embodiment of the present invention is applied to a first electronic device, where the first electronic device includes a sound collection unit, as shown in fig. 2, and the method includes:
step 201, a first instruction is obtained.
Step 202, responding to the first instruction, and reminding a user to respond to a communication request from the second electronic device.
Step 203, acquiring voice data through the sound acquisition unit; the voice data is used to respond to the communication request.
And step 204, extracting the user habit expression mode information and storing the information in a voice database.
Step 205, determining whether the first recognition result exists in the voice database, if so, the first recognition result meets the preset condition, and generating the second instruction.
Step 206, responding to the second instruction, so that the first electronic device responds to the communication request.
And step 207, sending the voice data to the second electronic equipment.
By adopting the embodiment of the invention, the voice data can be collected by the sound collection unit in step 203; the voice data is used for responding to the communication request; the user habit expression mode information can be extracted through 204 and stored in a voice database, and the expression mode of the user habit is simulated through the mode of collecting historical records and automatically extracting so as to form the expression mode of the user habit and stored in the voice data simulation database, so that the personalized customization of the voice data is realized; through the step 205 and the step 206, it can be determined whether the first recognition result exists in the voice database, and if so, the first recognition result meets the preset condition, and the second instruction is generated; responding to the second instruction to enable the first electronic equipment to respond to the communication request; the voice data can be sent to the second electronic device through step 207, so that no manual operation of the user, such as a key operation, is needed to trigger the response, but the voice data is directly sent to respond to the communication request, and the communication connection is directly established.
It should be noted that the communication connection established by the embodiment of the present invention does not have the existing communication for making a call, and also includes the communication connection established by using an IM tool such as WeChat, QQ, etc. based on the internet communication mode.
For example, taking a mobile phone to answer a call as an example, in the prior art, when a ring tone or a vibration prompt is obtained and an incoming call needs to be answered, a physical key or a virtual key for answering the call on the mobile phone is pressed to trigger a response to the ring tone or the vibration prompt, and then voice data is sent out, such as "hello, i is one", to establish a communication connection, and start a call between users. When the embodiment of the invention is adopted to obtain the ring tone or the vibration prompt and answer the incoming call, the response to the ring tone or the vibration prompt triggered by the user operation is omitted, and the voice data is directly sent to establish the communication connection to start the communication between the users. Therefore, the voice interaction process is optimized, and the voice interaction efficiency is improved.
Because the embodiment of the invention can extract the user habit expression mode information through the step 204 and store the information in the voice database so as to compare and identify the information in the following steps 205 and 206, the identification rate can be improved, and the voice interaction is optimized, and the following scene of answering the call is described as follows:
and the voice data is directly sent out to answer the call in response to the ring or the vibration prompt, and can be defined as common greetings such as 'feed your good' and 'feed your place' which accord with the call receiving habit of the user, so that the use of the user is more natural. After the call is put through by the greeting, the recording corresponding to the voice data (such as 'feed your good', 'feed your place', etc.) can be sent to the opposite user, so that the interactive voice can be multiplexed. Through learning (recording) of the speech idioms when the user calls, the idioms of the user, such as 'which place to feed and please talk', are automatically extracted, so that the personalized customization of the speech data is realized; and through the voice and the corresponding recognition text, a special model of a specific user is trained, so that the difference between a speaker and a channel is made up, the model matching is more accurate, and the recognition accuracy is higher.
By adopting the voice interaction mode, the call receiving command words are defined through the habit greetings of the user, so that the interaction is more natural, and the self-adaptation to the habit words of the user is realized through the collection of the habit words of the user and the corresponding voices, so that the usage is more in line with the habit of the user; through training corresponding to the voice and the model, the difference between the speaker and the channel is avoided, so that the model matching is more accurate, and the recognition rate is better.
The third method embodiment:
an information processing method according to an embodiment of the present invention is applied to a first electronic device, where the first electronic device includes a sound collection unit, as shown in fig. 3, and the method includes:
step 301, a first instruction is obtained.
Step 302, responding to the first instruction, so as to remind the user to respond to the communication request from the second electronic equipment.
Step 303, acquiring voice data through the sound acquisition unit; the voice data is used to respond to the communication request.
And 304, carrying out voiceprint recognition on the voice data to obtain a voiceprint recognition result.
Step 305, determining whether the voiceprint recognition result meets a second preset condition, if the determination result indicates that the voiceprint recognition result meets the second preset condition, the first recognition result meets the preset condition, and generating the second instruction.
Step 306, responding to the second instruction, so that the first electronic device responds to the communication request.
Step 307, sending the voice data to a second electronic device.
By adopting the embodiment of the invention, the voice data can be collected by the sound collection unit in step 303; the voice data is used for responding to the communication request; the voice print recognition can be carried out on the voice data through 304, and a voice print recognition result is obtained; through the step 305 and the step 306, it can be determined whether the voiceprint recognition result meets a second preset condition, and if the determination result indicates that the voiceprint recognition result meets the second preset condition, the first recognition result meets the preset condition, and the second instruction is generated; responding to the second instruction to enable the first electronic equipment to respond to the communication request; the voice data can be sent to the second electronic device through step 307, so that no manual operation of the user, such as a key operation, is needed to trigger the response, but the voice data is directly sent to respond to the communication request, and the communication connection is directly established.
For example, taking a mobile phone to answer a call as an example, in the prior art, when a ring tone or a vibration prompt is obtained and an incoming call needs to be answered, a physical key or a virtual key for answering the call on the mobile phone is pressed to trigger a response to the ring tone or the vibration prompt, and then voice data is sent out, such as "hello, i is one", to establish a communication connection, and start a call between users. When the embodiment of the invention is adopted to obtain the ring tone or the vibration prompt and answer the incoming call, the response to the ring tone or the vibration prompt triggered by the user operation is omitted, and the voice data is directly sent to establish the communication connection to start the communication between the users. Therefore, the voice interaction process is optimized, and the voice interaction efficiency is improved.
Because the voice print recognition can be performed on the voice data through the step 304 to obtain the voice print recognition result, the following steps 305 and 306 can be compared and recognized conveniently, the recognition rate can be improved, and the voice interaction can be optimized. The principle of voiceprint recognition is similar to that of fingerprint recognition, feature information is extracted, and the embodiment of the invention extracts the feature information of voice for recognition, for example, the voice frequency of women is usually higher, the voice frequency of men is usually lower, and then the person can know that the person is a man or a woman calling through the specific voice frequency.
The method comprises the following steps:
an information processing method according to an embodiment of the present invention is applied to a first electronic device, where the first electronic device includes a sound collection unit, as shown in fig. 4, and the method includes:
step 401, a first instruction is obtained.
Step 402, responding to the first instruction to remind a user to respond to a communication request from the second electronic device.
Step 403, if the first instruction is used to request a user to respond in a manner including audio, acquiring mixed data including the audio and the voice data when the voice data is acquired by the sound acquisition unit; the voice data is used to respond to the communication request.
And step 404, extracting the habit expression mode information of the user and storing the habit expression mode information in a voice database.
Step 405, determining whether the first recognition result exists in the voice database, if so, the first recognition result meets the preset condition, and generating the second instruction.
And step 406, responding to the second instruction, so that the first electronic device responds to the communication request.
Step 407, determining an incoming call audio, and filtering the incoming call audio from the mixed data to obtain the voice data.
And step 408, sending the voice data to the second electronic equipment.
With the embodiment of the present invention, in step 403, when the first instruction is used to request a user to respond in a manner including audio, and the voice data is collected by the sound collection unit, mixed data including the voice data and the audio is obtained; the voice data is used for responding to the communication request; the user habit expression mode information can be extracted through 404 and stored in a voice database, and the expression mode of the user habit is simulated through the mode of collecting historical records and automatically extracting so as to form the expression mode of the user habit and stored in a voice data simulation database so as to realize the personalized customization of the voice data; whether the first recognition result exists in the voice database can be judged through the step 405 and the step 406, and if so, the first recognition result meets the preset condition and the second instruction is generated; responding to the second instruction to enable the first electronic equipment to respond to the communication request; the incoming call audio can be determined through step 407, and the incoming call audio is filtered from the mixed data to obtain the voice data; the voice data can be sent to the second electronic device through step 408, so that no manual operation of the user, such as a key operation, is needed to trigger the response, but the voice data is directly sent to respond to the communication request, and the communication connection is directly established.
For example, taking a mobile phone to answer a call as an example, in the prior art, when a ring tone or a vibration prompt is obtained and an incoming call needs to be answered, a physical key or a virtual key for answering the call on the mobile phone is pressed to trigger a response to the ring tone or the vibration prompt, and then voice data is sent out, such as "hello, i is one", to establish a communication connection, and start a call between users. When the embodiment of the invention is adopted to obtain the ring tone or the vibration prompt and answer the incoming call, the response to the ring tone or the vibration prompt triggered by the user operation is omitted, and the voice data is directly sent to establish the communication connection to start the communication between the users. Therefore, the voice interaction process is optimized, and the voice interaction efficiency is improved.
Because the embodiment of the invention can extract the user habit expression mode information through the step 404 and store the user habit expression mode information in the voice database, so that the comparison and identification can be carried out in the following steps 405 and 406, the identification rate can be improved, the voice interaction can be optimized, and the following scene of answering the call is described as follows:
and the voice data is directly sent out to answer the call in response to the ring or the vibration prompt, and can be defined as common greetings such as 'feed your good' and 'feed your place' which accord with the call receiving habit of the user, so that the use of the user is more natural. After the call is put through by the greeting, the recording corresponding to the voice data (such as 'feed your good', 'feed your place', etc.) can be sent to the opposite user, so that the interactive voice can be multiplexed. Through learning (recording) of the speech idioms when the user calls, the idioms of the user, such as 'which place to feed and please talk', are automatically extracted, so that the personalized customization of the speech data is realized; and through the voice and the corresponding recognition text, a special model of a specific user is trained, so that the difference between a speaker and a channel is made up, the model matching is more accurate, and the recognition accuracy is higher.
By adopting the voice interaction mode, the call receiving command words are defined through the habit greetings of the user, so that the interaction is more natural, and the self-adaptation to the habit words of the user is realized through the collection of the habit words of the user and the corresponding voices, so that the usage is more in line with the habit of the user; through training corresponding to the voice and the model, the difference between the speaker and the channel is avoided, so that the model matching is more accurate, and the recognition rate is better.
In addition, when driving, the user is reminded of the existence of a call by using a mobile phone ring or a voice engine (TTS) in voice navigation, and the like, and at this time, if the user directly uses voice data to respond, the call is answered or hung up, and the recognition rate of the voice data is sharply reduced due to the mixing of the incoming call audio data of the ring and the voice data of the user. Even if the user uses the echo cancellation scheme to cancel the outgoing incoming call audio as much as possible, the following problems still cannot be solved:
because the initial incoming call audio frequency of the outgoing call has large deformation after passing through the loudspeaker, sound propagation and mic transcription on the mobile phone, and the distortion function (or called as a data model) of the deformation is inconsistent due to the difference of the relative positions of the loudspeaker and the mic of different mobile phones, the difference between the incoming call audio frequency for echo cancellation and the initial incoming call audio frequency of the actual outgoing transcription is large, so that the echo cancellation is inaccurate, the voice distortion of the echo cancellation is large, and the voice recognition rate is greatly reduced. In view of this, it becomes important how to eliminate the influence of the user's ringtone and TTS to improve the voice recognition rate.
With the embodiment of the present invention, the incoming call audio in step 407 is actually the incoming call audio after the deformation, and is not the initial audio that requests the user to respond in step 403. Because the embodiment of the invention can determine that the deformed incoming call audio is used for echo cancellation, the incoming call audio is filtered from the mixed data to obtain the voice command, and the recognition accuracy is higher, therefore, the interference of the incoming call audio to the voice command is avoided, the voice command can be correctly recognized, and the recognition rate is greatly improved. That is to say, through the transcription signal acquisition or distortion estimation of the telephone prompt tone, the distortion caused by a loudspeaker, sound transmission and mic transcription is avoided during echo cancellation, so that the voice recognition rate is improved, and the mobile phone can be efficiently answered or hung up by using voice data control.
In a preferred implementation manner of the embodiment of the present invention, if the first result meets a preset condition, the generating a second instruction specifically includes: and analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is successful, determining that the first recognition result meets the preset condition, and generating the second instruction for receiving the communication request.
In a preferred implementation manner of the embodiment of the present invention, the method further includes: and analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is unsuccessful, determining that the first recognition result does not accord with the preset condition, and generating a third instruction for rejecting the communication request.
In a preferred implementation manner of the embodiment of the present invention, the preset condition includes a signal-to-noise ratio and/or an identification content.
Here, it should be noted that: the following description of the electronic device items is similar to the description of the method, and the description of the beneficial effects of the method is omitted for brevity. For technical details not disclosed in the embodiments of the electronic device of the present invention, refer to the description of the embodiments of the method of the present invention.
The first embodiment of the electronic device:
as shown in fig. 5, the electronic device according to the embodiment of the present invention includes:
a first obtaining unit configured to obtain a first instruction;
the first response unit is used for responding to the first instruction so as to remind a user of responding to a communication request from the second electronic equipment;
the voice acquisition unit is used for acquiring voice data; the voice data is used for responding to the communication request;
the first processing unit is used for identifying the voice data to obtain a first identification result; if the first recognition result meets a preset condition, generating a second instruction;
a second response unit, configured to respond to the second instruction, so as to enable the first electronic device to respond to the communication request;
and the sending unit is used for sending the voice data to the second electronic equipment.
Second embodiment of the electronic device:
as shown in fig. 6, the electronic device according to the embodiment of the present invention includes:
a first obtaining unit configured to obtain a first instruction;
the first response unit is used for responding to the first instruction so as to remind a user of responding to a communication request from the second electronic equipment;
the voice acquisition unit is used for acquiring voice data; the voice data is used for responding to the communication request;
the extraction unit is used for extracting the habit expression mode information of the user and storing the habit expression mode information in a voice database;
the first processing unit is used for identifying the voice data to obtain a first identification result; judging whether the first recognition result exists in the voice database; if the first identification result meets the preset condition, generating the second instruction;
a second response unit, configured to respond to the second instruction, so as to enable the first electronic device to respond to the communication request;
and the sending unit is used for sending the voice data to the second electronic equipment.
Electronic equipment embodiment three:
as shown in fig. 5, the electronic device according to the embodiment of the present invention includes:
a first obtaining unit configured to obtain a first instruction;
the first response unit is used for responding to the first instruction so as to remind a user of responding to a communication request from the second electronic equipment;
the voice acquisition unit is used for acquiring voice data; the voice data is used for responding to the communication request;
the first processing unit is used for carrying out voiceprint recognition on the voice data to obtain a voiceprint recognition result; judging whether the voiceprint recognition result meets a second preset condition, if so, judging that the first recognition result meets the preset condition and generating a second instruction;
a second response unit, configured to respond to the second instruction, so as to enable the first electronic device to respond to the communication request;
and the sending unit is used for sending the voice data to the second electronic equipment.
The fourth embodiment of the electronic device:
as shown in fig. 7, the electronic device according to the embodiment of the present invention includes:
a first obtaining unit configured to obtain a first instruction;
the first response unit is used for responding to the first instruction so as to remind a user of responding to a communication request from the second electronic equipment;
the voice acquisition unit is used for acquiring voice data; the voice data is used for responding to the communication request;
the extraction unit is used for extracting the habit expression mode information of the user and storing the habit expression mode information in a voice database;
the first processing unit is used for identifying the voice data to obtain a first identification result; judging whether the first recognition result exists in the voice database; if the first identification result meets the preset condition, generating the second instruction;
a second response unit, configured to respond to the second instruction, so as to enable the first electronic device to respond to the communication request;
the second processing unit is used for acquiring mixed data containing the voice data and the audio when the voice data is acquired by the sound acquisition unit if the first instruction is used for requesting a user to respond in a mode of including the audio; determining incoming call audio, and filtering the incoming call audio from the mixed data to obtain the voice data;
and the sending unit is used for sending the voice data to the second electronic equipment.
In a preferred implementation manner of the embodiment of the present invention, the first processing unit is further configured to analyze the first recognition result, match the first recognition result with data in the voice data simulation library, determine that the first recognition result meets the preset condition when matching is successful, and generate the second instruction for accepting the communication request.
In a preferred implementation manner of the embodiment of the present invention, the electronic device further includes:
and the third processing unit is used for analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is unsuccessful, determining that the first recognition result does not accord with the preset condition and generating a third instruction for rejecting the communication request.
In a preferred implementation manner of the embodiment of the present invention, the preset condition includes a signal-to-noise ratio and/or an identification content.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. The above-described device embodiments are merely illustrative, for example, the division of the unit is only a logical functional division, and there may be other division ways in actual implementation, such as: multiple units or components may be combined, or may be integrated into another system, or some features may be omitted, or not implemented. In addition, the coupling, direct coupling or communication connection between the components shown or discussed may be through some interfaces, and the indirect coupling or communication connection between the devices or units may be electrical, mechanical or other forms.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed on a plurality of network units; some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, all the functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may be separately regarded as one unit, or two or more units may be integrated into one unit; the integrated unit can be realized in a form of hardware, or in a form of hardware plus a software functional unit.
Those of ordinary skill in the art will understand that: all or part of the steps for implementing the method embodiments may be implemented by hardware related to program instructions, and the program may be stored in a computer readable storage medium, and when executed, the program performs the steps including the method embodiments; and the aforementioned storage medium includes: a mobile storage device, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
Alternatively, the integrated unit of the present invention may be stored in a computer-readable storage medium if it is implemented in the form of a software functional module and sold or used as a separate product. Based on such understanding, the technical solutions of the embodiments of the present invention may be essentially implemented or a part contributing to the prior art may be embodied in the form of a software product, which is stored in a storage medium and includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the methods described in the embodiments of the present invention. And the aforementioned storage medium includes: a mobile storage device, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the appended claims.
Claims (14)
1. An information processing method is applied to a first electronic device, wherein the first electronic device comprises a sound acquisition unit, and the method comprises the following steps:
acquiring a first instruction;
responding to the first instruction to remind a user to respond to a communication request from a second electronic device;
collecting voice data through the sound collection unit; the voice data is used for responding to the communication request;
recognizing the voice data to obtain a first recognition result;
if the first recognition result meets a preset condition, generating a second instruction;
responding to the second instruction to enable the first electronic equipment to respond to the communication request;
and sending the voice data to the second electronic equipment.
2. The method of claim 1, further comprising:
extracting user habit expression mode information and storing the information in a voice database; wherein,
recognizing the voice data to obtain a first recognition result;
if the first recognition result meets a preset condition, generating a second instruction, specifically comprising:
judging whether the first recognition result exists in the voice database;
and if so, the first identification result meets the preset condition, and the second instruction is generated.
3. The method of claim 1, wherein,
recognizing the voice data to obtain a first recognition result;
if the first recognition result meets a preset condition, generating a second instruction, specifically comprising:
performing voiceprint recognition on the voice data to obtain a voiceprint recognition result;
and judging whether the voiceprint recognition result meets a second preset condition, if so, judging that the first recognition result meets the preset condition and generating the second instruction.
4. The method of claim 1, further comprising:
if the first instruction is used for requesting a user to respond in a mode of including audio, mixed data containing the voice data and the audio is obtained when the voice data is collected through the sound collection unit;
and determining incoming call audio, and filtering the incoming call audio from the mixed data to obtain the voice data.
5. The method according to any one of claims 1 to 4, wherein if the first result meets a predetermined condition, generating a second instruction specifically comprises:
and analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is successful, determining that the first recognition result meets the preset condition, and generating the second instruction for receiving the communication request.
6. The method of any of claims 1 to 4, further comprising: and analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is unsuccessful, determining that the first recognition result does not accord with the preset condition, and generating a third instruction for rejecting the communication request.
7. The method of claim 1, the preset condition comprising a signal-to-noise ratio, and/or an identification content.
8. An electronic device, the electronic device comprising:
a first obtaining unit configured to obtain a first instruction;
the first response unit is used for responding to the first instruction so as to remind a user of responding to a communication request from the second electronic equipment;
the voice acquisition unit is used for acquiring voice data; the voice data is used for responding to the communication request;
the first processing unit is used for identifying the voice data to obtain a first identification result; if the first recognition result meets a preset condition, generating a second instruction;
a second response unit, configured to respond to the second instruction, so as to enable the first electronic device to respond to the communication request;
and the sending unit is used for sending the voice data to the second electronic equipment.
9. The electronic device of claim 8, further comprising:
the extraction unit is used for extracting the habit expression mode information of the user and storing the habit expression mode information in a voice database;
the first processing unit is further configured to determine whether the first recognition result exists in the voice database; and if so, the first identification result meets the preset condition, and the second instruction is generated.
10. The electronic device of claim 8, wherein the first processing unit is further configured to perform voiceprint recognition on the voice data to obtain a voiceprint recognition result; and judging whether the voiceprint recognition result meets a second preset condition, if so, judging that the first recognition result meets the preset condition and generating the second instruction.
11. The electronic device of claim 8, further comprising:
the second processing unit is used for acquiring mixed data containing the voice data and the audio when the voice data is acquired by the sound acquisition unit if the first instruction is used for requesting a user to respond in a mode of including the audio; and determining incoming call audio, and filtering the incoming call audio from the mixed data to obtain the voice data.
12. The electronic device according to any one of claims 8 to 11, wherein the first processing unit is further configured to parse the first recognition result, perform matching with data in the voice data simulation library, determine that the first recognition result meets the preset condition when matching is successful, and generate the second instruction for accepting the communication request.
13. The electronic device of any of claims 8-11, further comprising:
and the third processing unit is used for analyzing the first recognition result, matching the first recognition result with the data in the voice data simulation library, and when the matching is unsuccessful, determining that the first recognition result does not accord with the preset condition and generating a third instruction for rejecting the communication request.
14. The electronic device of claim 8, the preset condition comprising a signal-to-noise ratio, and/or an identification content.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410101297.0A CN103929532A (en) | 2014-03-18 | 2014-03-18 | Information processing method and electronic equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410101297.0A CN103929532A (en) | 2014-03-18 | 2014-03-18 | Information processing method and electronic equipment |
Publications (1)
Publication Number | Publication Date |
---|---|
CN103929532A true CN103929532A (en) | 2014-07-16 |
Family
ID=51147603
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410101297.0A Pending CN103929532A (en) | 2014-03-18 | 2014-03-18 | Information processing method and electronic equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103929532A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104952447A (en) * | 2015-04-30 | 2015-09-30 | 深圳市全球锁安防系统工程有限公司 | Intelligent wearing equipment for safety and health service for old people and voice recognition method |
CN106155707A (en) * | 2015-03-23 | 2016-11-23 | 联想(北京)有限公司 | Information processing method and electronic equipment |
CN106997343A (en) * | 2017-03-28 | 2017-08-01 | 联想(北京)有限公司 | Information processing method and equipment |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20060065916A (en) * | 2004-12-10 | 2006-06-14 | 엘지전자 주식회사 | Automatic answering method for mobile terminal |
CN101657033A (en) * | 2008-08-22 | 2010-02-24 | 环达电脑(上海)有限公司 | Portable communication apparatus and method with voice control |
CN103220423A (en) * | 2013-04-10 | 2013-07-24 | 威盛电子股份有限公司 | Voice answering method and mobile terminal device |
CN103634480A (en) * | 2013-12-17 | 2014-03-12 | 百度在线网络技术(北京)有限公司 | Method and device for communication in mobile communication terminal |
-
2014
- 2014-03-18 CN CN201410101297.0A patent/CN103929532A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20060065916A (en) * | 2004-12-10 | 2006-06-14 | 엘지전자 주식회사 | Automatic answering method for mobile terminal |
CN101657033A (en) * | 2008-08-22 | 2010-02-24 | 环达电脑(上海)有限公司 | Portable communication apparatus and method with voice control |
CN103220423A (en) * | 2013-04-10 | 2013-07-24 | 威盛电子股份有限公司 | Voice answering method and mobile terminal device |
CN103634480A (en) * | 2013-12-17 | 2014-03-12 | 百度在线网络技术(北京)有限公司 | Method and device for communication in mobile communication terminal |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106155707A (en) * | 2015-03-23 | 2016-11-23 | 联想(北京)有限公司 | Information processing method and electronic equipment |
CN106155707B (en) * | 2015-03-23 | 2020-02-21 | 联想(北京)有限公司 | Information processing method and electronic equipment |
CN104952447A (en) * | 2015-04-30 | 2015-09-30 | 深圳市全球锁安防系统工程有限公司 | Intelligent wearing equipment for safety and health service for old people and voice recognition method |
CN104952447B (en) * | 2015-04-30 | 2020-03-27 | 深圳市全球锁安防系统工程有限公司 | Intelligent wearable device for elderly people's health service and voice recognition method |
CN106997343A (en) * | 2017-03-28 | 2017-08-01 | 联想(北京)有限公司 | Information processing method and equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103888581B (en) | A kind of communication terminal and its method for recording call-information | |
CN103856624B (en) | Identify method and the mobile terminal of identity | |
WO2017161741A1 (en) | Method and device for communicating information with deaf-mutes, smart terminal | |
CN102984666B (en) | Address list voice information processing method in a kind of communication process and system | |
CN101485188A (en) | Method and system for providing voice analysis service, and apparatus therefor | |
CN107978320A (en) | One kind call method of adjustment, device, equipment and medium | |
CN103841272A (en) | Method and device for sending voice messages | |
CN104954532B (en) | The method and device and mobile terminal of speech recognition | |
CN106713575A (en) | Method and system of recording contact information in cellphone call | |
CN109151148B (en) | Call content recording method, device, terminal and computer readable storage medium | |
US20230247131A1 (en) | Presentation of communications | |
CN104851423B (en) | Sound information processing method and device | |
CN106953962B (en) | A kind of call recording method and device | |
CN103929532A (en) | Information processing method and electronic equipment | |
CN111970295B (en) | Multi-terminal-based call transaction management method and device | |
EP2913822B1 (en) | Speaker recognition | |
CN108924330A (en) | Information processing method, device and electronic equipment | |
CN104348948B (en) | Information processing method and electronic equipment | |
US8861699B2 (en) | Electronic display method and device | |
CN106453882A (en) | Information processing method and device and electronic equipment | |
CN110232919A (en) | Real-time voice stream extracts and speech recognition system and method | |
CN104079703B (en) | A kind of information processing method and electronic equipment | |
CN112863511B (en) | Signal processing method, device and storage medium | |
CN105374364B (en) | Signal processing method and electronic equipment | |
CN110708418B (en) | Method and device for identifying attributes of calling party |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20140716 |
|
RJ01 | Rejection of invention patent application after publication |