CN109240639A - Acquisition methods, device, storage medium and the terminal of audio data - Google Patents
Acquisition methods, device, storage medium and the terminal of audio data Download PDFInfo
- Publication number
- CN109240639A CN109240639A CN201811000969.3A CN201811000969A CN109240639A CN 109240639 A CN109240639 A CN 109240639A CN 201811000969 A CN201811000969 A CN 201811000969A CN 109240639 A CN109240639 A CN 109240639A
- Authority
- CN
- China
- Prior art keywords
- audio data
- obtains
- signal
- audio
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/165—Management of the audio stream, e.g. setting of volume, audio stream path
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The embodiment of the present application discloses acquisition methods, device, storage medium and the terminal of a kind of audio data, this method comprises: receiving audio data obtains signal;Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;Denoising is carried out to first audio data, obtains target audio data, can be improved the accuracy that intelligent wearable device obtains audio data.
Description
Technical field
The invention relates to the acquisition methods of intelligent terminal technical field more particularly to audio data, device, storages
Medium and terminal.
Background technique
With the continuous development of intelligent wearable device, have the function of intelligent terminal, and is able to satisfy portability and large screen view
Feel that the intelligent wearable device of demand for experience gradually obtains the approval of user.
Currently, intelligent wearable device in noisy environment, can not accurately obtain the audio data of user's sending, and then influence
The normal use of user.
Summary of the invention
The purpose of the embodiment of the present application is to provide acquisition methods, device, storage medium and the terminal of a kind of audio data, can
To improve the accuracy that intelligent wearable device obtains audio data.
In a first aspect, the embodiment of the present application provides a kind of acquisition methods of audio data, comprising:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, then the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Second aspect, the embodiment of the present application provide a kind of acquisition device of audio data, comprising:
Signal receiving module obtains signal for receiving audio data;
First audio obtains module, and the audio data for being received according to the signal receiving module obtains letter
Number, the first audio data is obtained by osteoacusis mode;
Denoising module is denoised for obtaining first audio data that module obtains to first audio
Processing, obtains target audio data.
The third aspect, the embodiment of the present application provide a kind of computer readable storage medium, are stored thereon with computer journey
Sequence realizes the acquisition methods of the audio data as shown in first aspect when the program is executed by processor.
Fourth aspect, the embodiment of the present application provide a kind of terminal, including memory, and processor and storage are on a memory
And the computer program that can be run in processor, the processor are realized as shown in first aspect when executing the computer program
Audio data acquisition methods.
The acquisition scheme of the audio data provided in the embodiment of the present application, first reception audio data obtain signal;Then,
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;Finally, to the first audio number
According to denoising is carried out, target audio data are obtained, can be improved the accuracy that intelligent wearable device obtains audio data.
Detailed description of the invention
Fig. 1 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application;
Fig. 2 is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 3 A is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 3 B is that user provided by the embodiments of the present application passes through intelligent wearable device and Remote progress remote guide
Schematic diagram of a scenario;
Fig. 4 is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 5 is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 6 is a kind of structural schematic diagram of the acquisition device of audio data provided by the embodiments of the present application;
Fig. 7 is a kind of structural schematic diagram of intelligent wearable device provided by the embodiments of the present application;
Fig. 8 is a kind of signal pictorial diagram of intelligent wearable device provided by the embodiments of the present application;
Fig. 9 is a kind of structural schematic diagram of terminal device provided by the embodiments of the present application.
Specific embodiment
Further illustrate the technical solution of the application below with reference to the accompanying drawings and specific embodiments.It is understood that
It is that specific embodiment described herein is used only for explaining the application, rather than the restriction to the application.It further needs exist for illustrating
, part relevant to the application is illustrated only for ease of description, in attached drawing rather than entire infrastructure.
It should be mentioned that some exemplary embodiments are described as before exemplary embodiment is discussed in greater detail
The processing or method described as flow chart.Although each step is described as the processing of sequence by flow chart, many of these
Step can be implemented concurrently, concomitantly or simultaneously.In addition, the sequence of each step can be rearranged.When its operation
The processing can be terminated when completion, it is also possible to have the additional step being not included in attached drawing.The processing can be with
Corresponding to method, function, regulation, subroutine, subprogram etc..
With the continuous development of intelligent wearable device, there is cell-phone function, and be able to satisfy portability and large screen vision body
The intelligent wearable device for testing demand gradually obtains the approval of user.Currently, intelligent wearable device noisy environment (e.g., noise compared with
Big workshop, concert scene etc.) in, the audio data noise of acquisition is larger, so that user's input can not be obtained accurately
Audio data, and then influence user to the normal use and control of intelligent wearable device.
The embodiment of the present application provides a kind of acquisition methods of audio data, can obtain signal receiving audio data
Afterwards, the first audio data is obtained by osteoacusis mode, obtains target audio data after carrying out denoising.And then avoid user
When speaking in noisy environment, can not accurately obtain user input audio data the problem of, improve intelligent wearable device obtain
The accuracy of audio data.Concrete scheme is as follows:
Fig. 1 is the flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, and this method is for intelligence
Wearable device obtains the case where audio data of user's input, and this method can be set by the intelligence wearing for being equipped with bone conduction device
Standby or terminal device executes, for example, the intelligent wearable device can be intelligent glasses, intelligent helmet, Intelligent bracelet etc., terminal
Equipment can be smart phone, tablet computer, laptop etc..This method specifically comprises the following steps:
Step 110 receives audio data acquisition signal.
Wherein, audio data, which obtains signal, can be intelligent wearable device or other terminal device (such as mobile phones, tablet computer
Deng) receive the trigger action of user and generate for indicating that user wants the signal of input audio data.Optionally, if sound
Frequency data acquisition signal is the signal that user triggers the generation of other terminal devices, then other terminal devices can be by the audio number of generation
Intelligent wearable device is sent to by terminal device according to signal is obtained.For example, user clicks the record button on mobile phone, then can touch
Hair generates audio data and obtains signal, and the audio data of generation can be obtained signal and be sent to intelligent wearable device by mobile phone at this time.
Optionally, in the embodiment of the present application, the mode that triggering generates that audio data obtains signal has very much, the application couple
This without limiting, can be user click it is corresponding with audio signal is obtained on intelligent wearable device or other terminal devices
Fixed key for example, needing to obtain the audio data of user's input when recording, therefore can be worn when detection user clicks intelligence
When wearing record button in equipment, i.e., triggering generates audio data and obtains signal;Can also be detection intelligent wearable device or other
Whether preset affiliate application starts in terminal device, and optionally, affiliate application may include voice/video chat
Program, recorded program, remote guide program.For example, will start video call function when opening remote guide program, at this time must
It will be related to the acquisition of audio user information, therefore, affiliate application can be set by remote guide program, when detecting
When remote guide program starts, i.e., triggering generates audio data and obtains signal;The bone for being also possible to detect intelligent wearable device passes
It leads and whether produces bone vibration at equipment, for example, when user speaks, process of the audio signal based on osteoacusis can cause bone to shake
Dynamic, there are when bone vibration at detection bone conduction device place, i.e., triggering generates audio data and obtains signal.
Optionally, in the embodiment of the present application, reception audio data acquisition signal, which can be, is detecting that intelligently wearing is set
The standby receive capabilities for starting or opening when being kept in motion audio data and obtain signal, for example, the intelligence wearing that user wears
When equipment starts or is kept in motion, usually indicate that user will use the intelligent wearable device, it is therefore more likely that can trigger tone
Frequency data acquisition signal such as issues phonetic control command.Therefore shape can be moved when detecting intelligent wearable device starting or being in
When state, the receive capabilities that audio data obtains signal are opened, audio data is received and obtains signal.Optionally, audio number is being received
When according to obtaining signal, it can be and be received once every preset time (such as 5 minutes), be also possible to real-time perfoming reception, Zhi Daozhi
It can wearable device closing.
Further, it when receiving audio data acquisition signal, can be in current location information and preset position information
Timing receives audio data and obtains signal.
Wherein, preset position information can be that user is preset according to self-demand, can also be system default,
For example, it may be in view of being not easy to obtain the audio data of user's input in noisy environment, therefore can be by noisy environment, such as
Market, stadium, street etc. are set as preset position information.Current location information refers to user position information, can
To obtain current location information by the navigation positioning unit in intelligent wearable device, can also be interacted with intelligent terminal,
Current location information is obtained by the navigation positioning unit on intelligent terminal.
Optionally, if current location information is matched with preset position information, illustrate that current location is in noisy environment, then
Start audio data and obtain semiotic function, the audio data of user's input is obtained by osteoacusis mode.If current location information
It is mismatched with preset position information, then illustrates that current location is not belonging to noisy environment, for example, user is at home, at this time by normal
The air transmitted mode of rule can accurately obtain the audio data of user's input.
Step 120 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Wherein, osteoacusis mode is the slight vibration of caused incidence bone when being talked using people voice signal is received
Collection gets up to switch to the mode of electric signal.Sound is acquired by air transmitted since it is different from conventional microphone, so in very noise
The outflow of sound high-resolution can also be come in miscellaneous environment.Optionally, the embodiment of the present application can by bone conduction device come
The first audio data is obtained, if be intelligent glasses when intelligence wearing, bone conduction device be can be set in intelligent glasses temple portion
Position, when wearing, the bone conduction device that intelligent glasses temple position is arranged in contacts user with the skull near user's ear, can
Voice signal is collected to shake well by skull.First audio data is the audio data of user's input.
In the embodiment of the present application, if step 110 receives audio data and obtains signal, start bone conduction device
Speech voice input function collects the audio data of user's input, first obtained by osteoacusis mode by osteoacusis mode
Audio data, clarity is higher, will not enter bone conduction device for the noisy noise in environment by air transmitted.For example,
User is sought help using the remote guide function of intelligent wearable device to Remote in the biggish shop maintenance machine of noise, this
When, it can be very good to filter out machine sound in noisy environment in the audio data of the user's input obtained by bone conduction device
Sound ensure that the user of acquisition inputs the clarity of sound.
Optionally, if received audio data, which obtains signal, to be generated by detection associated program starting, due to association
Application program is the software that may relate to obtain audio data in preset operational process, but does not represent starting association and answer
With audio data is centainly obtained after program, therefore, a time interval can be preset, if in the prefixed time interval, it is bone-free
It shakes (i.e. user does not have input audio data), then closes bone conduction device, wait until that receiving audio data next time obtains signal
When, restart the audio data that bone conduction device obtains user's input.For example, detect that user opens associated program " wechat ",
It then generates audio data and obtains signal, starting bone conduction device obtains user's input after receiving audio data and obtaining signal
Audio data, if user only sees wechat circle of friends, and there is no voice-enabled chat or views in prefixed time interval (such as 1 minute)
Frequency is conversed, and closes bone conduction device at this time.If receive after a while due to detecting that user triggers the video in wechat again
Talk button and the audio data that generates obtains signal, then be again started up the audio data that bone conduction device obtains user's input.
It should be noted that needing to detect intelligence wearing before obtaining the first audio data by osteoacusis mode and setting
Whether standby bone conduction device is normal, if contacts with user's skull, if bone conduction device abnormal state or not and with account
Bone contact, then bone conduction device is the audio data that user's input can not be obtained by osteoacusis mode, is lost in order to prevent at this time
The audio data for leaking user's input needs to start microphone (microphone based on control conduction pattern work) or spare osteoacusis
Equipment temporarily to obtain the audio data of user's input and prompt user has dressed intelligent wearable device etc..
Step 130 carries out denoising to the first audio data, obtains target audio data.
It is smaller by Environmental Noise Influence although obtaining the first audio data by osteoacusis mode, obtaining by the
Still there can be certain noise in one audio data, the sound of first-class generation for example, user coughs, grits one's teeth, scratching, these sound
May be unobvious by air transmitted mode, but obtained then very clearly by osteoacusis mode, therefore, passed for passing through bone
The audio data that the mode of leading obtains is still to carry out denoising, and the invalid data removed in audio data (such as coughs, grits one's teeth, scratching
First-class sound), obtain not having noisy clarity high uses the first audio data as target audio data.
In the embodiment of the present application, the mode for carrying out denoising to the first audio data has very much, not to this application
It is defined, can be and pre- first pass through a large amount of audio data sample (comprising the audio data with noise and muting audio
Data) training filter, denoising is carried out by trained filter, for example, the filter of training can be minimum
Square error (LMS) filter.When carrying out denoising to the first audio data, the first audio number that analysis obtains can also be
According to audio frequency characteristics (such as tone color, tone, frequency, loudness), the audio of the audio data inputted due to noise and user is special
Sign is different, therefore can be corresponding by the first audio frequency characteristics of variation abnormality according to the situation of change of audio data sound intermediate frequency feature
Audio data (i.e. noise section) removal.For example, the 2nd second to the 5th second audio frequency characteristics in the first audio data of acquisition
Variation abnormality then illustrates that the period corresponding audio data is noise, needs to be deleted.
Optionally, it since obtained target audio data only include the audio data of user's input, and makes an uproar not comprising environment
Sound, therefore, if obtained target audio data are the dialog contexts of user and other users, the target audio that can will be obtained
Data are sent to call another party after carrying out data encryption.If obtained target audio data are the recording number of user's input
According to can then obtained target audio data be stored in the corresponding storage unit of recorded program.If obtained target audio
Data are control instruction of the user to intelligent wearable device or intelligent terminal, then intelligent wearable device can be directly in response to obtaining
Target audio data are sent to intelligent terminal, so that intelligent terminal responds the target audio data.For example, working as target audio
Data are " weather forecast ", then intelligent wearable device can respond starting weather prompting function at once, export current weather to user.
The acquisition methods of the audio data provided in the embodiment of the present application, first reception audio data obtain signal;Then,
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;Finally, to the first audio number
According to denoising is carried out, target audio data are obtained.Audio data is obtained relative to by air transmitted mode, the application is implemented
Example can obtain the first audio data by osteoacusis mode, after carrying out denoising after receiving audio data and obtaining signal
Obtain target audio data.And then when avoiding user and speaking in noisy environment, the audio of user's input can not be accurately obtained
The problem of data, improves the accuracy that intelligent wearable device obtains audio data.
Fig. 2 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, is suitable for obtaining
The case where when audio data, not only obtaining user, what is said or talked about, also to obtain ambient sound, for example, user thinks one section of record certainly
When the section of football match video that oneself explains, the live sound of football match should be obtained, also to obtain the sound of oneself explanation.As
Further explanation to above-described embodiment, comprising:
Step 210 receives audio data acquisition signal.
The embodiment of the present application will not only execute step 210 and pass through osteoacusis mode after receiving audio data and obtaining signal
The first audio data is obtained, step 230 is also executed and obtain environmental audio data.Illustratively, user is watching a football
When match, the live video that one section of record explains oneself is thought, at this point, step 220 should be executed, obtain the sound of oneself explanation, also
Step 230 is executed, the live sound of football match is obtained, to restore true match scene.
Step 220 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Illustratively, it can be very good to filter out noisy ambient sound by bone conduction device, get user's input
Audio data.
Step 230 obtains signal acquisition environmental audio data according to audio data.
Illustratively, environmental audio data can be obtained by the microphone to be worked based on air transmitted mode, i.e., existing
Intelligent wearable device obtain audio data mode, can be arranged in intelligent wearable device at least one be based on air transmitted
Microphone that mode works obtains ambient sound.
Step 240 generates second audio data according to environmental audio data and the first audio data.
Wherein, second audio data is counted after environmental audio data and the first audio data are carried out fusion treatment
According to optionally, carrying out the mode that fusion treatment generates second audio data to environmental audio data and the first audio data has very
More, the embodiment of the present application without limiting, can be the volume of the volume of environmental audio data and the first audio data this
After being adjusted, it is fused together generation second audio data, for example, the football live audio data that will acquire is turned down, will be used
After the explanation audio data of family input tunes up, two-way audio data fusion is obtained into second audio data together.
Optionally, it is contemplated that also include the first audio data in the environmental audio data of acquisition, only it is covered in noise
It in miscellaneous ambient sound, can not distinguish well, therefore generate the according to environmental audio data and the first audio data
When two audio datas, it can also be the first audio data of part replaced the first audio data in environmental audio data, obtain
To second audio data.For example, explaining audio data, the use in football live audio data that will acquire according to the user of acquisition
Family explains audio data and filters out, and the user of the football live audio data after then filtering out and acquisition explains audio data fusion
Together, second audio data is obtained.
Step 250 carries out denoising to second audio data, obtains target audio data.
Optionally, when carrying out denoising to second audio data, noise in the first audio data is not only filtered out (such as
Cough, grit one's teeth, scratching first-class sound), when also generating second audio data caused by new noise.For example, if the second audio number
According to being to merge to generate after the first audio data and environmental audio data to be carried out to volume adjustment, then can go out in second audio data
The first audio data being now overlapped (is explained in audio data and environmental audio data including the user that bone conduction device obtains and is wrapped
The user contained explains audio data), denoising at this time can be the first audio data portion in environmental audio data
It filters out.
Optionally, if second audio data is the first audio data replaced the first audio data in environmental audio data
It generates, is then likely to occur ambient sound in second audio data and is greater than the first audio data, it is excessive so as to cause ambient noise,
The case where can't hear the first audio data, denoising at this time, which can be, reduces the volumes of environmental audio data, improves the
The volume of one audio data.
The acquisition methods of audio data provided by the embodiments of the present application can lead to after receiving audio data and obtaining signal
It crosses osteoacusis mode and air transmitted mode obtains audio data jointly and generates second audio data, second audio data is carried out
Target audio data are obtained after denoising.It can be improved the accuracy that intelligent wearable device obtains audio data.
Fig. 3 A is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, and Fig. 3 B is this Shen
Please the user that provides of embodiment the schematic diagram of a scenario of remote guide is carried out by intelligent wearable device and Remote.The application is real
It applies example and is not only used to obtain audio data, but also the feelings for exporting audio data suitable for the bone conduction device of intelligent wearable device
Condition.For example, bone conduction device both send out by available user's word when user uses the remote guide function of intelligent wearable device
Long-range expert is given, and expert's word can be exported to user.Further explanation in conjunction with Fig. 3 B to above-described embodiment,
Include:
Step 310 starts.
Step 320 judges whether that receiving audio data obtains signal, if so, step 330 is executed, if it is not, executing step
350。
Since bone conduction device not only has the function of obtaining audio data, also have the function of transmitting audio data, and
It can replace earphone and loudspeaker, audio data to be played exported to user.Audio data is transmitted by bone conduction device,
Ears can be liberated, also it will be noted that the variation of ambient enviroment while receiving audio data, use more safety for example, and
Other side can be also allowed to be clear from the instruction of expression when speaking volume is smaller.
Optionally, it in order to enable can also accurately obtain in a noisy environment and playing audio-fequency data, can be
When receiving audio data acquisition signal, step 330 is executed, the first audio data is obtained by osteoacusis mode;Do not connecing
When receiving audio data acquisition signal, step 350 is executed, audio data to be played is exported by osteoacusis mode.Optionally,
Audio data, which obtains signal and can be, at this time shakes what triggering generated by bone.For example, user is when using remote guide function, it can
By being shaken at detection bone conduction device with the presence or absence of bone, and if it exists, illustrate that user is speaking, generate audio data and obtain
Signal, the audio data for receiving generation obtain execution step 330 after signal, the sound of user's input are obtained by osteoacusis mode
Frequency evidence, bone shakes if it does not exist, then illustrates that user in the guidance for listening expert, executes step 350 at this time, passes through osteoacusis mode
Audio data is instructed to user output expert.
Step 330 obtains signal according to the audio data, obtains the first audio data by osteoacusis mode.
Illustratively, in user and Remote communication process, when a user speaks, sound, which is based on osteoacusis mode, to be drawn
Bone vibration is played, triggering generates audio data and obtains signal when generating bone vibration, and the audio data for obtaining generation obtains letter
After number, then it can control bone conduction device and be based on osteoacusis mode and obtain user what is said or talked about.
Step 340 carries out denoising to the first audio data, obtains target audio data.
Step 350 obtains audio data to be played, exports audio data to be played by osteoacusis mode.
Wherein, audio data to be played, which can be, needs the audio data that plays to user, when being not received by audio
When data acquisition signal, the audio data remotely sended over that intelligent wearable device meeting automated communications unit obtains, and pass through
The audio data that bone conduction device will acquire is based on osteoacusis mode and exports to user.For example, audio data to be played is to use
Family use remote guide function when, what is said or talked about for Remote, user in Remote communication process, if user does not say
When words, bone vibration will not be generated, i.e. acquisition audio data obtains signal, at this point, expert's guidance that bone conduction device can will acquire
Audio data is based on osteoacusis mode and exports to user.
Optionally, the audio data to be played that osteoacusis mode exports in the embodiment of the present application is not limited to user and leads to
Dialog context when words can also be that the playing function that bone conduction device is replaced to earphone, data to be played at this time can be
Music to be played.
If step 360 when exporting audio data to be played by osteoacusis mode, receives audio data and obtains letter
Number, then start spare osteoacusis mode and obtains third audio data.
Wherein, third audio data is also the audio data of user's input, which, which can be, is passing through osteoacusis
The audio data that user inputs when mode exports audio data to be played.
Optionally, the bone conduction device in intelligent wearable device can be set multiple, and one is main bone conduction device, remaining
For spare bone conduction device, main bone conduction device is based on main osteoacusis mode and works, and backup bone conduction device is based on backup
Osteoacusis mode works.Wherein, when needing to carry out the acquisition and output of audio data, main osteoacusis is selected to set first
It is standby, when main bone conduction device is working or when abnormal state, enables backup bone conduction device, can guarantee to set even if osteoacusis
When for damage or being used, the audio data of user's input also can be accurately obtained.
Optionally, even if current time main bone conduction device, which is in, plays operating mode, audio data acquisition is also received
Signal, if being received when main bone conduction device exports audio data to be played to user based on main osteoacusis mode
Audio data obtains signal, then main bone conduction device continues to execute and exports audio data to be played to user at this time, starts standby
Equipment is passed to bone, the third audio data that user inputs is obtained based on backup osteoacusis mode, and execute step 340, to acquisition
Audio data carry out denoising, obtain target audio data.For example, user, which listens main bone to pass to equipment, sends back the expert come
During guidance, query is produced, he may interrupt him during expert speaks, at this point, main bone conduction device also exists
Play expert guidance, and receive because user speak generation voice data obtain signal, then will start spare osteoacusis and set
It is standby that user's audio data currently entered is obtained based on backup osteoacusis mode, and after carrying out denoising, by remotely leading to
Letter is sent to equipment where expert.
The acquisition methods of audio data provided by the embodiments of the present application can lead to when receiving audio data acquisition signal
It crosses osteoacusis mode and obtains the first audio data, when not triggering audio acquisition signal, exported by osteoacusis mode to be played
Audio data.The acquisition of audio data is completed by osteoacusis mode and plays two functions, and improves intelligent wearable device
Obtain the accuracy with playing audio-fequency data.
Fig. 4 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, as to above-mentioned
The further explanation of embodiment, comprising:
Step 410 receives audio data acquisition signal.
Step 420 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Step 430 extracts the corresponding audio frequency characteristics of the first audio data.
Wherein, acoustic characteristic may include one or more in the corresponding tone color of audio data, tone, frequency, loudness etc.
A, the audio frequency characteristics of the difference of sound source, corresponding audio data are also different, for example, the sound that user speaks in the first audio data
Frequency feature audio frequency characteristics corresponding with sound difficult to tackle are entirely different.Sound source is identical, and the audio frequency characteristics of corresponding audio data are also different
Fixed identical, for example, user speaks and coughs, the corresponding audio frequency characteristics of sound issued are also different.
It optionally, may incessantly include the corresponding audio data of a sound source in the first audio data that step 420 obtains,
The corresponding each audio frequency characteristics of the whole section audio of the first audio data are extracted respectively, for example, both saying including user in whole section audio data
The audio of words further includes the audio that user scratches one's head, then extracts the audio frequency characteristics that user speaks respectively and the audio that user scratches one's head is special
Sign.Optionally, the corresponding audio frequency characteristics of the first audio data are extracted, can be and carried out by trained audio feature extraction model
It extracts, can also be and extracted by preset audio feature extraction algorithm, the application is to this without limiting.
Step 440, according to audio frequency characteristics, whether there is invalid data in audio data.
Wherein, invalid data refers to the data of non-user voice input in the first audio data extracted, for example, user
The sound etc. that the sound difficult to tackle generated or user's cough generate.
Optionally, according to audio frequency characteristics, it whether there is invalid data in audio data, can be and acquire and deposit in advance
The audio frequency characteristics for storing up a variety of invalid datas, when judging, the audio frequency characteristics for the first audio data that will acquire with it is pre-stored
The audio frequency characteristics of a variety of invalid datas are compared, and determine the invalid data in the first audio data obtained, for example, if obtaining
The first audio data in have the tone of section audio data and tone color tone corresponding with sound difficult to tackle and tone color identical, then the section
Audio data is invalid data.The every kind of audio frequency characteristics (such as frequency) for being also possible to the first audio data that will acquire are depicted as
Full curve analyzes the tendency of curve, and the corresponding audio data in part that tendency mutates is considered as invalid data.
Step 450, if there is invalid data, then filter out the invalid data in the first audio data, obtain target audio
Data.
Optionally, if there are invalid datas in the first audio data obtained, in the first audio data of acquisition
Except ineffective data part, after remaining audio data is spliced, it can be completed and the first audio data of acquisition is gone
It makes an uproar the process of processing, target audio data can be obtained.
The acquisition methods of audio data provided by the embodiments of the present application can lead to after receiving audio data and obtaining signal
It crosses osteoacusis mode and obtains the first audio data, according to the audio frequency characteristics of the first audio data, filter out in the first audio data
Invalid data obtains target audio data.And then when user being avoided to speak in noisy environment, user's input can not be accurately obtained
Audio data the problem of, improve intelligent wearable device obtain audio data accuracy.
Fig. 5 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, and the application is implemented
Example is suitable for the case where user speech controls intelligent wearable device, as the further explanation to above-described embodiment, packet
It includes:
Step 510 receives audio data acquisition signal.
Step 520 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Step 530 carries out denoising to the first audio data, obtains target audio data.
Step 540 carries out intention analysis to target audio data.
Optionally, target audio data are carried out with the process for being intended to analysis, semantic point as is carried out to the first audio data
The process of analysis.The analysis that can be intended to by the intention analysis model pre-established in the embodiment of the present application.Wherein, it is intended that
Analysis model, which can be, is arranged keyword as much as possible for all control instructions in advance, and establishes control instruction and keyword
Between mapping relations, input neural network model learnt.In the intention point for actually carrying out target audio data
When analysis, it can be and first target audio data are segmented, the keyword after participle is then inputted into trained intention and is analyzed
In model, model analyzes the keyword of input based on learning algorithm when training, the corresponding intention of output keyword.Example
Such as, target audio data are " today, how is weather? ", to the word segmented to obtain " today ", " weather ", " how
Sample ", and these three words are inputted and are intended to analysis model, it obtains being intended to analysis result being that user wants to understand the weather of today.
Step 550, response are intended to the corresponding control instruction of analysis result.
Optionally, the corresponding control instruction of target audio data is determined according to semantic analysis result, is then responding to the control
Instruction, such as, it is intended that analysis result is that user wants to understand the weather of today, and the corresponding control instruction of the intention is " starting weather
Prompt ", then system responds the instruction, starts weather prompting function, exports current weather to user.
Optionally, the corresponding control instruction of target audio data can be the control to intelligent wearable device of user's sending
System instruction can also be the control instruction to terminal devices such as mobile phones that user issues.To this without limiting, if control instruction
For the corresponding control instruction of terminal device, intelligent wearable device can be logical between terminal device by it by the control instruction
Letter connection, is sent to terminal device, so that terminal device executes the control instruction.
The acquisition methods of audio data provided by the embodiments of the present application can lead to after receiving audio data and obtaining signal
It crosses osteoacusis mode and obtains the first audio data, denoised and be intended to analyze, thus control of the response to intelligent wearable device.
The accuracy that intelligent wearable device obtains audio data is improved, accurately intelligence can be worn by voice in a noisy environment
Equipment is worn to be controlled.
Fig. 6 is a kind of structural schematic diagram of the acquisition device of audio data provided by the embodiments of the present application.As shown in fig. 6,
The device includes: signal detection module 610, and audio obtains module 620, denoising module 630.
Signal receiving module 610 obtains signal for receiving audio data;
First audio obtains module 620, and the audio data for being received according to the signal receiving module 610 obtains
The number of winning the confidence obtains the first audio data by osteoacusis mode;;
Denoising module 630, for first audio obtain module 620 obtain first audio data into
Row denoising obtains target audio data.
Further, above-mentioned apparatus further include:
Environmental audio obtains module, for obtaining signal acquisition environmental audio data according to the audio data;
Correspondingly, denoising module 630 is used for, generated according to the environmental audio data and first audio data
Second audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, above-mentioned apparatus further include:
Audio output module exports the audio to be played by osteoacusis mode for obtaining audio data to be played
Data.
Further, the first audio obtains module 620 and is also used to, if exported by the osteoacusis mode described wait broadcast
When putting audio data, receives audio data and obtain signal, then start spare osteoacusis mode and obtain third audio data.
Further, denoising module 630 is used for, and extracts the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, above-mentioned apparatus further include:
It is intended to analysis module, for carrying out intention analysis to the target audio data;
Respond module is instructed, for responding the corresponding control instruction of the intention analysis result.
Further, signal receiving module 610 is also used to, and when current location information is matched with preset position information, is connect
It receives audio data and obtains signal.
The acquisition device of the audio data provided in the embodiment of the present application, firstly, signal receiving module 610 receives audio number
According to acquisition signal;Then, the first audio obtains module 620 and obtains signal according to the audio data, is obtained by osteoacusis mode
Take the first audio data;Finally, denoising module 630 carries out denoising to first audio data, target sound is obtained
Frequency evidence.The first audio data is obtained relative to by air transmitted mode, the embodiment of the present application can receive audio data
After obtaining signal, the first audio data is obtained by osteoacusis mode, obtains target audio data after carrying out denoising.In turn
When user being avoided to speak in noisy environment, can not accurately obtain user input audio data the problem of, improve intelligence wearing
The accuracy of equipment acquisition audio data.
Method provided by the aforementioned all embodiments of the application can be performed in above-mentioned apparatus, and it is corresponding to have the execution above method
Functional module and beneficial effect.The not technical detail of detailed description in the present embodiment, reference can be made to the aforementioned all implementations of the application
Method provided by example.
The present embodiment provides a kind of intelligent wearable device on the basis of the various embodiments described above, and Fig. 7 is that the application is implemented
A kind of structural schematic diagram for intelligent wearable device that example provides, Fig. 8 is a kind of intelligent wearable device provided by the embodiments of the present application
Signal pictorial diagram.As shown in Figure 7 and Figure 8, which includes: memory 701, processor (Central
Processing Unit, CPU) 702, display unit 703, touch panel 704, heart rate detection mould group 705, range sensor
706, camera 707, bone-conduction speaker 708, microphone 709, breath light 710, these components pass through one or more communications
Bus or signal wire 711 communicate.
It should be understood that diagram intelligent wearable device is only an example of intelligent wearable device, and intelligently wear
Wearing equipment can have than shown in the drawings more or less component, can combine two or more components, or
Person can have different component configurations.Various parts shown in the drawings can include one or more signal processings and/
Or it is realized in the combination of hardware including specific integrated circuit, software or hardware and software.
Just the intelligent wearable device provided in this embodiment obtained for audio data is described in detail below, the intelligence
Energy wearable device is by taking intelligent glasses as an example.
Memory 701, the memory 701 can be accessed by CPU702, and the memory 701 may include that high speed is random
Access memory, can also include nonvolatile memory, for example, one or more disk memory, flush memory device or its
His volatile solid-state part.
Display unit 703, can be used for the operation and control interface of display image data and operating system, and display unit 703 is embedded in
In the frame of intelligent wearable device, frame is internally provided with inner transmission lines 711, the inner transmission lines 711 and display
Component 703 connects.
Touch panel 704, which is arranged in the outside of at least one intelligent wearable device temple, for obtaining
Touch data is taken, touch panel 704 is connected by inner transmission lines 711 with CPU702.Wherein, touch panel 704 is detectable
The finger sliding of user, clicking operation, and the data detected are transmitted to processor 702 accordingly and are handled with generation pair
The control instruction answered, illustratively, can be left shift instruction, right shift instruction, move up instruction, move down instruction etc..Illustratively, it shows
Show component 703 can video-stream processor 702 transmit virtual image data, which can be accordingly according to touch panel
704 user's operations that detect carry out corresponding changes, specifically, can be carry out screen switching, when detect left shift instruction or
Switch upper one or next virtual image picture after right shift instruction accordingly;When display unit 703 shows video playing information
When, which, which can be, plays out playbacking for content, and right shift instruction can be the F.F. for playing out content;Work as display
The display of component 703 is when being editable word content, and the left shift instruction, right shift instruction move up instruction, move down instruction and can be pair
The displacement operation of cursor, the i.e. position of cursor can move the touch operation of touch tablet according to user;Work as display unit
When the contents of 703 displays are game animation picture, the left shift instruction, right shift instruction move up instruction, move down instruction and can be to trip
Object in play is controlled, in machine game like flying, can by the left shift instruction, right shift instruction, move up instruction, move down instruction point
Not Kong Zhi aircraft heading;When display unit 703 can show the video pictures of different channel, which is moved to right
It instructs, moves up instruction, moves down the switching for instructing and can carrying out different channel, wherein moving up instruction and moving down instruction can be switching
To pre-set channel (the common channel that such as user uses);When display unit 703 shows static images, which is moved to right
It instructs, moves up instruction, moves down the switching that instructs and can carry out between different pictures, wherein left shift instruction can be to switch to one
Width picture, right shift instruction, which can be, switches to next width figure, and an atlas can be to switch to by moving up instruction, and moving down instruction can be with
It is to switch to next atlas.The touch panel 704 can also be used to control the display switch of display unit 703, exemplary
, when long pressing 704 touch area of touch panel, display unit 703, which is powered, shows graphic interface, when long pressing touch again
When 704 touch area of panel, display unit 703 is powered off, can be by carrying out in touch panel 704 after display unit 703 is powered
Upper cunning and operation of gliding are to adjust the brightness or resolution ratio that show image in display unit 703.
Heart rate detection mould group 705, for measuring the heart rate data of user, heart rate refers to beats per minute, the heart rate
Mould group 705 is detected to be arranged on the inside of temple.Specifically, the heart rate detection mould group 705 can be in such a way that electric pulse measures
Human body electrocardio data are obtained using stemness electrode, heart rate size is determined according to the amplitude peak in electrocardiogram (ECG) data;The heart rate detection
Mould group 705 can also be by being formed using the light transmitting and light receiver of photoelectric method measurement heart rate, correspondingly, the heart rate is examined
Mould group 705 is surveyed to be arranged at temple bottom, the ear-lobe of human body auricle.Heart rate detection mould group 705 can phase after collecting heart rate data
The progress data processing in processor 702 that is sent to answered has obtained the current heart rate value of wearer, in one embodiment, processing
Device 702, can be by the heart rate value real-time display in display unit 703 after determining the heart rate value of user, optional processor
702 are determining that heart rate value lower (such as less than 50) or higher (such as larger than 100) can trigger alarm accordingly, while by the heart
Rate value and/or the warning message of generation are sent to server by communication module.
Range sensor 706, may be provided on frame, the distance which is used to incude face to frame,
The realization of infrared induction principle can be used in the range sensor 706.Specifically, the range sensor 706 is by the range data of acquisition
It is sent to processor 702, data control the bright dark of display unit 703 to processor 702 according to this distance.Illustratively, work as determination
When the collected distance of range sensor 706 is less than 5 centimetres out, the corresponding control display unit 703 of processor 702, which is in, to be lighted
State, when determine range sensor be detected with object close to when, it is corresponding control display unit 704 and be in close shape
State.
Breath light 710 may be provided at the edge of frame, when display unit 703 closes display picture, the breath light 710
It can be lighted according to the control of processor 702 in the bright dark effect of gradual change.
Camera 707 can be the position that the upper side frame of frame is arranged in, and acquire the proactive of the image data in front of user
As module, the rear photographing module of user eyeball information can also be acquired, is also possible to the combination of the two.Specifically, camera 707
When acquiring forward image, the image of acquisition is sent to the identification of processor 702, processing, and trigger accordingly according to recognition result
Trigger event.Illustratively, when user wears the wearable device at home, by being identified to the forward image of acquisition,
If recognizing article of furniture, corresponding inquiry whether there is corresponding control event, if it is present accordingly by the control
The corresponding control interface of event processed is shown in display unit 703, and user can carry out corresponding furniture object by touch panel 704
The control of product, wherein the article of furniture and intelligent wearable device are connected to the network by bluetooth or wireless self-networking;Work as user
When wearing the wearable device outdoors, target identification mode can be opened accordingly, which can be used to identify specific
People, camera 707 by the image of acquisition be sent to processor 702 carry out recognition of face processing, if recognizing the pre- of setting
If face, then the loudspeaker that can be integrated accordingly by intelligent wearable device carries out sound casting, which may be used also
With plants different for identification, for example, processor 702 is adopted according to the touch operation of touch panel 704 with recording camera 707
The present image of collection is simultaneously sent to server by communication module to be identified, server carries out the plant in acquisition image
It identifies and feeds back relevant botanical name, introduce to intelligent wearable device, and feedback data is shown in display unit 703.
Camera 707 can also be the image for acquiring user's eye such as eyeball, be generated by the identification of the rotation to eyeball different
Control instruction illustratively move up control instruction as eyeball is rotated up generation, eyeball, which rotates down generation and moves down control, to be referred to
It enables, the eyeball generation that turns left moves to left control instruction, and the eyeball generation that turns right moves to right control instruction, wherein display unit 703
Can video-stream processor 702 transmit virtual image data, what which can detect according to camera 707 accordingly
Control instruction that the mobile variation of user eyeball generates and change, specifically, can be carry out screen switching, moved to left when detecting
Control instruction switches upper one or next virtual image picture after moving to right control instruction accordingly;When display unit 703 is aobvious
When showing video playing information, this, which moves to left control instruction and can be, plays out playbacking for content, move to right control instruction can be into
The F.F. of row broadcasting content;When the display of display unit 703 is editable word content, this moves to left control instruction, moves to right control
System instruction moves up control instruction, moves down control instruction and can be displacement operation to cursor, i.e. the position of cursor can be according to user
The touch operation of touch tablet is moved;When the content that display unit 703 is shown is game animation picture, this moves to left control
System instruction moves to right control instruction, moves up control instruction, moving down control instruction and can be and control the object in game, such as
In aircraft game, control instruction can be moved to left by this, control instruction is moved to right, moves up control instruction, moving down control instruction and control respectively
The heading of aircraft processed;When display unit 703 can show the video pictures of different channel, this moves to left control instruction, moves to right
Control instruction moves up control instruction, moves down control instruction and can carry out the switching of different channel, wherein move up control instruction and under
Pre-set channel (the common channel that such as user uses) can be to switch to by moving control instruction;When display unit 703 shows static map
When piece, this moves to left control instruction, moves to right control instruction, moves up control instruction, moving down control instruction and can carry out between different pictures
Switching, wherein a width picture can be to switch to by moving to left control instruction, moved to right control instruction and be can be and switch to next width
Figure, an atlas can be to switch to by moving up control instruction, moved down control instruction and be can be and switch to next atlas.
The inner wall side of at least one temple is arranged in bone-conduction speaker 708, bone-conduction speaker 708, for that will receive
To the audio signal that sends of processor 702 be converted to vibration signal, or for the vibration signal received to be converted to audio
Signal is sent to processor 702.Wherein, sound can be transferred to human body inner ear by skull by bone-conduction speaker 708, pass through by
The electric signal of audio is changed into vibration signal and is transmitted in skull cochlea, then is perceived by auditory nerve.It can also will be by bone
The sound of transmitting is sent to processor 702, after vibration signal is changed into the electric signal of audio, everywhere by electric signal transmitting
Manage device 702.It is used as sounding device or sound receiver by bone-conduction speaker 708, reduces hardware configuration thickness, weight
It is lighter, while electromagnetic-radiation-free will not be influenced by electromagnetic radiation, and have antinoise, waterproof and liberation ears
A little.
Microphone 709, may be provided on the lower frame of frame, for acquiring external (user, environment) sound and being transmitted to
Processor 702 is handled.Illustratively, the sound that microphone 709 issues user be acquired and pass through processor 702 into
Row Application on Voiceprint Recognition can receive subsequent voice control, specifically, user if being identified as the vocal print of certification user accordingly
Collected voice is sent to processor 702 and identified according to recognition result generation pair by capable of emitting voice, microphone 709
The control instruction answered, such as " booting ", " shutdown ", " promoting display brightness ", " reducing display brightness ", the subsequent basis of processor 702
The control instruction of the generation executes corresponding control processing.
In the present embodiment, processor 702 is used for:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Further, after the reception audio data acquisition signal, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, before the reception audio data acquisition signal, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
Further, if export the audio data to be played by the osteoacusis mode, audio number is received
According to signal is obtained, then starts spare osteoacusis mode and obtain third audio data.
It is further, described that denoising is carried out to first audio data, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, it is described obtain target audio data after, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
Further, the reception audio data obtains signal, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
The acquisition device and intelligent wearable device of the audio data of the intelligent wearable device provided in above-described embodiment can be held
The acquisition methods of row audio data provided by any embodiment of the invention have and execute the corresponding functional module of this method and have
Beneficial effect.The not technical detail of detailed description in the above-described embodiments, reference can be made to audio provided by any embodiment of the invention
The acquisition methods of data.
Fig. 9 is the structural schematic diagram of another terminal device provided by the embodiments of the present application.As shown in figure 9, the terminal can
To include: shell (not shown), memory 901, central processing unit (Central Processing Unit, CPU) 902
(also known as processor, hereinafter referred to as CPU), the computer program that is stored on memory 901 and can be run on processor 902,
Circuit board (not shown) and power circuit (not shown).The circuit board is placed in the space that the shell surrounds
Portion;The CPU902 and the memory 901 are arranged on the circuit board;The power circuit, for for the terminal
Each circuit or device power supply;The memory 901, for storing executable program code;The CPU902 is by reading institute
The executable program code that stores in memory 901 is stated to run program corresponding with the executable program code.
The terminal further include: Peripheral Interface 903, RF (Radio Frequency, radio frequency) circuit 905, voicefrequency circuit
906, loudspeaker 911, power management chip 908, input/output (I/O) subsystem 909, touch screen 912, other input/controls
Equipment 910 and outside port 904, these components are communicated by one or more communication bus or signal wire 907.
It should be understood that graphic terminal 900 is only an example of terminal, and terminal device 900 can be with
With than shown in the drawings more or less component, two or more components can be combined, or can have
Different component configurations.Various parts shown in the drawings can include one or more signal processings and/or dedicated integrated
It is realized in the combination of hardware, software or hardware and software including circuit.
Just provided in this embodiment below to be described in detail for a kind of terminal device, the terminal device is with intelligent hand
For machine.
Memory 901, the memory 901 can be accessed by CPU902, Peripheral Interface 903 etc., and the memory 901 can
It can also include nonvolatile memory to include high-speed random access memory, such as one or more disk memory,
Flush memory device or other volatile solid-state parts.
The peripheral hardware that outputs and inputs of equipment can be connected to CPU902 and deposited by Peripheral Interface 903, the Peripheral Interface 903
Reservoir 901.
I/O subsystem 909, the I/O subsystem 909 can be by the input/output peripherals in equipment, such as touch screen 912
With other input/control devicess 910, it is connected to Peripheral Interface 903.I/O subsystem 909 may include 9091 He of display controller
For controlling one or more input controllers 9092 of other input/control devicess 910.Wherein, one or more input controls
Device 9092 processed receives electric signal from other input/control devicess 910 or sends electric signal to other input/control devicess 910,
Other input/control devicess 910 may include physical button (push button, rocker buttons etc.), dial, slide switch, behaviour
Vertical pole clicks idler wheel.It is worth noting that input controller 9092 can with it is following any one connect: keyboard, infrared port,
The indicating equipment of USB interface and such as mouse.
Wherein, according to the working principle of touch screen and transmission information medium classification, touch screen 912 can for resistance-type,
Capacitor induction type, infrared-type or surface acoustic wave type.Classify according to mounting means, touch screen 912 can be with are as follows: external hanging type, built-in
Formula or monoblock type.Classify according to technical principle, touch screen 912 can be with are as follows: vector pressure sensing technology touch screen, resistive technologies touching
Touch screen, capacitance technology touch screen, infrared technology touch screen or surface acoustic wave technique touch screen.
Touch screen 912, the touch screen 912 are the input interface and output interface between user terminal and user, can
It is shown to user depending on output, visual output may include figure, text, icon, video etc..Optionally, touch screen 912 is by user
The electric signal (electric signal of such as contact surface) triggered on touch screen curtain, is sent to processor 902.
Display controller 9091 in I/O subsystem 909 receives electric signal from touch screen 912 or sends out to touch screen 912
Electric signals.Touch screen 912 detects the contact on touch screen, and the contact that display controller 9091 will test is converted to and is shown
The interaction of user interface object on touch screen 912, i.e. realization human-computer interaction, the user interface being shown on touch screen 912
Object can be the icon of running game, the icon for being networked to corresponding network etc..It is worth noting that equipment can also include light
Mouse, light mouse are the extensions for the touch sensitive surface for not showing the touch sensitive surface visually exported, or formed by touch screen.
RF circuit 905 is mainly used for establishing the communication of intelligent sound box Yu wireless network (i.e. network side), realizes intelligent sound box
Data receiver and transmission with wireless network.Such as transmitting-receiving short message, Email etc..
Voicefrequency circuit 906 is mainly used for receiving audio data from Peripheral Interface 903, which is converted to telecommunications
Number, and the electric signal is sent to loudspeaker 911.
Loudspeaker 911 is reduced to for intelligent sound box to be passed through RF circuit 905 from the received voice signal of wireless network
Sound simultaneously plays the sound to user.
Power management chip 908, the hardware for being connected by CPU902, I/O subsystem and Peripheral Interface are powered
And power management.
In the present embodiment, central processing unit 902 is used for:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Further, after the reception audio data acquisition signal, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, before the reception audio data acquisition signal, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
Further, if export the audio data to be played by the osteoacusis mode, audio number is received
According to signal is obtained, then starts spare osteoacusis mode and obtain third audio data.
It is further, described that denoising is carried out to first audio data, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, it is described obtain target audio data after, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
Further, the reception audio data obtains signal, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
The embodiment of the present application also provides a kind of storage medium comprising terminal device executable instruction, and the terminal device can
It executes instruction when being executed by terminal device processor for executing a kind of acquisition methods of audio data, this method comprises:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Further, after the reception audio data acquisition signal, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, before the reception audio data acquisition signal, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
Further, if export the audio data to be played by the osteoacusis mode, audio number is received
According to signal is obtained, then starts spare osteoacusis mode and obtain third audio data.
It is further, described that denoising is carried out to first audio data, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, it is described obtain target audio data after, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
Further, the reception audio data obtains signal, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
The computer storage medium of the embodiment of the present application, can be using any of one or more computer-readable media
Combination.Computer-readable medium can be computer-readable signal media or computer readable storage medium.It is computer-readable
Storage medium for example may be-but not limited to-the system of electricity, magnetic, optical, electromagnetic, infrared ray or semiconductor, device or
Device, or any above combination.The more specific example (non exhaustive list) of computer readable storage medium includes: tool
There are electrical connection, the portable computer diskette, hard disk, random access memory (RAM), read-only memory of one or more conducting wires
(ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, portable compact disc read-only memory (CD-
ROM), light storage device, magnetic memory device or above-mentioned any appropriate combination.In this document, computer-readable storage
Medium can be any tangible medium for including or store program, which can be commanded execution system, device or device
Using or it is in connection.
Computer-readable signal media may include in a base band or as carrier wave a part propagate data-signal,
Wherein carry computer-readable program code.The data-signal of this propagation can take various forms, including but unlimited
In electromagnetic signal, optical signal or above-mentioned any appropriate combination.Computer-readable signal media can also be that computer can
Any computer-readable medium other than storage medium is read, which can send, propagates or transmit and be used for
By the use of instruction execution system, device or device or program in connection.
The program code for including on computer-readable medium can transmit with any suitable medium, including --- but it is unlimited
In wireless, electric wire, optical cable, RF etc. or above-mentioned any appropriate combination.
Can with one or more programming languages or combinations thereof come write for execute the application operation computer
Program code, programming language include object oriented program language-such as Java, Smalltalk, C++, are also wrapped
Include conventional procedural programming language-such as " C " language or similar programming language.Program code can be complete
Ground executes on the user computer, partly executes on the user computer, executing as an independent software package, partially existing
Part executes on the remote computer or executes on a remote computer or server completely on subscriber computer.It is being related to
In the situation of remote computer, remote computer can pass through the network of any kind --- including local area network (LAN) or wide area
Net (WAN)-be connected to subscriber computer, or, it may be connected to outer computer (such as utilize ISP
To be connected by internet).
Certainly, a kind of storage medium comprising computer executable instructions, computer provided by the embodiment of the present application
Operation is recommended in the application that executable instruction is not limited to the described above, and application provided by the application any embodiment can also be performed
Relevant operation in recommended method.
Note that above are only the preferred embodiment and institute's application technology principle of the application.It will be appreciated by those skilled in the art that
The application is not limited to specific embodiment described here, be able to carry out for a person skilled in the art it is various it is apparent variation,
The protection scope readjusted and substituted without departing from the application.Therefore, although being carried out by above embodiments to the application
It is described in further detail, but the application is not limited only to above embodiments, in the case where not departing from the application design, also
It may include more other equivalent embodiments, and scope of the present application is determined by the scope of the appended claims.
Claims (10)
1. a kind of acquisition methods of audio data characterized by comprising
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
2. the acquisition methods of audio data according to claim 1, which is characterized in that the reception audio data obtains letter
After number, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
3. the acquisition methods of audio data according to claim 1 or 2, which is characterized in that the reception audio data obtains
Before the number of winning the confidence, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
4. the acquisition methods of audio data according to claim 3, which is characterized in that
If export the audio data to be played by the osteoacusis mode, receives audio data and obtain signal, then
Start spare osteoacusis mode and obtains third audio data.
5. the acquisition methods of audio data according to claim 1 or 2, which is characterized in that described to first audio
Data carry out denoising, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
6. the acquisition methods of audio data according to claim 1 or 2, which is characterized in that described to obtain target sound frequency
According to later, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
7. the acquisition methods of audio data according to claim 1, which is characterized in that the reception audio data obtains letter
Number, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
8. a kind of acquisition device of audio data characterized by comprising
Signal receiving module obtains signal for receiving audio data;
First audio obtains module, and the audio data for being received according to the signal receiving module obtains signal, leads to
It crosses osteoacusis mode and obtains the first audio data;
Denoising module carries out at denoising for obtaining first audio data that module obtains to first audio
Reason, obtains target audio data.
9. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the program is held by processor
The acquisition methods of the audio data as described in any in claim 1-7 are realized when row.
10. a kind of terminal, including memory, processor and storage are on a memory and can be in the computer journey of processor operation
Sequence, which is characterized in that the processor realizes audio number as claimed in claim 1 when executing the computer program
According to acquisition methods.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811000969.3A CN109240639A (en) | 2018-08-30 | 2018-08-30 | Acquisition methods, device, storage medium and the terminal of audio data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811000969.3A CN109240639A (en) | 2018-08-30 | 2018-08-30 | Acquisition methods, device, storage medium and the terminal of audio data |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109240639A true CN109240639A (en) | 2019-01-18 |
Family
ID=65069811
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811000969.3A Pending CN109240639A (en) | 2018-08-30 | 2018-08-30 | Acquisition methods, device, storage medium and the terminal of audio data |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109240639A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111050248A (en) * | 2020-01-14 | 2020-04-21 | Oppo广东移动通信有限公司 | Wireless earphone and control method thereof |
CN111491176A (en) * | 2020-04-27 | 2020-08-04 | 百度在线网络技术(北京)有限公司 | Video processing method, device, equipment and storage medium |
CN111859008A (en) * | 2019-04-29 | 2020-10-30 | 深圳市冠旭电子股份有限公司 | Music recommending method and terminal |
CN112581970A (en) * | 2019-09-12 | 2021-03-30 | 深圳市韶音科技有限公司 | System and method for audio signal generation |
CN113053371A (en) * | 2019-12-27 | 2021-06-29 | 阿里巴巴集团控股有限公司 | Voice control system and method, voice suite, bone conduction and voice processing device |
WO2022141244A1 (en) * | 2020-12-30 | 2022-07-07 | 雷铭科技有限公司 | Audio file processing method and apparatus, and bone conduction device |
US11902759B2 (en) | 2019-09-12 | 2024-02-13 | Shenzhen Shokz Co., Ltd. | Systems and methods for audio signal generation |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102084668A (en) * | 2008-05-22 | 2011-06-01 | 伯恩同通信有限公司 | A method and a system for processing signals |
US20140337036A1 (en) * | 2013-05-09 | 2014-11-13 | Dsp Group Ltd. | Low power activation of a voice activated device |
US9324313B1 (en) * | 2013-10-23 | 2016-04-26 | Google Inc. | Methods and systems for implementing bone conduction-based noise cancellation for air-conducted sound |
US9578423B2 (en) * | 2012-12-11 | 2017-02-21 | Beijing Lenovo Software Ltd. | Electronic device and sound capturing method |
CN106686494A (en) * | 2016-12-27 | 2017-05-17 | 广东小天才科技有限公司 | Voice input control method of wearable device and wearable device |
CN106847275A (en) * | 2016-12-27 | 2017-06-13 | 广东小天才科技有限公司 | Method for controlling wearable device and wearable device |
-
2018
- 2018-08-30 CN CN201811000969.3A patent/CN109240639A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102084668A (en) * | 2008-05-22 | 2011-06-01 | 伯恩同通信有限公司 | A method and a system for processing signals |
US9578423B2 (en) * | 2012-12-11 | 2017-02-21 | Beijing Lenovo Software Ltd. | Electronic device and sound capturing method |
US20140337036A1 (en) * | 2013-05-09 | 2014-11-13 | Dsp Group Ltd. | Low power activation of a voice activated device |
US9324313B1 (en) * | 2013-10-23 | 2016-04-26 | Google Inc. | Methods and systems for implementing bone conduction-based noise cancellation for air-conducted sound |
CN106686494A (en) * | 2016-12-27 | 2017-05-17 | 广东小天才科技有限公司 | Voice input control method of wearable device and wearable device |
CN106847275A (en) * | 2016-12-27 | 2017-06-13 | 广东小天才科技有限公司 | Method for controlling wearable device and wearable device |
Non-Patent Citations (2)
Title |
---|
何国民主编: "《现代大学计算机基础》", 30 September 2015, 西安电子科技大学出版社 * |
迈克尔•拉毕格等编著: "《导演创作完全手册 第5版 插图修订版》", 31 July 2016 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111859008A (en) * | 2019-04-29 | 2020-10-30 | 深圳市冠旭电子股份有限公司 | Music recommending method and terminal |
CN111859008B (en) * | 2019-04-29 | 2023-11-10 | 深圳市冠旭电子股份有限公司 | Music recommending method and terminal |
CN112581970A (en) * | 2019-09-12 | 2021-03-30 | 深圳市韶音科技有限公司 | System and method for audio signal generation |
US11902759B2 (en) | 2019-09-12 | 2024-02-13 | Shenzhen Shokz Co., Ltd. | Systems and methods for audio signal generation |
CN113053371A (en) * | 2019-12-27 | 2021-06-29 | 阿里巴巴集团控股有限公司 | Voice control system and method, voice suite, bone conduction and voice processing device |
CN111050248A (en) * | 2020-01-14 | 2020-04-21 | Oppo广东移动通信有限公司 | Wireless earphone and control method thereof |
CN111050248B (en) * | 2020-01-14 | 2021-10-01 | Oppo广东移动通信有限公司 | Wireless earphone and control method thereof |
CN111491176A (en) * | 2020-04-27 | 2020-08-04 | 百度在线网络技术(北京)有限公司 | Video processing method, device, equipment and storage medium |
WO2022141244A1 (en) * | 2020-12-30 | 2022-07-07 | 雷铭科技有限公司 | Audio file processing method and apparatus, and bone conduction device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109240639A (en) | Acquisition methods, device, storage medium and the terminal of audio data | |
CN109120790B (en) | Call control method and device, storage medium and wearable device | |
CN110785735B (en) | Apparatus and method for voice command scenario | |
CN106164823B (en) | The non-vision feedback that vision changes in staring tracking and equipment | |
CN110874129A (en) | Display system | |
CN110326300B (en) | Information processing apparatus, information processing method, and computer-readable storage medium | |
CN109259724B (en) | Eye monitoring method and device, storage medium and wearable device | |
US20230045237A1 (en) | Wearable apparatus for active substitution | |
CN109036410A (en) | Audio recognition method, device, storage medium and terminal | |
CN109254659A (en) | Control method, device, storage medium and the wearable device of wearable device | |
CN109040462A (en) | Stroke reminding method, apparatus, storage medium and wearable device | |
CN109238306A (en) | Step counting data verification method, device, storage medium and terminal based on wearable device | |
CN108761795A (en) | A kind of Wearable | |
CN108874130B (en) | Play control method and related product | |
CN109255064A (en) | Information search method, device, intelligent glasses and storage medium | |
US20180054688A1 (en) | Personal Audio Lifestyle Analytics and Behavior Modification Feedback | |
CN108683790B (en) | Voice processing method and related product | |
CN109360549A (en) | A kind of data processing method, device and the device for data processing | |
CN109061903A (en) | Data display method, device, intelligent glasses and storage medium | |
CN109257490B (en) | Audio processing method and device, wearable device and storage medium | |
CN110097875A (en) | Interactive voice based on microphone signal wakes up electronic equipment, method and medium | |
CN110223711A (en) | Interactive voice based on microphone signal wakes up electronic equipment, method and medium | |
CN109241900A (en) | Control method, device, storage medium and the wearable device of wearable device | |
CN110111776A (en) | Interactive voice based on microphone signal wakes up electronic equipment, method and medium | |
CN106685459A (en) | Wearable device operation control method and wearable device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190118 |