[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN109240639A - Acquisition methods, device, storage medium and the terminal of audio data - Google Patents

Acquisition methods, device, storage medium and the terminal of audio data Download PDF

Info

Publication number
CN109240639A
CN109240639A CN201811000969.3A CN201811000969A CN109240639A CN 109240639 A CN109240639 A CN 109240639A CN 201811000969 A CN201811000969 A CN 201811000969A CN 109240639 A CN109240639 A CN 109240639A
Authority
CN
China
Prior art keywords
audio data
obtains
signal
audio
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811000969.3A
Other languages
Chinese (zh)
Inventor
林肇堃
魏苏龙
麦绮兰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority to CN201811000969.3A priority Critical patent/CN109240639A/en
Publication of CN109240639A publication Critical patent/CN109240639A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The embodiment of the present application discloses acquisition methods, device, storage medium and the terminal of a kind of audio data, this method comprises: receiving audio data obtains signal;Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;Denoising is carried out to first audio data, obtains target audio data, can be improved the accuracy that intelligent wearable device obtains audio data.

Description

Acquisition methods, device, storage medium and the terminal of audio data
Technical field
The invention relates to the acquisition methods of intelligent terminal technical field more particularly to audio data, device, storages Medium and terminal.
Background technique
With the continuous development of intelligent wearable device, have the function of intelligent terminal, and is able to satisfy portability and large screen view Feel that the intelligent wearable device of demand for experience gradually obtains the approval of user.
Currently, intelligent wearable device in noisy environment, can not accurately obtain the audio data of user's sending, and then influence The normal use of user.
Summary of the invention
The purpose of the embodiment of the present application is to provide acquisition methods, device, storage medium and the terminal of a kind of audio data, can To improve the accuracy that intelligent wearable device obtains audio data.
In a first aspect, the embodiment of the present application provides a kind of acquisition methods of audio data, comprising:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, then the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Second aspect, the embodiment of the present application provide a kind of acquisition device of audio data, comprising:
Signal receiving module obtains signal for receiving audio data;
First audio obtains module, and the audio data for being received according to the signal receiving module obtains letter Number, the first audio data is obtained by osteoacusis mode;
Denoising module is denoised for obtaining first audio data that module obtains to first audio Processing, obtains target audio data.
The third aspect, the embodiment of the present application provide a kind of computer readable storage medium, are stored thereon with computer journey Sequence realizes the acquisition methods of the audio data as shown in first aspect when the program is executed by processor.
Fourth aspect, the embodiment of the present application provide a kind of terminal, including memory, and processor and storage are on a memory And the computer program that can be run in processor, the processor are realized as shown in first aspect when executing the computer program Audio data acquisition methods.
The acquisition scheme of the audio data provided in the embodiment of the present application, first reception audio data obtain signal;Then, Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;Finally, to the first audio number According to denoising is carried out, target audio data are obtained, can be improved the accuracy that intelligent wearable device obtains audio data.
Detailed description of the invention
Fig. 1 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application;
Fig. 2 is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 3 A is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 3 B is that user provided by the embodiments of the present application passes through intelligent wearable device and Remote progress remote guide Schematic diagram of a scenario;
Fig. 4 is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 5 is the flow diagram of the acquisition methods of another audio data provided by the embodiments of the present application;
Fig. 6 is a kind of structural schematic diagram of the acquisition device of audio data provided by the embodiments of the present application;
Fig. 7 is a kind of structural schematic diagram of intelligent wearable device provided by the embodiments of the present application;
Fig. 8 is a kind of signal pictorial diagram of intelligent wearable device provided by the embodiments of the present application;
Fig. 9 is a kind of structural schematic diagram of terminal device provided by the embodiments of the present application.
Specific embodiment
Further illustrate the technical solution of the application below with reference to the accompanying drawings and specific embodiments.It is understood that It is that specific embodiment described herein is used only for explaining the application, rather than the restriction to the application.It further needs exist for illustrating , part relevant to the application is illustrated only for ease of description, in attached drawing rather than entire infrastructure.
It should be mentioned that some exemplary embodiments are described as before exemplary embodiment is discussed in greater detail The processing or method described as flow chart.Although each step is described as the processing of sequence by flow chart, many of these Step can be implemented concurrently, concomitantly or simultaneously.In addition, the sequence of each step can be rearranged.When its operation The processing can be terminated when completion, it is also possible to have the additional step being not included in attached drawing.The processing can be with Corresponding to method, function, regulation, subroutine, subprogram etc..
With the continuous development of intelligent wearable device, there is cell-phone function, and be able to satisfy portability and large screen vision body The intelligent wearable device for testing demand gradually obtains the approval of user.Currently, intelligent wearable device noisy environment (e.g., noise compared with Big workshop, concert scene etc.) in, the audio data noise of acquisition is larger, so that user's input can not be obtained accurately Audio data, and then influence user to the normal use and control of intelligent wearable device.
The embodiment of the present application provides a kind of acquisition methods of audio data, can obtain signal receiving audio data Afterwards, the first audio data is obtained by osteoacusis mode, obtains target audio data after carrying out denoising.And then avoid user When speaking in noisy environment, can not accurately obtain user input audio data the problem of, improve intelligent wearable device obtain The accuracy of audio data.Concrete scheme is as follows:
Fig. 1 is the flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, and this method is for intelligence Wearable device obtains the case where audio data of user's input, and this method can be set by the intelligence wearing for being equipped with bone conduction device Standby or terminal device executes, for example, the intelligent wearable device can be intelligent glasses, intelligent helmet, Intelligent bracelet etc., terminal Equipment can be smart phone, tablet computer, laptop etc..This method specifically comprises the following steps:
Step 110 receives audio data acquisition signal.
Wherein, audio data, which obtains signal, can be intelligent wearable device or other terminal device (such as mobile phones, tablet computer Deng) receive the trigger action of user and generate for indicating that user wants the signal of input audio data.Optionally, if sound Frequency data acquisition signal is the signal that user triggers the generation of other terminal devices, then other terminal devices can be by the audio number of generation Intelligent wearable device is sent to by terminal device according to signal is obtained.For example, user clicks the record button on mobile phone, then can touch Hair generates audio data and obtains signal, and the audio data of generation can be obtained signal and be sent to intelligent wearable device by mobile phone at this time.
Optionally, in the embodiment of the present application, the mode that triggering generates that audio data obtains signal has very much, the application couple This without limiting, can be user click it is corresponding with audio signal is obtained on intelligent wearable device or other terminal devices Fixed key for example, needing to obtain the audio data of user's input when recording, therefore can be worn when detection user clicks intelligence When wearing record button in equipment, i.e., triggering generates audio data and obtains signal;Can also be detection intelligent wearable device or other Whether preset affiliate application starts in terminal device, and optionally, affiliate application may include voice/video chat Program, recorded program, remote guide program.For example, will start video call function when opening remote guide program, at this time must It will be related to the acquisition of audio user information, therefore, affiliate application can be set by remote guide program, when detecting When remote guide program starts, i.e., triggering generates audio data and obtains signal;The bone for being also possible to detect intelligent wearable device passes It leads and whether produces bone vibration at equipment, for example, when user speaks, process of the audio signal based on osteoacusis can cause bone to shake Dynamic, there are when bone vibration at detection bone conduction device place, i.e., triggering generates audio data and obtains signal.
Optionally, in the embodiment of the present application, reception audio data acquisition signal, which can be, is detecting that intelligently wearing is set The standby receive capabilities for starting or opening when being kept in motion audio data and obtain signal, for example, the intelligence wearing that user wears When equipment starts or is kept in motion, usually indicate that user will use the intelligent wearable device, it is therefore more likely that can trigger tone Frequency data acquisition signal such as issues phonetic control command.Therefore shape can be moved when detecting intelligent wearable device starting or being in When state, the receive capabilities that audio data obtains signal are opened, audio data is received and obtains signal.Optionally, audio number is being received When according to obtaining signal, it can be and be received once every preset time (such as 5 minutes), be also possible to real-time perfoming reception, Zhi Daozhi It can wearable device closing.
Further, it when receiving audio data acquisition signal, can be in current location information and preset position information Timing receives audio data and obtains signal.
Wherein, preset position information can be that user is preset according to self-demand, can also be system default, For example, it may be in view of being not easy to obtain the audio data of user's input in noisy environment, therefore can be by noisy environment, such as Market, stadium, street etc. are set as preset position information.Current location information refers to user position information, can To obtain current location information by the navigation positioning unit in intelligent wearable device, can also be interacted with intelligent terminal, Current location information is obtained by the navigation positioning unit on intelligent terminal.
Optionally, if current location information is matched with preset position information, illustrate that current location is in noisy environment, then Start audio data and obtain semiotic function, the audio data of user's input is obtained by osteoacusis mode.If current location information It is mismatched with preset position information, then illustrates that current location is not belonging to noisy environment, for example, user is at home, at this time by normal The air transmitted mode of rule can accurately obtain the audio data of user's input.
Step 120 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Wherein, osteoacusis mode is the slight vibration of caused incidence bone when being talked using people voice signal is received Collection gets up to switch to the mode of electric signal.Sound is acquired by air transmitted since it is different from conventional microphone, so in very noise The outflow of sound high-resolution can also be come in miscellaneous environment.Optionally, the embodiment of the present application can by bone conduction device come The first audio data is obtained, if be intelligent glasses when intelligence wearing, bone conduction device be can be set in intelligent glasses temple portion Position, when wearing, the bone conduction device that intelligent glasses temple position is arranged in contacts user with the skull near user's ear, can Voice signal is collected to shake well by skull.First audio data is the audio data of user's input.
In the embodiment of the present application, if step 110 receives audio data and obtains signal, start bone conduction device Speech voice input function collects the audio data of user's input, first obtained by osteoacusis mode by osteoacusis mode Audio data, clarity is higher, will not enter bone conduction device for the noisy noise in environment by air transmitted.For example, User is sought help using the remote guide function of intelligent wearable device to Remote in the biggish shop maintenance machine of noise, this When, it can be very good to filter out machine sound in noisy environment in the audio data of the user's input obtained by bone conduction device Sound ensure that the user of acquisition inputs the clarity of sound.
Optionally, if received audio data, which obtains signal, to be generated by detection associated program starting, due to association Application program is the software that may relate to obtain audio data in preset operational process, but does not represent starting association and answer With audio data is centainly obtained after program, therefore, a time interval can be preset, if in the prefixed time interval, it is bone-free It shakes (i.e. user does not have input audio data), then closes bone conduction device, wait until that receiving audio data next time obtains signal When, restart the audio data that bone conduction device obtains user's input.For example, detect that user opens associated program " wechat ", It then generates audio data and obtains signal, starting bone conduction device obtains user's input after receiving audio data and obtaining signal Audio data, if user only sees wechat circle of friends, and there is no voice-enabled chat or views in prefixed time interval (such as 1 minute) Frequency is conversed, and closes bone conduction device at this time.If receive after a while due to detecting that user triggers the video in wechat again Talk button and the audio data that generates obtains signal, then be again started up the audio data that bone conduction device obtains user's input.
It should be noted that needing to detect intelligence wearing before obtaining the first audio data by osteoacusis mode and setting Whether standby bone conduction device is normal, if contacts with user's skull, if bone conduction device abnormal state or not and with account Bone contact, then bone conduction device is the audio data that user's input can not be obtained by osteoacusis mode, is lost in order to prevent at this time The audio data for leaking user's input needs to start microphone (microphone based on control conduction pattern work) or spare osteoacusis Equipment temporarily to obtain the audio data of user's input and prompt user has dressed intelligent wearable device etc..
Step 130 carries out denoising to the first audio data, obtains target audio data.
It is smaller by Environmental Noise Influence although obtaining the first audio data by osteoacusis mode, obtaining by the Still there can be certain noise in one audio data, the sound of first-class generation for example, user coughs, grits one's teeth, scratching, these sound May be unobvious by air transmitted mode, but obtained then very clearly by osteoacusis mode, therefore, passed for passing through bone The audio data that the mode of leading obtains is still to carry out denoising, and the invalid data removed in audio data (such as coughs, grits one's teeth, scratching First-class sound), obtain not having noisy clarity high uses the first audio data as target audio data.
In the embodiment of the present application, the mode for carrying out denoising to the first audio data has very much, not to this application It is defined, can be and pre- first pass through a large amount of audio data sample (comprising the audio data with noise and muting audio Data) training filter, denoising is carried out by trained filter, for example, the filter of training can be minimum Square error (LMS) filter.When carrying out denoising to the first audio data, the first audio number that analysis obtains can also be According to audio frequency characteristics (such as tone color, tone, frequency, loudness), the audio of the audio data inputted due to noise and user is special Sign is different, therefore can be corresponding by the first audio frequency characteristics of variation abnormality according to the situation of change of audio data sound intermediate frequency feature Audio data (i.e. noise section) removal.For example, the 2nd second to the 5th second audio frequency characteristics in the first audio data of acquisition Variation abnormality then illustrates that the period corresponding audio data is noise, needs to be deleted.
Optionally, it since obtained target audio data only include the audio data of user's input, and makes an uproar not comprising environment Sound, therefore, if obtained target audio data are the dialog contexts of user and other users, the target audio that can will be obtained Data are sent to call another party after carrying out data encryption.If obtained target audio data are the recording number of user's input According to can then obtained target audio data be stored in the corresponding storage unit of recorded program.If obtained target audio Data are control instruction of the user to intelligent wearable device or intelligent terminal, then intelligent wearable device can be directly in response to obtaining Target audio data are sent to intelligent terminal, so that intelligent terminal responds the target audio data.For example, working as target audio Data are " weather forecast ", then intelligent wearable device can respond starting weather prompting function at once, export current weather to user.
The acquisition methods of the audio data provided in the embodiment of the present application, first reception audio data obtain signal;Then, Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;Finally, to the first audio number According to denoising is carried out, target audio data are obtained.Audio data is obtained relative to by air transmitted mode, the application is implemented Example can obtain the first audio data by osteoacusis mode, after carrying out denoising after receiving audio data and obtaining signal Obtain target audio data.And then when avoiding user and speaking in noisy environment, the audio of user's input can not be accurately obtained The problem of data, improves the accuracy that intelligent wearable device obtains audio data.
Fig. 2 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, is suitable for obtaining The case where when audio data, not only obtaining user, what is said or talked about, also to obtain ambient sound, for example, user thinks one section of record certainly When the section of football match video that oneself explains, the live sound of football match should be obtained, also to obtain the sound of oneself explanation.As Further explanation to above-described embodiment, comprising:
Step 210 receives audio data acquisition signal.
The embodiment of the present application will not only execute step 210 and pass through osteoacusis mode after receiving audio data and obtaining signal The first audio data is obtained, step 230 is also executed and obtain environmental audio data.Illustratively, user is watching a football When match, the live video that one section of record explains oneself is thought, at this point, step 220 should be executed, obtain the sound of oneself explanation, also Step 230 is executed, the live sound of football match is obtained, to restore true match scene.
Step 220 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Illustratively, it can be very good to filter out noisy ambient sound by bone conduction device, get user's input Audio data.
Step 230 obtains signal acquisition environmental audio data according to audio data.
Illustratively, environmental audio data can be obtained by the microphone to be worked based on air transmitted mode, i.e., existing Intelligent wearable device obtain audio data mode, can be arranged in intelligent wearable device at least one be based on air transmitted Microphone that mode works obtains ambient sound.
Step 240 generates second audio data according to environmental audio data and the first audio data.
Wherein, second audio data is counted after environmental audio data and the first audio data are carried out fusion treatment According to optionally, carrying out the mode that fusion treatment generates second audio data to environmental audio data and the first audio data has very More, the embodiment of the present application without limiting, can be the volume of the volume of environmental audio data and the first audio data this After being adjusted, it is fused together generation second audio data, for example, the football live audio data that will acquire is turned down, will be used After the explanation audio data of family input tunes up, two-way audio data fusion is obtained into second audio data together.
Optionally, it is contemplated that also include the first audio data in the environmental audio data of acquisition, only it is covered in noise It in miscellaneous ambient sound, can not distinguish well, therefore generate the according to environmental audio data and the first audio data When two audio datas, it can also be the first audio data of part replaced the first audio data in environmental audio data, obtain To second audio data.For example, explaining audio data, the use in football live audio data that will acquire according to the user of acquisition Family explains audio data and filters out, and the user of the football live audio data after then filtering out and acquisition explains audio data fusion Together, second audio data is obtained.
Step 250 carries out denoising to second audio data, obtains target audio data.
Optionally, when carrying out denoising to second audio data, noise in the first audio data is not only filtered out (such as Cough, grit one's teeth, scratching first-class sound), when also generating second audio data caused by new noise.For example, if the second audio number According to being to merge to generate after the first audio data and environmental audio data to be carried out to volume adjustment, then can go out in second audio data The first audio data being now overlapped (is explained in audio data and environmental audio data including the user that bone conduction device obtains and is wrapped The user contained explains audio data), denoising at this time can be the first audio data portion in environmental audio data It filters out.
Optionally, if second audio data is the first audio data replaced the first audio data in environmental audio data It generates, is then likely to occur ambient sound in second audio data and is greater than the first audio data, it is excessive so as to cause ambient noise, The case where can't hear the first audio data, denoising at this time, which can be, reduces the volumes of environmental audio data, improves the The volume of one audio data.
The acquisition methods of audio data provided by the embodiments of the present application can lead to after receiving audio data and obtaining signal It crosses osteoacusis mode and air transmitted mode obtains audio data jointly and generates second audio data, second audio data is carried out Target audio data are obtained after denoising.It can be improved the accuracy that intelligent wearable device obtains audio data.
Fig. 3 A is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, and Fig. 3 B is this Shen Please the user that provides of embodiment the schematic diagram of a scenario of remote guide is carried out by intelligent wearable device and Remote.The application is real It applies example and is not only used to obtain audio data, but also the feelings for exporting audio data suitable for the bone conduction device of intelligent wearable device Condition.For example, bone conduction device both send out by available user's word when user uses the remote guide function of intelligent wearable device Long-range expert is given, and expert's word can be exported to user.Further explanation in conjunction with Fig. 3 B to above-described embodiment, Include:
Step 310 starts.
Step 320 judges whether that receiving audio data obtains signal, if so, step 330 is executed, if it is not, executing step 350。
Since bone conduction device not only has the function of obtaining audio data, also have the function of transmitting audio data, and It can replace earphone and loudspeaker, audio data to be played exported to user.Audio data is transmitted by bone conduction device, Ears can be liberated, also it will be noted that the variation of ambient enviroment while receiving audio data, use more safety for example, and Other side can be also allowed to be clear from the instruction of expression when speaking volume is smaller.
Optionally, it in order to enable can also accurately obtain in a noisy environment and playing audio-fequency data, can be When receiving audio data acquisition signal, step 330 is executed, the first audio data is obtained by osteoacusis mode;Do not connecing When receiving audio data acquisition signal, step 350 is executed, audio data to be played is exported by osteoacusis mode.Optionally, Audio data, which obtains signal and can be, at this time shakes what triggering generated by bone.For example, user is when using remote guide function, it can By being shaken at detection bone conduction device with the presence or absence of bone, and if it exists, illustrate that user is speaking, generate audio data and obtain Signal, the audio data for receiving generation obtain execution step 330 after signal, the sound of user's input are obtained by osteoacusis mode Frequency evidence, bone shakes if it does not exist, then illustrates that user in the guidance for listening expert, executes step 350 at this time, passes through osteoacusis mode Audio data is instructed to user output expert.
Step 330 obtains signal according to the audio data, obtains the first audio data by osteoacusis mode.
Illustratively, in user and Remote communication process, when a user speaks, sound, which is based on osteoacusis mode, to be drawn Bone vibration is played, triggering generates audio data and obtains signal when generating bone vibration, and the audio data for obtaining generation obtains letter After number, then it can control bone conduction device and be based on osteoacusis mode and obtain user what is said or talked about.
Step 340 carries out denoising to the first audio data, obtains target audio data.
Step 350 obtains audio data to be played, exports audio data to be played by osteoacusis mode.
Wherein, audio data to be played, which can be, needs the audio data that plays to user, when being not received by audio When data acquisition signal, the audio data remotely sended over that intelligent wearable device meeting automated communications unit obtains, and pass through The audio data that bone conduction device will acquire is based on osteoacusis mode and exports to user.For example, audio data to be played is to use Family use remote guide function when, what is said or talked about for Remote, user in Remote communication process, if user does not say When words, bone vibration will not be generated, i.e. acquisition audio data obtains signal, at this point, expert's guidance that bone conduction device can will acquire Audio data is based on osteoacusis mode and exports to user.
Optionally, the audio data to be played that osteoacusis mode exports in the embodiment of the present application is not limited to user and leads to Dialog context when words can also be that the playing function that bone conduction device is replaced to earphone, data to be played at this time can be Music to be played.
If step 360 when exporting audio data to be played by osteoacusis mode, receives audio data and obtains letter Number, then start spare osteoacusis mode and obtains third audio data.
Wherein, third audio data is also the audio data of user's input, which, which can be, is passing through osteoacusis The audio data that user inputs when mode exports audio data to be played.
Optionally, the bone conduction device in intelligent wearable device can be set multiple, and one is main bone conduction device, remaining For spare bone conduction device, main bone conduction device is based on main osteoacusis mode and works, and backup bone conduction device is based on backup Osteoacusis mode works.Wherein, when needing to carry out the acquisition and output of audio data, main osteoacusis is selected to set first It is standby, when main bone conduction device is working or when abnormal state, enables backup bone conduction device, can guarantee to set even if osteoacusis When for damage or being used, the audio data of user's input also can be accurately obtained.
Optionally, even if current time main bone conduction device, which is in, plays operating mode, audio data acquisition is also received Signal, if being received when main bone conduction device exports audio data to be played to user based on main osteoacusis mode Audio data obtains signal, then main bone conduction device continues to execute and exports audio data to be played to user at this time, starts standby Equipment is passed to bone, the third audio data that user inputs is obtained based on backup osteoacusis mode, and execute step 340, to acquisition Audio data carry out denoising, obtain target audio data.For example, user, which listens main bone to pass to equipment, sends back the expert come During guidance, query is produced, he may interrupt him during expert speaks, at this point, main bone conduction device also exists Play expert guidance, and receive because user speak generation voice data obtain signal, then will start spare osteoacusis and set It is standby that user's audio data currently entered is obtained based on backup osteoacusis mode, and after carrying out denoising, by remotely leading to Letter is sent to equipment where expert.
The acquisition methods of audio data provided by the embodiments of the present application can lead to when receiving audio data acquisition signal It crosses osteoacusis mode and obtains the first audio data, when not triggering audio acquisition signal, exported by osteoacusis mode to be played Audio data.The acquisition of audio data is completed by osteoacusis mode and plays two functions, and improves intelligent wearable device Obtain the accuracy with playing audio-fequency data.
Fig. 4 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, as to above-mentioned The further explanation of embodiment, comprising:
Step 410 receives audio data acquisition signal.
Step 420 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Step 430 extracts the corresponding audio frequency characteristics of the first audio data.
Wherein, acoustic characteristic may include one or more in the corresponding tone color of audio data, tone, frequency, loudness etc. A, the audio frequency characteristics of the difference of sound source, corresponding audio data are also different, for example, the sound that user speaks in the first audio data Frequency feature audio frequency characteristics corresponding with sound difficult to tackle are entirely different.Sound source is identical, and the audio frequency characteristics of corresponding audio data are also different Fixed identical, for example, user speaks and coughs, the corresponding audio frequency characteristics of sound issued are also different.
It optionally, may incessantly include the corresponding audio data of a sound source in the first audio data that step 420 obtains, The corresponding each audio frequency characteristics of the whole section audio of the first audio data are extracted respectively, for example, both saying including user in whole section audio data The audio of words further includes the audio that user scratches one's head, then extracts the audio frequency characteristics that user speaks respectively and the audio that user scratches one's head is special Sign.Optionally, the corresponding audio frequency characteristics of the first audio data are extracted, can be and carried out by trained audio feature extraction model It extracts, can also be and extracted by preset audio feature extraction algorithm, the application is to this without limiting.
Step 440, according to audio frequency characteristics, whether there is invalid data in audio data.
Wherein, invalid data refers to the data of non-user voice input in the first audio data extracted, for example, user The sound etc. that the sound difficult to tackle generated or user's cough generate.
Optionally, according to audio frequency characteristics, it whether there is invalid data in audio data, can be and acquire and deposit in advance The audio frequency characteristics for storing up a variety of invalid datas, when judging, the audio frequency characteristics for the first audio data that will acquire with it is pre-stored The audio frequency characteristics of a variety of invalid datas are compared, and determine the invalid data in the first audio data obtained, for example, if obtaining The first audio data in have the tone of section audio data and tone color tone corresponding with sound difficult to tackle and tone color identical, then the section Audio data is invalid data.The every kind of audio frequency characteristics (such as frequency) for being also possible to the first audio data that will acquire are depicted as Full curve analyzes the tendency of curve, and the corresponding audio data in part that tendency mutates is considered as invalid data.
Step 450, if there is invalid data, then filter out the invalid data in the first audio data, obtain target audio Data.
Optionally, if there are invalid datas in the first audio data obtained, in the first audio data of acquisition Except ineffective data part, after remaining audio data is spliced, it can be completed and the first audio data of acquisition is gone It makes an uproar the process of processing, target audio data can be obtained.
The acquisition methods of audio data provided by the embodiments of the present application can lead to after receiving audio data and obtaining signal It crosses osteoacusis mode and obtains the first audio data, according to the audio frequency characteristics of the first audio data, filter out in the first audio data Invalid data obtains target audio data.And then when user being avoided to speak in noisy environment, user's input can not be accurately obtained Audio data the problem of, improve intelligent wearable device obtain audio data accuracy.
Fig. 5 is a kind of flow diagram of the acquisition methods of audio data provided by the embodiments of the present application, and the application is implemented Example is suitable for the case where user speech controls intelligent wearable device, as the further explanation to above-described embodiment, packet It includes:
Step 510 receives audio data acquisition signal.
Step 520 obtains signal according to audio data, obtains the first audio data by osteoacusis mode.
Step 530 carries out denoising to the first audio data, obtains target audio data.
Step 540 carries out intention analysis to target audio data.
Optionally, target audio data are carried out with the process for being intended to analysis, semantic point as is carried out to the first audio data The process of analysis.The analysis that can be intended to by the intention analysis model pre-established in the embodiment of the present application.Wherein, it is intended that Analysis model, which can be, is arranged keyword as much as possible for all control instructions in advance, and establishes control instruction and keyword Between mapping relations, input neural network model learnt.In the intention point for actually carrying out target audio data When analysis, it can be and first target audio data are segmented, the keyword after participle is then inputted into trained intention and is analyzed In model, model analyzes the keyword of input based on learning algorithm when training, the corresponding intention of output keyword.Example Such as, target audio data are " today, how is weather? ", to the word segmented to obtain " today ", " weather ", " how Sample ", and these three words are inputted and are intended to analysis model, it obtains being intended to analysis result being that user wants to understand the weather of today.
Step 550, response are intended to the corresponding control instruction of analysis result.
Optionally, the corresponding control instruction of target audio data is determined according to semantic analysis result, is then responding to the control Instruction, such as, it is intended that analysis result is that user wants to understand the weather of today, and the corresponding control instruction of the intention is " starting weather Prompt ", then system responds the instruction, starts weather prompting function, exports current weather to user.
Optionally, the corresponding control instruction of target audio data can be the control to intelligent wearable device of user's sending System instruction can also be the control instruction to terminal devices such as mobile phones that user issues.To this without limiting, if control instruction For the corresponding control instruction of terminal device, intelligent wearable device can be logical between terminal device by it by the control instruction Letter connection, is sent to terminal device, so that terminal device executes the control instruction.
The acquisition methods of audio data provided by the embodiments of the present application can lead to after receiving audio data and obtaining signal It crosses osteoacusis mode and obtains the first audio data, denoised and be intended to analyze, thus control of the response to intelligent wearable device. The accuracy that intelligent wearable device obtains audio data is improved, accurately intelligence can be worn by voice in a noisy environment Equipment is worn to be controlled.
Fig. 6 is a kind of structural schematic diagram of the acquisition device of audio data provided by the embodiments of the present application.As shown in fig. 6, The device includes: signal detection module 610, and audio obtains module 620, denoising module 630.
Signal receiving module 610 obtains signal for receiving audio data;
First audio obtains module 620, and the audio data for being received according to the signal receiving module 610 obtains The number of winning the confidence obtains the first audio data by osteoacusis mode;;
Denoising module 630, for first audio obtain module 620 obtain first audio data into Row denoising obtains target audio data.
Further, above-mentioned apparatus further include:
Environmental audio obtains module, for obtaining signal acquisition environmental audio data according to the audio data;
Correspondingly, denoising module 630 is used for, generated according to the environmental audio data and first audio data Second audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, above-mentioned apparatus further include:
Audio output module exports the audio to be played by osteoacusis mode for obtaining audio data to be played Data.
Further, the first audio obtains module 620 and is also used to, if exported by the osteoacusis mode described wait broadcast When putting audio data, receives audio data and obtain signal, then start spare osteoacusis mode and obtain third audio data.
Further, denoising module 630 is used for, and extracts the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, above-mentioned apparatus further include:
It is intended to analysis module, for carrying out intention analysis to the target audio data;
Respond module is instructed, for responding the corresponding control instruction of the intention analysis result.
Further, signal receiving module 610 is also used to, and when current location information is matched with preset position information, is connect It receives audio data and obtains signal.
The acquisition device of the audio data provided in the embodiment of the present application, firstly, signal receiving module 610 receives audio number According to acquisition signal;Then, the first audio obtains module 620 and obtains signal according to the audio data, is obtained by osteoacusis mode Take the first audio data;Finally, denoising module 630 carries out denoising to first audio data, target sound is obtained Frequency evidence.The first audio data is obtained relative to by air transmitted mode, the embodiment of the present application can receive audio data After obtaining signal, the first audio data is obtained by osteoacusis mode, obtains target audio data after carrying out denoising.In turn When user being avoided to speak in noisy environment, can not accurately obtain user input audio data the problem of, improve intelligence wearing The accuracy of equipment acquisition audio data.
Method provided by the aforementioned all embodiments of the application can be performed in above-mentioned apparatus, and it is corresponding to have the execution above method Functional module and beneficial effect.The not technical detail of detailed description in the present embodiment, reference can be made to the aforementioned all implementations of the application Method provided by example.
The present embodiment provides a kind of intelligent wearable device on the basis of the various embodiments described above, and Fig. 7 is that the application is implemented A kind of structural schematic diagram for intelligent wearable device that example provides, Fig. 8 is a kind of intelligent wearable device provided by the embodiments of the present application Signal pictorial diagram.As shown in Figure 7 and Figure 8, which includes: memory 701, processor (Central Processing Unit, CPU) 702, display unit 703, touch panel 704, heart rate detection mould group 705, range sensor 706, camera 707, bone-conduction speaker 708, microphone 709, breath light 710, these components pass through one or more communications Bus or signal wire 711 communicate.
It should be understood that diagram intelligent wearable device is only an example of intelligent wearable device, and intelligently wear Wearing equipment can have than shown in the drawings more or less component, can combine two or more components, or Person can have different component configurations.Various parts shown in the drawings can include one or more signal processings and/ Or it is realized in the combination of hardware including specific integrated circuit, software or hardware and software.
Just the intelligent wearable device provided in this embodiment obtained for audio data is described in detail below, the intelligence Energy wearable device is by taking intelligent glasses as an example.
Memory 701, the memory 701 can be accessed by CPU702, and the memory 701 may include that high speed is random Access memory, can also include nonvolatile memory, for example, one or more disk memory, flush memory device or its His volatile solid-state part.
Display unit 703, can be used for the operation and control interface of display image data and operating system, and display unit 703 is embedded in In the frame of intelligent wearable device, frame is internally provided with inner transmission lines 711, the inner transmission lines 711 and display Component 703 connects.
Touch panel 704, which is arranged in the outside of at least one intelligent wearable device temple, for obtaining Touch data is taken, touch panel 704 is connected by inner transmission lines 711 with CPU702.Wherein, touch panel 704 is detectable The finger sliding of user, clicking operation, and the data detected are transmitted to processor 702 accordingly and are handled with generation pair The control instruction answered, illustratively, can be left shift instruction, right shift instruction, move up instruction, move down instruction etc..Illustratively, it shows Show component 703 can video-stream processor 702 transmit virtual image data, which can be accordingly according to touch panel 704 user's operations that detect carry out corresponding changes, specifically, can be carry out screen switching, when detect left shift instruction or Switch upper one or next virtual image picture after right shift instruction accordingly;When display unit 703 shows video playing information When, which, which can be, plays out playbacking for content, and right shift instruction can be the F.F. for playing out content;Work as display The display of component 703 is when being editable word content, and the left shift instruction, right shift instruction move up instruction, move down instruction and can be pair The displacement operation of cursor, the i.e. position of cursor can move the touch operation of touch tablet according to user;Work as display unit When the contents of 703 displays are game animation picture, the left shift instruction, right shift instruction move up instruction, move down instruction and can be to trip Object in play is controlled, in machine game like flying, can by the left shift instruction, right shift instruction, move up instruction, move down instruction point Not Kong Zhi aircraft heading;When display unit 703 can show the video pictures of different channel, which is moved to right It instructs, moves up instruction, moves down the switching for instructing and can carrying out different channel, wherein moving up instruction and moving down instruction can be switching To pre-set channel (the common channel that such as user uses);When display unit 703 shows static images, which is moved to right It instructs, moves up instruction, moves down the switching that instructs and can carry out between different pictures, wherein left shift instruction can be to switch to one Width picture, right shift instruction, which can be, switches to next width figure, and an atlas can be to switch to by moving up instruction, and moving down instruction can be with It is to switch to next atlas.The touch panel 704 can also be used to control the display switch of display unit 703, exemplary , when long pressing 704 touch area of touch panel, display unit 703, which is powered, shows graphic interface, when long pressing touch again When 704 touch area of panel, display unit 703 is powered off, can be by carrying out in touch panel 704 after display unit 703 is powered Upper cunning and operation of gliding are to adjust the brightness or resolution ratio that show image in display unit 703.
Heart rate detection mould group 705, for measuring the heart rate data of user, heart rate refers to beats per minute, the heart rate Mould group 705 is detected to be arranged on the inside of temple.Specifically, the heart rate detection mould group 705 can be in such a way that electric pulse measures Human body electrocardio data are obtained using stemness electrode, heart rate size is determined according to the amplitude peak in electrocardiogram (ECG) data;The heart rate detection Mould group 705 can also be by being formed using the light transmitting and light receiver of photoelectric method measurement heart rate, correspondingly, the heart rate is examined Mould group 705 is surveyed to be arranged at temple bottom, the ear-lobe of human body auricle.Heart rate detection mould group 705 can phase after collecting heart rate data The progress data processing in processor 702 that is sent to answered has obtained the current heart rate value of wearer, in one embodiment, processing Device 702, can be by the heart rate value real-time display in display unit 703 after determining the heart rate value of user, optional processor 702 are determining that heart rate value lower (such as less than 50) or higher (such as larger than 100) can trigger alarm accordingly, while by the heart Rate value and/or the warning message of generation are sent to server by communication module.
Range sensor 706, may be provided on frame, the distance which is used to incude face to frame, The realization of infrared induction principle can be used in the range sensor 706.Specifically, the range sensor 706 is by the range data of acquisition It is sent to processor 702, data control the bright dark of display unit 703 to processor 702 according to this distance.Illustratively, work as determination When the collected distance of range sensor 706 is less than 5 centimetres out, the corresponding control display unit 703 of processor 702, which is in, to be lighted State, when determine range sensor be detected with object close to when, it is corresponding control display unit 704 and be in close shape State.
Breath light 710 may be provided at the edge of frame, when display unit 703 closes display picture, the breath light 710 It can be lighted according to the control of processor 702 in the bright dark effect of gradual change.
Camera 707 can be the position that the upper side frame of frame is arranged in, and acquire the proactive of the image data in front of user As module, the rear photographing module of user eyeball information can also be acquired, is also possible to the combination of the two.Specifically, camera 707 When acquiring forward image, the image of acquisition is sent to the identification of processor 702, processing, and trigger accordingly according to recognition result Trigger event.Illustratively, when user wears the wearable device at home, by being identified to the forward image of acquisition, If recognizing article of furniture, corresponding inquiry whether there is corresponding control event, if it is present accordingly by the control The corresponding control interface of event processed is shown in display unit 703, and user can carry out corresponding furniture object by touch panel 704 The control of product, wherein the article of furniture and intelligent wearable device are connected to the network by bluetooth or wireless self-networking;Work as user When wearing the wearable device outdoors, target identification mode can be opened accordingly, which can be used to identify specific People, camera 707 by the image of acquisition be sent to processor 702 carry out recognition of face processing, if recognizing the pre- of setting If face, then the loudspeaker that can be integrated accordingly by intelligent wearable device carries out sound casting, which may be used also With plants different for identification, for example, processor 702 is adopted according to the touch operation of touch panel 704 with recording camera 707 The present image of collection is simultaneously sent to server by communication module to be identified, server carries out the plant in acquisition image It identifies and feeds back relevant botanical name, introduce to intelligent wearable device, and feedback data is shown in display unit 703. Camera 707 can also be the image for acquiring user's eye such as eyeball, be generated by the identification of the rotation to eyeball different Control instruction illustratively move up control instruction as eyeball is rotated up generation, eyeball, which rotates down generation and moves down control, to be referred to It enables, the eyeball generation that turns left moves to left control instruction, and the eyeball generation that turns right moves to right control instruction, wherein display unit 703 Can video-stream processor 702 transmit virtual image data, what which can detect according to camera 707 accordingly Control instruction that the mobile variation of user eyeball generates and change, specifically, can be carry out screen switching, moved to left when detecting Control instruction switches upper one or next virtual image picture after moving to right control instruction accordingly;When display unit 703 is aobvious When showing video playing information, this, which moves to left control instruction and can be, plays out playbacking for content, move to right control instruction can be into The F.F. of row broadcasting content;When the display of display unit 703 is editable word content, this moves to left control instruction, moves to right control System instruction moves up control instruction, moves down control instruction and can be displacement operation to cursor, i.e. the position of cursor can be according to user The touch operation of touch tablet is moved;When the content that display unit 703 is shown is game animation picture, this moves to left control System instruction moves to right control instruction, moves up control instruction, moving down control instruction and can be and control the object in game, such as In aircraft game, control instruction can be moved to left by this, control instruction is moved to right, moves up control instruction, moving down control instruction and control respectively The heading of aircraft processed;When display unit 703 can show the video pictures of different channel, this moves to left control instruction, moves to right Control instruction moves up control instruction, moves down control instruction and can carry out the switching of different channel, wherein move up control instruction and under Pre-set channel (the common channel that such as user uses) can be to switch to by moving control instruction;When display unit 703 shows static map When piece, this moves to left control instruction, moves to right control instruction, moves up control instruction, moving down control instruction and can carry out between different pictures Switching, wherein a width picture can be to switch to by moving to left control instruction, moved to right control instruction and be can be and switch to next width Figure, an atlas can be to switch to by moving up control instruction, moved down control instruction and be can be and switch to next atlas.
The inner wall side of at least one temple is arranged in bone-conduction speaker 708, bone-conduction speaker 708, for that will receive To the audio signal that sends of processor 702 be converted to vibration signal, or for the vibration signal received to be converted to audio Signal is sent to processor 702.Wherein, sound can be transferred to human body inner ear by skull by bone-conduction speaker 708, pass through by The electric signal of audio is changed into vibration signal and is transmitted in skull cochlea, then is perceived by auditory nerve.It can also will be by bone The sound of transmitting is sent to processor 702, after vibration signal is changed into the electric signal of audio, everywhere by electric signal transmitting Manage device 702.It is used as sounding device or sound receiver by bone-conduction speaker 708, reduces hardware configuration thickness, weight It is lighter, while electromagnetic-radiation-free will not be influenced by electromagnetic radiation, and have antinoise, waterproof and liberation ears A little.
Microphone 709, may be provided on the lower frame of frame, for acquiring external (user, environment) sound and being transmitted to Processor 702 is handled.Illustratively, the sound that microphone 709 issues user be acquired and pass through processor 702 into Row Application on Voiceprint Recognition can receive subsequent voice control, specifically, user if being identified as the vocal print of certification user accordingly Collected voice is sent to processor 702 and identified according to recognition result generation pair by capable of emitting voice, microphone 709 The control instruction answered, such as " booting ", " shutdown ", " promoting display brightness ", " reducing display brightness ", the subsequent basis of processor 702 The control instruction of the generation executes corresponding control processing.
In the present embodiment, processor 702 is used for:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Further, after the reception audio data acquisition signal, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, before the reception audio data acquisition signal, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
Further, if export the audio data to be played by the osteoacusis mode, audio number is received According to signal is obtained, then starts spare osteoacusis mode and obtain third audio data.
It is further, described that denoising is carried out to first audio data, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, it is described obtain target audio data after, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
Further, the reception audio data obtains signal, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
The acquisition device and intelligent wearable device of the audio data of the intelligent wearable device provided in above-described embodiment can be held The acquisition methods of row audio data provided by any embodiment of the invention have and execute the corresponding functional module of this method and have Beneficial effect.The not technical detail of detailed description in the above-described embodiments, reference can be made to audio provided by any embodiment of the invention The acquisition methods of data.
Fig. 9 is the structural schematic diagram of another terminal device provided by the embodiments of the present application.As shown in figure 9, the terminal can To include: shell (not shown), memory 901, central processing unit (Central Processing Unit, CPU) 902 (also known as processor, hereinafter referred to as CPU), the computer program that is stored on memory 901 and can be run on processor 902, Circuit board (not shown) and power circuit (not shown).The circuit board is placed in the space that the shell surrounds Portion;The CPU902 and the memory 901 are arranged on the circuit board;The power circuit, for for the terminal Each circuit or device power supply;The memory 901, for storing executable program code;The CPU902 is by reading institute The executable program code that stores in memory 901 is stated to run program corresponding with the executable program code.
The terminal further include: Peripheral Interface 903, RF (Radio Frequency, radio frequency) circuit 905, voicefrequency circuit 906, loudspeaker 911, power management chip 908, input/output (I/O) subsystem 909, touch screen 912, other input/controls Equipment 910 and outside port 904, these components are communicated by one or more communication bus or signal wire 907.
It should be understood that graphic terminal 900 is only an example of terminal, and terminal device 900 can be with With than shown in the drawings more or less component, two or more components can be combined, or can have Different component configurations.Various parts shown in the drawings can include one or more signal processings and/or dedicated integrated It is realized in the combination of hardware, software or hardware and software including circuit.
Just provided in this embodiment below to be described in detail for a kind of terminal device, the terminal device is with intelligent hand For machine.
Memory 901, the memory 901 can be accessed by CPU902, Peripheral Interface 903 etc., and the memory 901 can It can also include nonvolatile memory to include high-speed random access memory, such as one or more disk memory, Flush memory device or other volatile solid-state parts.
The peripheral hardware that outputs and inputs of equipment can be connected to CPU902 and deposited by Peripheral Interface 903, the Peripheral Interface 903 Reservoir 901.
I/O subsystem 909, the I/O subsystem 909 can be by the input/output peripherals in equipment, such as touch screen 912 With other input/control devicess 910, it is connected to Peripheral Interface 903.I/O subsystem 909 may include 9091 He of display controller For controlling one or more input controllers 9092 of other input/control devicess 910.Wherein, one or more input controls Device 9092 processed receives electric signal from other input/control devicess 910 or sends electric signal to other input/control devicess 910, Other input/control devicess 910 may include physical button (push button, rocker buttons etc.), dial, slide switch, behaviour Vertical pole clicks idler wheel.It is worth noting that input controller 9092 can with it is following any one connect: keyboard, infrared port, The indicating equipment of USB interface and such as mouse.
Wherein, according to the working principle of touch screen and transmission information medium classification, touch screen 912 can for resistance-type, Capacitor induction type, infrared-type or surface acoustic wave type.Classify according to mounting means, touch screen 912 can be with are as follows: external hanging type, built-in Formula or monoblock type.Classify according to technical principle, touch screen 912 can be with are as follows: vector pressure sensing technology touch screen, resistive technologies touching Touch screen, capacitance technology touch screen, infrared technology touch screen or surface acoustic wave technique touch screen.
Touch screen 912, the touch screen 912 are the input interface and output interface between user terminal and user, can It is shown to user depending on output, visual output may include figure, text, icon, video etc..Optionally, touch screen 912 is by user The electric signal (electric signal of such as contact surface) triggered on touch screen curtain, is sent to processor 902.
Display controller 9091 in I/O subsystem 909 receives electric signal from touch screen 912 or sends out to touch screen 912 Electric signals.Touch screen 912 detects the contact on touch screen, and the contact that display controller 9091 will test is converted to and is shown The interaction of user interface object on touch screen 912, i.e. realization human-computer interaction, the user interface being shown on touch screen 912 Object can be the icon of running game, the icon for being networked to corresponding network etc..It is worth noting that equipment can also include light Mouse, light mouse are the extensions for the touch sensitive surface for not showing the touch sensitive surface visually exported, or formed by touch screen.
RF circuit 905 is mainly used for establishing the communication of intelligent sound box Yu wireless network (i.e. network side), realizes intelligent sound box Data receiver and transmission with wireless network.Such as transmitting-receiving short message, Email etc..
Voicefrequency circuit 906 is mainly used for receiving audio data from Peripheral Interface 903, which is converted to telecommunications Number, and the electric signal is sent to loudspeaker 911.
Loudspeaker 911 is reduced to for intelligent sound box to be passed through RF circuit 905 from the received voice signal of wireless network Sound simultaneously plays the sound to user.
Power management chip 908, the hardware for being connected by CPU902, I/O subsystem and Peripheral Interface are powered And power management.
In the present embodiment, central processing unit 902 is used for:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Further, after the reception audio data acquisition signal, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, before the reception audio data acquisition signal, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
Further, if export the audio data to be played by the osteoacusis mode, audio number is received According to signal is obtained, then starts spare osteoacusis mode and obtain third audio data.
It is further, described that denoising is carried out to first audio data, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, it is described obtain target audio data after, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
Further, the reception audio data obtains signal, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
The embodiment of the present application also provides a kind of storage medium comprising terminal device executable instruction, and the terminal device can It executes instruction when being executed by terminal device processor for executing a kind of acquisition methods of audio data, this method comprises:
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
Further, after the reception audio data acquisition signal, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
Further, before the reception audio data acquisition signal, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
Further, if export the audio data to be played by the osteoacusis mode, audio number is received According to signal is obtained, then starts spare osteoacusis mode and obtain third audio data.
It is further, described that denoising is carried out to first audio data, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
Further, it is described obtain target audio data after, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
Further, the reception audio data obtains signal, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
The computer storage medium of the embodiment of the present application, can be using any of one or more computer-readable media Combination.Computer-readable medium can be computer-readable signal media or computer readable storage medium.It is computer-readable Storage medium for example may be-but not limited to-the system of electricity, magnetic, optical, electromagnetic, infrared ray or semiconductor, device or Device, or any above combination.The more specific example (non exhaustive list) of computer readable storage medium includes: tool There are electrical connection, the portable computer diskette, hard disk, random access memory (RAM), read-only memory of one or more conducting wires (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, portable compact disc read-only memory (CD- ROM), light storage device, magnetic memory device or above-mentioned any appropriate combination.In this document, computer-readable storage Medium can be any tangible medium for including or store program, which can be commanded execution system, device or device Using or it is in connection.
Computer-readable signal media may include in a base band or as carrier wave a part propagate data-signal, Wherein carry computer-readable program code.The data-signal of this propagation can take various forms, including but unlimited In electromagnetic signal, optical signal or above-mentioned any appropriate combination.Computer-readable signal media can also be that computer can Any computer-readable medium other than storage medium is read, which can send, propagates or transmit and be used for By the use of instruction execution system, device or device or program in connection.
The program code for including on computer-readable medium can transmit with any suitable medium, including --- but it is unlimited In wireless, electric wire, optical cable, RF etc. or above-mentioned any appropriate combination.
Can with one or more programming languages or combinations thereof come write for execute the application operation computer Program code, programming language include object oriented program language-such as Java, Smalltalk, C++, are also wrapped Include conventional procedural programming language-such as " C " language or similar programming language.Program code can be complete Ground executes on the user computer, partly executes on the user computer, executing as an independent software package, partially existing Part executes on the remote computer or executes on a remote computer or server completely on subscriber computer.It is being related to In the situation of remote computer, remote computer can pass through the network of any kind --- including local area network (LAN) or wide area Net (WAN)-be connected to subscriber computer, or, it may be connected to outer computer (such as utilize ISP To be connected by internet).
Certainly, a kind of storage medium comprising computer executable instructions, computer provided by the embodiment of the present application Operation is recommended in the application that executable instruction is not limited to the described above, and application provided by the application any embodiment can also be performed Relevant operation in recommended method.
Note that above are only the preferred embodiment and institute's application technology principle of the application.It will be appreciated by those skilled in the art that The application is not limited to specific embodiment described here, be able to carry out for a person skilled in the art it is various it is apparent variation, The protection scope readjusted and substituted without departing from the application.Therefore, although being carried out by above embodiments to the application It is described in further detail, but the application is not limited only to above embodiments, in the case where not departing from the application design, also It may include more other equivalent embodiments, and scope of the present application is determined by the scope of the appended claims.

Claims (10)

1. a kind of acquisition methods of audio data characterized by comprising
It receives audio data and obtains signal;
Signal is obtained according to the audio data, the first audio data is obtained by osteoacusis mode;
Denoising is carried out to first audio data, obtains target audio data.
2. the acquisition methods of audio data according to claim 1, which is characterized in that the reception audio data obtains letter After number, further includes:
Signal acquisition environmental audio data are obtained according to the audio data;
Correspondingly, described carry out denoising to the audio data, target audio data are obtained, comprising:
Second audio data is generated according to the environmental audio data and first audio data;
Denoising is carried out to the second audio data, obtains target audio data.
3. the acquisition methods of audio data according to claim 1 or 2, which is characterized in that the reception audio data obtains Before the number of winning the confidence, further includes:
Audio data to be played is obtained, the audio data to be played is exported by osteoacusis mode.
4. the acquisition methods of audio data according to claim 3, which is characterized in that
If export the audio data to be played by the osteoacusis mode, receives audio data and obtain signal, then Start spare osteoacusis mode and obtains third audio data.
5. the acquisition methods of audio data according to claim 1 or 2, which is characterized in that described to first audio Data carry out denoising, comprising:
Extract the corresponding audio frequency characteristics of first audio data;
According to the audio frequency characteristics, judge in the audio data with the presence or absence of invalid data;
If there is invalid data, then the invalid data in first audio data is filtered out.
6. the acquisition methods of audio data according to claim 1 or 2, which is characterized in that described to obtain target sound frequency According to later, further includes:
Intention analysis is carried out to the target audio data;
Respond the corresponding control instruction of the intention analysis result.
7. the acquisition methods of audio data according to claim 1, which is characterized in that the reception audio data obtains letter Number, comprising:
When current location information is matched with preset position information, receives audio data and obtain signal.
8. a kind of acquisition device of audio data characterized by comprising
Signal receiving module obtains signal for receiving audio data;
First audio obtains module, and the audio data for being received according to the signal receiving module obtains signal, leads to It crosses osteoacusis mode and obtains the first audio data;
Denoising module carries out at denoising for obtaining first audio data that module obtains to first audio Reason, obtains target audio data.
9. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the program is held by processor The acquisition methods of the audio data as described in any in claim 1-7 are realized when row.
10. a kind of terminal, including memory, processor and storage are on a memory and can be in the computer journey of processor operation Sequence, which is characterized in that the processor realizes audio number as claimed in claim 1 when executing the computer program According to acquisition methods.
CN201811000969.3A 2018-08-30 2018-08-30 Acquisition methods, device, storage medium and the terminal of audio data Pending CN109240639A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811000969.3A CN109240639A (en) 2018-08-30 2018-08-30 Acquisition methods, device, storage medium and the terminal of audio data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811000969.3A CN109240639A (en) 2018-08-30 2018-08-30 Acquisition methods, device, storage medium and the terminal of audio data

Publications (1)

Publication Number Publication Date
CN109240639A true CN109240639A (en) 2019-01-18

Family

ID=65069811

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811000969.3A Pending CN109240639A (en) 2018-08-30 2018-08-30 Acquisition methods, device, storage medium and the terminal of audio data

Country Status (1)

Country Link
CN (1) CN109240639A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111050248A (en) * 2020-01-14 2020-04-21 Oppo广东移动通信有限公司 Wireless earphone and control method thereof
CN111491176A (en) * 2020-04-27 2020-08-04 百度在线网络技术(北京)有限公司 Video processing method, device, equipment and storage medium
CN111859008A (en) * 2019-04-29 2020-10-30 深圳市冠旭电子股份有限公司 Music recommending method and terminal
CN112581970A (en) * 2019-09-12 2021-03-30 深圳市韶音科技有限公司 System and method for audio signal generation
CN113053371A (en) * 2019-12-27 2021-06-29 阿里巴巴集团控股有限公司 Voice control system and method, voice suite, bone conduction and voice processing device
WO2022141244A1 (en) * 2020-12-30 2022-07-07 雷铭科技有限公司 Audio file processing method and apparatus, and bone conduction device
US11902759B2 (en) 2019-09-12 2024-02-13 Shenzhen Shokz Co., Ltd. Systems and methods for audio signal generation

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102084668A (en) * 2008-05-22 2011-06-01 伯恩同通信有限公司 A method and a system for processing signals
US20140337036A1 (en) * 2013-05-09 2014-11-13 Dsp Group Ltd. Low power activation of a voice activated device
US9324313B1 (en) * 2013-10-23 2016-04-26 Google Inc. Methods and systems for implementing bone conduction-based noise cancellation for air-conducted sound
US9578423B2 (en) * 2012-12-11 2017-02-21 Beijing Lenovo Software Ltd. Electronic device and sound capturing method
CN106686494A (en) * 2016-12-27 2017-05-17 广东小天才科技有限公司 Voice input control method of wearable device and wearable device
CN106847275A (en) * 2016-12-27 2017-06-13 广东小天才科技有限公司 Method for controlling wearable device and wearable device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102084668A (en) * 2008-05-22 2011-06-01 伯恩同通信有限公司 A method and a system for processing signals
US9578423B2 (en) * 2012-12-11 2017-02-21 Beijing Lenovo Software Ltd. Electronic device and sound capturing method
US20140337036A1 (en) * 2013-05-09 2014-11-13 Dsp Group Ltd. Low power activation of a voice activated device
US9324313B1 (en) * 2013-10-23 2016-04-26 Google Inc. Methods and systems for implementing bone conduction-based noise cancellation for air-conducted sound
CN106686494A (en) * 2016-12-27 2017-05-17 广东小天才科技有限公司 Voice input control method of wearable device and wearable device
CN106847275A (en) * 2016-12-27 2017-06-13 广东小天才科技有限公司 Method for controlling wearable device and wearable device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
何国民主编: "《现代大学计算机基础》", 30 September 2015, 西安电子科技大学出版社 *
迈克尔•拉毕格等编著: "《导演创作完全手册 第5版 插图修订版》", 31 July 2016 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111859008A (en) * 2019-04-29 2020-10-30 深圳市冠旭电子股份有限公司 Music recommending method and terminal
CN111859008B (en) * 2019-04-29 2023-11-10 深圳市冠旭电子股份有限公司 Music recommending method and terminal
CN112581970A (en) * 2019-09-12 2021-03-30 深圳市韶音科技有限公司 System and method for audio signal generation
US11902759B2 (en) 2019-09-12 2024-02-13 Shenzhen Shokz Co., Ltd. Systems and methods for audio signal generation
CN113053371A (en) * 2019-12-27 2021-06-29 阿里巴巴集团控股有限公司 Voice control system and method, voice suite, bone conduction and voice processing device
CN111050248A (en) * 2020-01-14 2020-04-21 Oppo广东移动通信有限公司 Wireless earphone and control method thereof
CN111050248B (en) * 2020-01-14 2021-10-01 Oppo广东移动通信有限公司 Wireless earphone and control method thereof
CN111491176A (en) * 2020-04-27 2020-08-04 百度在线网络技术(北京)有限公司 Video processing method, device, equipment and storage medium
WO2022141244A1 (en) * 2020-12-30 2022-07-07 雷铭科技有限公司 Audio file processing method and apparatus, and bone conduction device

Similar Documents

Publication Publication Date Title
CN109240639A (en) Acquisition methods, device, storage medium and the terminal of audio data
CN109120790B (en) Call control method and device, storage medium and wearable device
CN110785735B (en) Apparatus and method for voice command scenario
CN106164823B (en) The non-vision feedback that vision changes in staring tracking and equipment
CN110874129A (en) Display system
CN110326300B (en) Information processing apparatus, information processing method, and computer-readable storage medium
CN109259724B (en) Eye monitoring method and device, storage medium and wearable device
US20230045237A1 (en) Wearable apparatus for active substitution
CN109036410A (en) Audio recognition method, device, storage medium and terminal
CN109254659A (en) Control method, device, storage medium and the wearable device of wearable device
CN109040462A (en) Stroke reminding method, apparatus, storage medium and wearable device
CN109238306A (en) Step counting data verification method, device, storage medium and terminal based on wearable device
CN108761795A (en) A kind of Wearable
CN108874130B (en) Play control method and related product
CN109255064A (en) Information search method, device, intelligent glasses and storage medium
US20180054688A1 (en) Personal Audio Lifestyle Analytics and Behavior Modification Feedback
CN108683790B (en) Voice processing method and related product
CN109360549A (en) A kind of data processing method, device and the device for data processing
CN109061903A (en) Data display method, device, intelligent glasses and storage medium
CN109257490B (en) Audio processing method and device, wearable device and storage medium
CN110097875A (en) Interactive voice based on microphone signal wakes up electronic equipment, method and medium
CN110223711A (en) Interactive voice based on microphone signal wakes up electronic equipment, method and medium
CN109241900A (en) Control method, device, storage medium and the wearable device of wearable device
CN110111776A (en) Interactive voice based on microphone signal wakes up electronic equipment, method and medium
CN106685459A (en) Wearable device operation control method and wearable device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190118