[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN106599110A - Artificial intelligence-based voice search method and device - Google Patents

Artificial intelligence-based voice search method and device Download PDF

Info

Publication number
CN106599110A
CN106599110A CN201611079405.4A CN201611079405A CN106599110A CN 106599110 A CN106599110 A CN 106599110A CN 201611079405 A CN201611079405 A CN 201611079405A CN 106599110 A CN106599110 A CN 106599110A
Authority
CN
China
Prior art keywords
voice
child
search results
artificial intelligence
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201611079405.4A
Other languages
Chinese (zh)
Inventor
李超
李先刚
孙珏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN201611079405.4A priority Critical patent/CN106599110A/en
Publication of CN106599110A publication Critical patent/CN106599110A/en
Priority to US15/823,663 priority patent/US10157619B2/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/26Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3343Query execution using phonetics
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3344Query execution using natural language analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/335Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/432Query formulation
    • G06F16/433Query formulation using audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/435Filtering based on additional data, e.g. user or group profiles
    • G06F16/436Filtering based on additional data, e.g. user or group profiles using biological or physiological data of a human being, e.g. blood pressure, facial expression, gestures
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/54Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for retrieval
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/93Discriminating between voiced and unvoiced parts of speech signals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/78Detection of presence or absence of voice signals
    • G10L2025/783Detection of presence or absence of voice signals based on threshold decision
    • G10L2025/786Adaptive threshold
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • G10L25/30Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Theoretical Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Physiology (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

Provided are an artificial intelligence-based voice search method and device. The method comprises: identifying a voice input by a user to judge whether the voice is a child's voice; if a judgment result is yes, filtering search results obtained according to the voice; and feeding the filtered search results back to the user. In the present invention, a voice input by a user is identified to determine whether the user is a child user, and after the user is determined as the child user, search results are filtered, and search results about the child user can be differentially set, so sensitive contents which are included in the search results and are not suitable to be read by a child can be eliminated, and contents pushed to the child user contribute to the physical and psychological health of the child.

Description

Voice search method and device based on artificial intelligence
Technical field
The present invention relates to technical field of information processing, more particularly to a kind of voice search method and dress based on artificial intelligence Put.
Background technology
Artificial intelligence (Artificial Intelligence), english abbreviation is AI.It is study, be developed for simulation, Extend and extend a new science of technology of theory, method, technology and the application system of the intelligence of people.Artificial intelligence is to calculate One branch of machine science, it attempts the essence for understanding intelligence, and produce it is a kind of it is new can be in the way of human intelligence be similar The intelligent machine made a response, the research in the field include robot, speech recognition, image recognition, natural language processing and specially Family's system etc..Wherein, the most important aspect of artificial intelligence is exactly speech recognition technology.
With the continuous development of speech recognition technology, the range of application of speech recognition technology is also more and more extensive.With language The lifting of sound recognition accuracy, increasing user are got used to using phonetic search mode, so as to avoid carrying out input through keyboard, Simplify search procedure.At present, developing rapidly with intelligent terminal's technology, carries out the youngster of phonetic search using intelligent terminal It is more and more, it should be noted that child is particularly preschooler and is likeed best by way of voice when using intelligent terminal Interact.
During phonetic search is carried out, by speech recognition technology, the voice that child is input into is converted into into text, from And the text for being converted is scanned for as search word.But the Search Results obtained based on search word usually contain it is irregular not Neat information, as child user is mostly scanned for using the intelligent terminal of the head of a family, and cannot determine that whether the user is Virgin user, leads to not carry out differentiation setting for the Search Results of child user colony.
The content of the invention
It is contemplated that at least solving one of technical problem in correlation technique to a certain extent.
For this purpose, first purpose of the present invention is to propose a kind of voice search method based on artificial intelligence, to realize It is whether the identification of child user to user, for solving in prior art as user cannot be determined for child, leads to not For its differentiation, the problem of Search Results is set.
Second object of the present invention is to propose a kind of voice searching device based on artificial intelligence.
Third object of the present invention is to propose another kind of voice searching device based on artificial intelligence.
Fourth object of the present invention is to propose a kind of non-transitorycomputer readable storage medium.
5th purpose of the present invention is to propose a kind of computer program.
It is that, up to above-mentioned purpose, first aspect present invention embodiment proposes a kind of phonetic search side based on artificial intelligence Method, including:
The voice being input into user is identified, and judges whether the voice is child's voice;
If it is judged that being yes, the Search Results to being obtained according to the voice are screened;
Search Results after screening are fed back to into the user.
The voice search method based on artificial intelligence of the embodiment of the present invention, is known by the voice being input into user Not, whether to determine the user as child user, after the user is determined for child user, then Search Results are sieved Choosing, realizing the Search Results to child user carries out differentiation setting, does not meet child such that it is able to weed out Search Results The sensitive content of reading so that the content pushed to child user has the physical and mental health using child.
It is that, up to above-mentioned purpose, second aspect present invention embodiment proposes a kind of phonetic search based on artificial intelligence and fills Put, including:
Judge module, the voice for being input into user are identified, and judge whether the voice is child's voice;
Screening module, for being yes in the judged result of the judge module, to the search knot obtained according to the voice Fruit is screened;
Feedback module, for the Search Results after screening are fed back to the user.
The voice searching device based on artificial intelligence of the embodiment of the present invention, is known by the voice being input into user Not, whether to determine the user as child user, after the user is determined for child user, then Search Results are sieved Choosing, realizing the Search Results to child user carries out differentiation setting, does not meet child such that it is able to weed out Search Results The sensitive content of reading so that the content pushed to child user has the physical and mental health using child.
It is that, up to above-mentioned purpose, third aspect present invention embodiment proposes another kind of phonetic search based on artificial intelligence and fills Put, including:Processor;For storing the memorizer of the processor executable;Wherein, the processor is configured to: The voice being input into user is identified, and judges whether the voice is child's voice;If it is judged that be yes, to according to described The Search Results that voice is obtained are screened;Search Results after screening are fed back to into the user.
To achieve these goals, fourth aspect present invention embodiment proposes a kind of non-transitory computer-readable storage Medium, when the instruction in the storage medium is performed by the processor of server end so that server end is able to carry out one The voice search method based on artificial intelligence is planted, methods described includes:The voice being input into user is identified, and judges described Whether voice is child's voice;If it is judged that being yes, the Search Results to being obtained according to the voice are screened;After screening Search Results feed back to the user.
To achieve these goals, fifth aspect present invention embodiment proposes a kind of computer program, when described When instruction processing unit in computer program is performed, a kind of voice search method based on artificial intelligence, the side are performed Method includes:The voice being input into user is identified, and judges whether the voice is child's voice;If it is judged that be it is yes, it is right Screened according to the Search Results that the voice is obtained;Search Results after screening are fed back to into the user.
The additional aspect of the present invention and advantage will be set forth in part in the description, and partly will become from the following description Obtain substantially, or recognized by the practice of the present invention.
Description of the drawings
Of the invention above-mentioned and/or additional aspect and advantage will become from the following description of the accompanying drawings of embodiments It is substantially and easy to understand, wherein:
A kind of schematic flow sheet of voice search method based on artificial intelligence that Fig. 1 is provided by the embodiment of the present invention;
Flow process of the another kind that Fig. 2 is provided by the embodiment of the present invention based on the voice search method of artificial intelligence is illustrated Figure;
A kind of training flow process of grader that Fig. 3 is provided by the embodiment of the present invention;
A kind of structural representation of voice searching device based on artificial intelligence that Fig. 4 is provided by the embodiment of the present invention;
Fig. 5 is a kind of structural representation of screening module provided in an embodiment of the present invention;
Fig. 6 is a kind of structural representation of judge module provided in an embodiment of the present invention.
Specific embodiment
Embodiments of the invention are described below in detail, the example of the embodiment is shown in the drawings, wherein from start to finish Same or similar label represents same or similar element or the element with same or like function.Below with reference to attached The embodiment of figure description is exemplary, it is intended to for explaining the present invention, and be not considered as limiting the invention.
Below with reference to the accompanying drawings the voice search method based on artificial intelligence and device of the embodiment of the present invention are described.
A kind of schematic flow sheet of voice search method based on artificial intelligence that Fig. 1 is provided by the embodiment of the present invention. Should be comprised the following steps based on the voice search method of artificial intelligence:
S101, the voice being input into user are identified, and judge whether voice is child's voice.
In the present embodiment, speech recognition modeling is provided with advance in a search engine, can be right by speech recognition modeling The voice is input into by user is identified, to determine the classification of the voice.The speech recognition modeling can be machine learning model, After the training to great amount of samples data and study, the machine learning model can be classified to the classification of voice.
If it is judged that being yes, then S102 is performed, if it is judged that being no, then terminate flow process.
S102, the Search Results to being obtained according to voice are screened.
Specifically, voice is carried out changing the content of text that can get voice, then using content of text as search Key word, scanned for based on the key word of the search, obtain Search Results corresponding with the content of text of the voice.
Due to the information on the Internet it is more various, according to may include in the result that the content of text of language is searched Sensitive content, or the bad network information.Due to having identified that in S101 voice is child's voice, illustrate to be input into the user of voice For child user, if Search Results are not carried out screening that Search Results are pushed to the child user directly, this is may result in Child user receives many unsound network consultings, and these bad network consultings or information can be to the physical and mental healths of child Adversely affect.
In the present embodiment, in order to preferably provide service or network environment well to child user, according to language Content of text get Search Results after, need to screen Search Results, be not suitable for weeding out in Search Results The sensitive content of children's reading.Specifically, Search Results can be analyzed, in judging the Search Results, whether includes discomfort Close children's reading sensitive content, if it is determined that go out Search Results go out protection have the sensitive content for being not suitable for children's reading, so that it may To go out to weed out these sensitive contents from the Search Results.
S103, the Search Results after screening are fed back to into user.
As the sensitive content for being not suitable for children's reading is eliminated from initial Search Results, complete to Search Results Screening, the Search Results after screening than being well suited to children's reading, then feed back to use by the content relative healths in Search Results Family.
The voice search method based on artificial intelligence that the present embodiment is provided, is known by the voice being input into user Not, whether to determine the user as child user, after the user is determined for child user, then Search Results are sieved Choosing, realizing the Search Results to child user carries out differentiation setting, does not meet child such that it is able to weed out Search Results The sensitive content of reading so that the content pushed to child user has the physical and mental health using child.
A kind of schematic flow sheet of voice search method based on artificial intelligence that Fig. 2 is provided by the embodiment of the present invention. Should be comprised the following steps based on the voice search method of artificial intelligence:
S201, voice is carried out voice activity detection remove voice in it is quiet, obtain tested speech.
As user is during speaking, may there is pause, in order to the searching resource for reducing occupancy can be gathered To the voice that is input into of user carry out voice activity detection (Voice Activity Detection, abbreviation VAD), with from sound The prolonged quiet phase is recognized and is eliminated in sound signal stream, to reach the purpose for saving bandwidth resources, and advantageously reduces use The time delay end to end that family is felt.
Specifically, framing can be carried out according to default first step length to voice, for example, can be right with 10ms as a frame Every frame of voice carries out voice activity detection, quiet in voice to remove, will remove it is quiet after voice as tested speech.
S202, the acoustic featuress for extracting tested speech.
Specifically, framing can be carried out according to default second step length to tested speech, for example, the second step-length can be with 25ms is a frame, then according to default 3rd step-length carries out extraction acoustic featuress per frame to tested speech, for example, the 3rd step-length Can be 10ms.In the present embodiment, acoustic featuress can be wave filter group (Filter bank40, abbreviation Fbank40) feature or Person's mel-frequency cepstrum coefficient (Mel-Frequency Cepstral Coefficients, abbreviation MFCC) feature.
S203, the acoustic featuress of tested speech are input in object classifiers it are identified, judges that whether voice be Child's voice.
Further, will remove it is quiet after voice as tested speech, in order to identify whether tested speech is child's voice, Need to extract the acoustic featuress of the tested speech, then the acoustic featuress of the tested speech are input to the mesh for having trained In mark grader, whether object classifiers can be identified to acoustic featuress to tested speech, to determine the voice as child's voice.
Specifically, after tested speech being input to object classifiers per frame corresponding acoustic featuress, object classifiers can To give a mark to each frame, then the comprehensive score per frame carries out final judgement, in the present embodiment, can obtaining every frame Point sum up it is average, according to plus and it is average after meansigma methodss as basis for estimation, when meansigma methodss are more than or equal to default After threshold value, then may determine that the voice is child's voice, if being less than the threshold value, may determine that the non-child's voice of voice.
In the present embodiment, S204 is then performed when judging that voice is child's voice, when it is non-child's voice to judge voice, terminate Flow process.
In the present embodiment, before voice is tested, need to be trained grader, to get target classification Device.Fig. 3 is the classifier training flow process that provides in the embodiment of the present invention.As shown in figure 3, the classifier training includes following step Suddenly:
S301, obtain sample voice for being trained to default grader.
In the present embodiment, the sound of child can be sampled as sample voice, the sample collected using these Voice is trained to default grader.Preferably, the voice for being labeled as child's voice in history by manual sort can be gathered As sample voice.
S302, sample voice is carried out voice activity detection remove training data in it is quiet, obtain train voice.
As user is during speaking, may there is pause, in order to reduce the searching resource of occupancy, get Sample voice is carried out after sample voice voice activity detection remove training data in it is quiet, obtain train voice.
Specifically, framing can be carried out according to default first step length to sample voice, for example, can be with 10ms as one Frame, carries out voice activity detection to every frame of sample voice, quiet in sample voice to remove, will remove it is quiet after sample Voice is used as training voice.
S303, the acoustic featuress for extracting training voice.
Specifically, can be to training voice to carry out framing according to default second step length, for example, can be with 25ms as one Frame, then according to default 3rd step-length extracts acoustic featuress to the every frame for training voice, wherein, acoustic featuress can be filtering Device group (Filter bank40, abbreviation Fbank40) feature or mel-frequency cepstrum coefficient (Mel-Frequency Cepstral Coefficients, abbreviation MFCC) feature.
S304, the acoustic featuress of training voice are input in grader and are trained, obtain object classifiers.
In the present embodiment, time recurrent neural network (Long Short-Time Memory can be based on Projection, abbreviation LSTMP), deep neural network (Deep Neural Networks, abbreviation DNN) or circulation nerve Network (Recurrent Neural Networks, abbreviation RNN) is modeled to grader.Calculation based on machine learning herein Method builds grader and is not defined.
After the acoustic featuress for getting training voice, acoustic featuress are input in grader and are trained, work as classification After the result stable convergence of device output, the training of the grader completes to obtain object classifiers.
S204, convert speech into content of text.
S205, scanned in the data base for setting up for child in advance according to content of text.
In order to ensure to provide preferably service or network environment for child, in the present embodiment, can build for child in advance Stand some data bases, such as test item bank, nursery rhymes animation library, developmental game storehouse etc.
After content of text is converted speech into, can be entered in the data base for setting up for child in advance according to content of text Line search, as these data bases are set up for children population, enables to the search that Search Results more meet child In demand, and data base, information has been carried out screening relative healths.
If the content of correlation is not searched in the data base for setting up for child in advance, enter in other data bases Line search, that is, perform S206;If the content of correlation is searched in the data base for setting up for child in advance, S207 is performed.
If S206, the content that correlation is not searched in the data base for setting up for child in advance, exist according to content of text Scan for obtaining Search Results in other data bases.
When the content of text according to voice does not search the content of correlation in the data base pre-build for child, it is Meet the demand of search user, can be scanned in other data bases according to content of text, think user search to phase The content of pass.
S207, Search Results are carried out screening removing be not suitable for the sensitive content of children's reading.
Due to the information on the Internet it is more various, according to may include in the result that the content of text of language is searched Sensitive content, or the bad network information.Due to having identified that in S203 voice is child's voice, illustrate to be input into the user of voice For child user, if Search Results are not carried out screening that Search Results are pushed to the child user directly, this is may result in Child user receives many unsound network consultings, and these bad network consultings or information can be to the physical and mental healths of child Adversely affect.
In the present embodiment, in order to preferably provide service or network environment well to child user, to Search Results Screened, to weed out the sensitive content for being not suitable for children's reading in Search Results.Specifically, Search Results can be carried out Whether analysis, include the sensitive content for being not suitable for children's reading in judging the Search Results, if it is determined that goes out Search Results and goes out to protect Shield has the sensitive content for being not suitable for children's reading, it is possible to go out to weed out these sensitive contents from the Search Results.
In the present embodiment, in order to preferably provide the network information, the knot searched in the data base for pre-building to child Fruit can also be screened, and further increasing the safety of information.
S208, the Search Results after screening are fed back to into user.
As the sensitive content for being not suitable for children's reading is eliminated from initial Search Results, complete to Search Results Screening, the Search Results after screening than being well suited to children's reading, then feed back to use by the content relative healths in Search Results Family.
The voice search method based on artificial intelligence that the present embodiment is provided, is known by the voice being input into user Not, whether to determine the user as child user, after the user is determined for child user, then Search Results are sieved Choosing, realizing the Search Results to child user carries out differentiation setting, does not meet child such that it is able to weed out Search Results The sensitive content of reading so that the content pushed to child user has the physical and mental health using child.
A kind of structural representation of voice searching device based on artificial intelligence that Fig. 4 is provided by the embodiment of the present invention. Should be included based on the voice searching device of artificial intelligence:Judge module 11, screening module 12 and feedback module 13.
Wherein, judge module 11, the voice for being input into user are identified, and judge whether the voice is virgin Sound.
Screening module 12, for being yes, to being obtained according to the voice search in the judged result of the judge module As a result screened.
Feedback module 13, for the Search Results after screening are fed back to the user.
A kind of structural representation of screening module that Fig. 5 is provided for the present embodiment.As shown in figure 5, the screening module 12 is wrapped Include:Converting unit 121, search unit 122 and screening unit 123.
Converting unit 121, for the voice is converted into content of text.
Search unit 122, for scanning for obtaining the Search Results according to the content of text.
Screening unit 123, removes the sensitive content for being not suitable for children's reading for carrying out screening to the Search Results.
Further, search unit 122, specifically for:
Scanned in the data base for setting up for child in advance according to the content of text.
If not searching the content of correlation in the database, according to the content of text in other data bases Scan for, to obtain the Search Results.
A kind of structural representation of judge module that Fig. 6 is provided for the present embodiment.As shown in fig. 6, the judge module 11 is wrapped Include:Acquiring unit 111, quiet removal unit 112, feature extraction unit 113, training unit 114 and identification judging unit 115.
Acquiring unit 111, for obtaining the sample voice for being trained to default grader.
Quiet removal unit 112, removes in the training data for voice activity detection is carried out to the sample voice It is quiet, obtain train voice.
Feature extraction unit 113, for extracting the acoustic featuress of the training voice.
Training unit 114, is trained for the acoustic featuress of the training voice are input in the grader, obtains To the object classifiers.
Further, quiet removal unit 112, specifically for:
According to default first step-length to the sample voice framing, and the every frame to the sample voice carries out voice work Move quiet during detecting is removed per frame, obtain the training voice.
Feature extraction unit 113, specifically for:
Framing is carried out to the training voice according to default second step-length.
According to default 3rd step-length to framing after the training voice carry out acoustic featuress extraction.
Further, quiet removal unit 112, is additionally operable to carry out the voice voice activity detection removal voice In it is quiet, obtain tested speech.
Feature extraction unit 113, is additionally operable to the acoustic featuress for extracting the tested speech.
Further, judge module 11, also include:
Identification judging unit 115, for being input in the object classifiers by the acoustic featuress of the tested speech Row identification, to judge the voice whether as child's voice.
Further, quiet removal unit 112, also particularly useful for:
According to default first step-length to voice framing, and every frame is carried out quiet during voice activity detection is removed per frame Sound, obtains the tested speech.
Feature extraction unit 113, also particularly useful for:
Framing is carried out to the tested speech according to default second step-length.
According to default 3rd step-length to framing after the tested speech carry out acoustic featuress extraction.
Identification judging unit 115, specifically for:
The tested speech is input in the object classifiers per the corresponding sound characteristic of frame and is given a mark.
Obtain the meansigma methodss of score of the tested speech per frame.
If the meansigma methodss exceed default threshold value, judge the tested speech as child's voice.
The voice searching device based on artificial intelligence that the present embodiment is provided, is known by the voice being input into user Not, whether to determine the user as child user, after the user is determined for child user, then Search Results are sieved Choosing, realizing the Search Results to child user carries out differentiation setting, does not meet child such that it is able to weed out Search Results The sensitive content of reading so that the content pushed to child user has the physical and mental health using child.
In the description of this specification, reference term " one embodiment ", " some embodiments ", " example ", " specifically show Example ", or the description of " some examples " etc. mean specific features with reference to the embodiment or example description, structure, material or spy Point is contained at least one embodiment or example of the present invention.In this manual, to the schematic representation of above-mentioned term not Identical embodiment or example must be directed to.And, the specific features of description, structure, material or feature can be with office Combined in one or more embodiments or example in an appropriate manner.Additionally, in the case of not conflicting, the skill of this area The feature of the different embodiments or example described in this specification and different embodiments or example can be tied by art personnel Close and combine.
Additionally, term " first ", " second " are only used for describing purpose, and it is not intended that indicating or implying relative importance Or the implicit quantity for indicating indicated technical characteristic.Thus, define " first ", the feature of " second " can express or Implicitly include at least one this feature.In describing the invention, " multiple " are meant that at least two, such as two, three It is individual etc., unless otherwise expressly limited specifically.
In flow chart or here any process described otherwise above or method description are construed as, expression includes It is one or more for realizing custom logic function or process the step of the module of code of executable instruction, fragment or portion Point, and the scope of the preferred embodiment of the present invention includes other realization, wherein the suitable of shown or discussion can not be pressed Sequence, including according to involved function by it is basic simultaneously in the way of or in the opposite order, carry out perform function, this should be of the invention Embodiment person of ordinary skill in the field understood.
Expression or here logic described otherwise above and/or step, for example, are considered use in flow charts In the order list of the executable instruction for realizing logic function, may be embodied in any computer-readable medium, for Instruction execution system, device or equipment (as computer based system, the system including processor or other can hold from instruction The system of row system, device or equipment instruction fetch execute instruction) use, or with reference to these instruction execution systems, device or set It is standby and use.For the purpose of this specification, " computer-readable medium " can any can be included, store, communicate, propagate or pass The dress that defeated program is used for instruction execution system, device or equipment or with reference to these instruction execution systems, device or equipment Put.The more specifically example (non-exhaustive list) of computer-readable medium is including following:With the electricity that one or more connect up Connecting portion (electronic installation), portable computer diskette box (magnetic device), random access memory (RAM), read only memory (ROM), erasable edit read-only storage (EPROM or flash memory), fiber device, and portable optic disk is read-only deposits Reservoir (CDROM).In addition, computer-readable medium can even is that the paper that can print described program thereon or other are suitable Medium, because for example by carrying out optical scanning to paper or other media edlin, interpretation can then be entered or if necessary with which His suitable method is processed to electronically obtain described program, is then stored in computer storage.
It should be appreciated that each several part of the present invention can be realized with hardware, software, firmware or combinations thereof.Above-mentioned In embodiment, the software that multiple steps or method can be performed in memory and by suitable instruction execution system with storage Or firmware is realizing.Such as, if realized with hardware with another embodiment, can be with following skill well known in the art Any one of art or their combination are realizing:With for data signal is realized the logic gates of logic function from Scattered logic circuit, the special IC with suitable combinational logic gate circuit, programmable gate array (PGA), scene can be compiled Journey gate array (FPGA) etc..
Those skilled in the art are appreciated that to realize all or part of step that above-described embodiment method is carried Suddenly the hardware that can be by program to instruct correlation is completed, and described program can be stored in a kind of computer-readable storage medium In matter, the program upon execution, including one or a combination set of the step of embodiment of the method.
Additionally, each functional unit in each embodiment of the invention can be integrated in a processing module, it is also possible to It is that unit is individually physically present, it is also possible to which two or more units are integrated in a module.Above-mentioned integrated mould Block both can be realized in the form of hardware, it would however also be possible to employ the form of software function module is realized.The integrated module is such as Fruit using in the form of software function module realize and as independent production marketing or use when, it is also possible to be stored in a computer In read/write memory medium.
Storage medium mentioned above can be read only memory, disk or CD etc..Although having shown that above and retouching Embodiments of the invention are stated, it is to be understood that above-described embodiment is exemplary, it is impossible to be interpreted as the limit to the present invention System, one of ordinary skill in the art can be changed to above-described embodiment, change, replace and become within the scope of the invention Type.

Claims (14)

1. a kind of voice search method based on artificial intelligence, it is characterised in that include:
The voice being input into user is identified, and judges whether the voice is child's voice;
If it is judged that being yes, the Search Results to being obtained according to the voice are screened;
Search Results after screening are fed back to into the user.
2. the voice search method based on artificial intelligence according to claim 1, it is characterised in that described to according to described The Search Results that voice is obtained are screened, including:
The voice is converted into into content of text;
Scan for obtaining the Search Results according to the content of text;
The Search Results are carried out screening and removes the sensitive content for being not suitable for children's reading.
3. the voice search method based on artificial intelligence according to claim 2, it is characterised in that described according to the text This content scans for obtaining the Search Results, including:
Scanned in the data base for setting up for child in advance according to the content of text;
If not searching the content of correlation in the database, carried out in other data bases according to the content of text Search, to obtain the Search Results.
4. the voice search method based on artificial intelligence according to any one of claim 1-3, it is characterised in that described right Before the voice is input into by user is identified, including:
Obtain the sample voice for being trained to default grader;
Voice activity detection is carried out to the sample voice and removes quiet in the training data, obtain training voice;
Extract the acoustic featuress of the training voice;
The acoustic featuress of the training voice are input in the grader and are trained, obtain the object classifiers.
5. the voice search method based on artificial intelligence according to claim 4, it is characterised in that described to the sample Voice carries out voice activity detection and removes quiet in the training data, obtains training voice, including:
According to default first step-length to the sample voice framing, and speech activity is carried out to every frame of the sample voice detect Survey quiet in removing per frame, obtain the training voice;
The acoustic featuress for extracting the training voice, including:
Framing is carried out to the training voice according to default second step-length,
According to default 3rd step-length to framing after the training voice carry out acoustic featuress extraction.
6. the voice search method based on artificial intelligence according to claim 4, it is characterised in that described defeated to user institute The voice for entering is identified, and judges whether the voice is child's voice, including:
Voice activity detection is carried out to the voice and removes quiet in the voice, obtain tested speech;
Extract the acoustic featuress of the tested speech;
The acoustic featuress of the tested speech are input in the object classifiers and are identified, whether to judge the voice For child's voice.
7. the voice search method based on artificial intelligence according to claim 6, it is characterised in that described to the voice Carry out voice activity detection and remove quiet in the voice, obtain tested speech, including:
It is according to default first step-length to voice framing and quiet during voice activity detection removal is carried out to every frame per frame, obtain To the tested speech;
The acoustic featuress for extracting the tested speech, including:
Framing is carried out to the tested speech according to default second step-length,
According to default 3rd step-length to framing after the tested speech carry out acoustic featuress extraction;
The acoustic featuress by the tested speech are identified in being input to the object classifiers, to judge the voice Whether it is child's voice, including:
The tested speech is input in the object classifiers per the corresponding sound characteristic of frame and is given a mark;
Obtain the meansigma methodss of score of the tested speech per frame;
If the meansigma methodss exceed default threshold value, judge the tested speech as child's voice.
8. a kind of voice searching device based on artificial intelligence, it is characterised in that include:
Judge module, the voice for being input into user are identified, and judge whether the voice is child's voice;
Screening module, for being yes in the judged result of the judge module, the Search Results to being obtained according to the voice enter Row screening;
Feedback module, for the Search Results after screening are fed back to the user.
9. the voice searching device based on artificial intelligence according to claim 8, it is characterised in that the screening module, Including:
Converting unit, for the voice is converted into content of text;
Search unit, for scanning for obtaining the Search Results according to the content of text;
Screening unit, removes the sensitive content for being not suitable for children's reading for carrying out screening to the Search Results.
10. the voice searching device based on artificial intelligence according to claim 9, it is characterised in that the search unit, Specifically for:
Scanned in the data base for setting up for child in advance according to the content of text;
If not searching the content of correlation in the database, carried out in other data bases according to the content of text Search, to obtain the Search Results.
11. voice searching devices based on artificial intelligence according to any one of claim 8-10, it is characterised in that described Judge module, including:
Acquiring unit, for obtaining the sample voice for being trained to default grader;
Quiet removal unit, it is quiet in the voice activity detection removal training data for carrying out to the sample voice, Obtain training voice;
Feature extraction unit, for extracting the acoustic featuress of the training voice;
Training unit, is trained for the acoustic featuress of the training voice are input in the grader, obtains described Object classifiers.
12. voice searching devices based on artificial intelligence according to claim 11, it is characterised in that the quiet removal Unit, specifically for:
According to default first step-length to the sample voice framing, and speech activity is carried out to every frame of the sample voice detect Survey quiet in removing per frame, obtain the training voice;
The feature extraction unit, specifically for:
Framing is carried out to the training voice according to default second step-length;
According to default 3rd step-length to framing after the training voice carry out acoustic featuress extraction.
13. voice searching devices based on artificial intelligence according to claim 11, it is characterised in that
The quiet removal unit, is additionally operable to carry out the voice voice activity detection and removes quiet in the voice, obtain To tested speech;
The feature extraction unit, is additionally operable to the acoustic featuress for extracting the tested speech;
The judge module, also includes:
Identification judging unit, is identified for the acoustic featuress of the tested speech are input in the object classifiers, To judge the voice whether as child's voice.
14. voice searching devices based on artificial intelligence according to claim 13, it is characterised in that the quiet removal Unit, also particularly useful for:
It is according to default first step-length to voice framing and quiet during voice activity detection removal is carried out to every frame per frame, obtain To the tested speech;
The feature extraction unit, also particularly useful for:
Framing is carried out to the tested speech according to default second step-length;
According to default 3rd step-length to framing after the tested speech carry out acoustic featuress extraction;
The identification judging unit, specifically for:
The tested speech is input in the object classifiers per the corresponding sound characteristic of frame and is given a mark;
Obtain the meansigma methodss of score of the tested speech per frame;
If the meansigma methodss exceed default threshold value, judge the tested speech as child's voice.
CN201611079405.4A 2016-11-29 2016-11-29 Artificial intelligence-based voice search method and device Pending CN106599110A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201611079405.4A CN106599110A (en) 2016-11-29 2016-11-29 Artificial intelligence-based voice search method and device
US15/823,663 US10157619B2 (en) 2016-11-29 2017-11-28 Method and device for searching according to speech based on artificial intelligence

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611079405.4A CN106599110A (en) 2016-11-29 2016-11-29 Artificial intelligence-based voice search method and device

Publications (1)

Publication Number Publication Date
CN106599110A true CN106599110A (en) 2017-04-26

Family

ID=58595724

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611079405.4A Pending CN106599110A (en) 2016-11-29 2016-11-29 Artificial intelligence-based voice search method and device

Country Status (2)

Country Link
US (1) US10157619B2 (en)
CN (1) CN106599110A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107610706A (en) * 2017-09-13 2018-01-19 百度在线网络技术(北京)有限公司 The processing method and processing unit of phonetic search result
CN108766431A (en) * 2018-06-06 2018-11-06 广东小天才科技有限公司 Automatic awakening method based on voice recognition and electronic equipment
CN109087633A (en) * 2018-08-23 2018-12-25 北京猎户星空科技有限公司 Voice assessment method, device and electronic equipment
CN110660393A (en) * 2019-10-31 2020-01-07 广东美的制冷设备有限公司 Voice interaction method, device, equipment and storage medium
CN112231440A (en) * 2020-10-09 2021-01-15 安徽讯呼信息科技有限公司 Voice search method based on artificial intelligence
CN113486208A (en) * 2021-06-09 2021-10-08 安徽沐峰数据科技有限公司 Voice search equipment based on artificial intelligence and search method thereof

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107221326B (en) * 2017-05-16 2021-05-28 百度在线网络技术(北京)有限公司 Voice awakening method and device based on artificial intelligence and computer equipment
CN111916065B (en) * 2020-08-05 2024-07-02 北京百度网讯科技有限公司 Method and device for processing voice
CN112614514B (en) * 2020-12-15 2024-02-13 中国科学技术大学 Effective voice fragment detection method, related equipment and readable storage medium
CN113514544A (en) * 2020-12-29 2021-10-19 大连理工大学 Mobile robot pavement material identification method based on sound characteristics

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8417530B1 (en) * 2010-08-20 2013-04-09 Google Inc. Accent-influenced search results
CN103151039A (en) * 2013-02-07 2013-06-12 中国科学院自动化研究所 Speaker age identification method based on SVM (Support Vector Machine)
CN104795067A (en) * 2014-01-20 2015-07-22 华为技术有限公司 Voice interaction method and device
CN105095406A (en) * 2015-07-09 2015-11-25 百度在线网络技术(北京)有限公司 Method and apparatus for voice search based on user feature
CN105677787A (en) * 2015-12-29 2016-06-15 广州神马移动信息科技有限公司 Information searching device and information searching method
US20160171109A1 (en) * 2014-12-12 2016-06-16 Ebay Inc. Web content filtering

Family Cites Families (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0127718B1 (en) * 1983-06-07 1987-03-18 International Business Machines Corporation Process for activity detection in a voice transmission system
TW356548B (en) * 1996-07-11 1999-04-21 Sega Entpr Corp Sound identifying device method of sound identification and the game machine using the said device
US6286001B1 (en) * 1999-02-24 2001-09-04 Doodlebug Online, Inc. System and method for authorizing access to data on content servers in a distributed network
US6336117B1 (en) * 1999-04-30 2002-01-01 International Business Machines Corporation Content-indexing search system and method providing search results consistent with content filtering and blocking policies implemented in a blocking engine
AU2001234011A1 (en) * 2000-01-28 2001-08-07 Sagi Cooper Apparatus and method for accessing multimedia content
US20040128282A1 (en) * 2001-03-07 2004-07-01 Paul Kleinberger System and method for computer searching
EP1402409A2 (en) * 2001-06-08 2004-03-31 W.W. Grainger, Inc. System and method for retrieving information from an electronic catalog
US6731239B2 (en) * 2002-01-18 2004-05-04 Ford Motor Company System and method for retrieving information using position coordinates
US7216121B2 (en) * 2002-12-31 2007-05-08 International Business Machines Corporation Search engine facility with automated knowledge retrieval, generation and maintenance
US7672931B2 (en) * 2005-06-30 2010-03-02 Microsoft Corporation Searching for content using voice search queries
US20080183694A1 (en) * 2007-01-31 2008-07-31 Daniel Cane Method and system presenting search results using relationship information
US7882102B2 (en) * 2007-09-10 2011-02-01 Mitac International Corporation Nearest-neighbor geographic search
US8160877B1 (en) * 2009-08-06 2012-04-17 Narus, Inc. Hierarchical real-time speaker recognition for biometric VoIP verification and targeting
US20110302633A1 (en) * 2010-06-07 2011-12-08 Windschauer Robert J Safe Internet Browser
US8484219B2 (en) * 2010-09-21 2013-07-09 Sony Computer Entertainment America Llc Developing a knowledge base associated with a user that facilitates evolution of an intelligent user interface
US8768782B1 (en) * 2011-06-10 2014-07-01 Linkedin Corporation Optimized cloud computing fact checking
KR101971008B1 (en) * 2012-06-29 2019-04-22 삼성전자주식회사 Control method for terminal using context-aware and terminal thereof
US20150287410A1 (en) * 2013-03-15 2015-10-08 Google Inc. Speech and semantic parsing for content selection
US9489411B2 (en) * 2013-07-29 2016-11-08 Sybase, Inc. High performance index creation
US10083237B2 (en) * 2015-08-31 2018-09-25 Google Llc Protecting users from inappropriate sensitive or offensive search results
US20170337610A1 (en) * 2016-05-20 2017-11-23 Monroney Labels, LLC Motor Vehicle Data Retrieval, Processing and Presentation Process

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8417530B1 (en) * 2010-08-20 2013-04-09 Google Inc. Accent-influenced search results
CN103151039A (en) * 2013-02-07 2013-06-12 中国科学院自动化研究所 Speaker age identification method based on SVM (Support Vector Machine)
CN104795067A (en) * 2014-01-20 2015-07-22 华为技术有限公司 Voice interaction method and device
US20160171109A1 (en) * 2014-12-12 2016-06-16 Ebay Inc. Web content filtering
CN105095406A (en) * 2015-07-09 2015-11-25 百度在线网络技术(北京)有限公司 Method and apparatus for voice search based on user feature
CN105677787A (en) * 2015-12-29 2016-06-15 广州神马移动信息科技有限公司 Information searching device and information searching method

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107610706A (en) * 2017-09-13 2018-01-19 百度在线网络技术(北京)有限公司 The processing method and processing unit of phonetic search result
CN108766431A (en) * 2018-06-06 2018-11-06 广东小天才科技有限公司 Automatic awakening method based on voice recognition and electronic equipment
CN109087633A (en) * 2018-08-23 2018-12-25 北京猎户星空科技有限公司 Voice assessment method, device and electronic equipment
CN110660393A (en) * 2019-10-31 2020-01-07 广东美的制冷设备有限公司 Voice interaction method, device, equipment and storage medium
CN110660393B (en) * 2019-10-31 2021-12-03 广东美的制冷设备有限公司 Voice interaction method, device, equipment and storage medium
CN112231440A (en) * 2020-10-09 2021-01-15 安徽讯呼信息科技有限公司 Voice search method based on artificial intelligence
CN113486208A (en) * 2021-06-09 2021-10-08 安徽沐峰数据科技有限公司 Voice search equipment based on artificial intelligence and search method thereof

Also Published As

Publication number Publication date
US20180151183A1 (en) 2018-05-31
US10157619B2 (en) 2018-12-18

Similar Documents

Publication Publication Date Title
CN106599110A (en) Artificial intelligence-based voice search method and device
CN110728997B (en) Multi-modal depression detection system based on context awareness
KR102601848B1 (en) Device and method of data recognition model construction, and data recognition devicce
CN1157710C (en) Speech datas extraction
CN105893478B (en) A kind of tag extraction method and apparatus
CN108320734A (en) Audio signal processing method and device, storage medium, electronic equipment
CN110910283A (en) Method, device, equipment and storage medium for generating legal document
CN105095415B (en) The determination method and apparatus of network mood
CN103765506A (en) Method for tone/intonation recognition using auditory attention cues
CN105023573A (en) Speech syllable/vowel/phone boundary detection using auditory attention cues
CN103744953A (en) Network hotspot mining method based on Chinese text emotion recognition
CN108899033B (en) Method and device for determining speaker characteristics
CN108345587A (en) A kind of the authenticity detection method and system of comment
Chittaragi et al. Automatic text-independent Kannada dialect identification system
CN106649849A (en) Text information base building method and device and searching method, device and system
CN108897732A (en) Statement type recognition methods and device, storage medium and electronic device
KR20190083143A (en) Sensory evaluation method and apparatus
CN110119443A (en) A kind of sentiment analysis method towards recommendation service
CN112418172A (en) Multimode information fusion emotion analysis method based on multimode information intelligent processing unit
CN107948730A (en) Method, apparatus, equipment and storage medium based on picture generation video
CN105183808A (en) Problem classification method and apparatus
CN106844340A (en) News in brief generation and display methods, apparatus and system based on artificial intelligence
CN109241993B (en) Evaluation object emotion classification method and device integrating user and overall evaluation information
CN109478405A (en) Information processing equipment, information processing method and program
Kello et al. Scale-free networks in phonological and orthographic wordform lexicons

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20170426