[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN109522921A - Statement similarity method of discrimination and equipment - Google Patents

Statement similarity method of discrimination and equipment Download PDF

Info

Publication number
CN109522921A
CN109522921A CN201811088787.6A CN201811088787A CN109522921A CN 109522921 A CN109522921 A CN 109522921A CN 201811088787 A CN201811088787 A CN 201811088787A CN 109522921 A CN109522921 A CN 109522921A
Authority
CN
China
Prior art keywords
word
sentence
sequence
feature
full articulamentum
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811088787.6A
Other languages
Chinese (zh)
Inventor
黄建强
陆晨昱
章乐
舒畅
武拥珍
何永
李传丰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yi Language Intelligent Technology (shanghai) Co Ltd
Original Assignee
Yi Language Intelligent Technology (shanghai) Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yi Language Intelligent Technology (shanghai) Co Ltd filed Critical Yi Language Intelligent Technology (shanghai) Co Ltd
Priority to CN201811088787.6A priority Critical patent/CN109522921A/en
Publication of CN109522921A publication Critical patent/CN109522921A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The object of the present invention is to provide a kind of statement similarity method of discrimination and equipment, the method that the present invention proposes the Fusion Features for extracting data mining technology and stack convolutional neural networks can be improved the accuracy of semantic similarity judgement.

Description

Statement similarity method of discrimination and equipment
Technical field
The present invention relates to computer field more particularly to a kind of statement similarity method of discrimination and equipment.
Background technique
In recent years, with the rapid development in deep learning field, the relevant task of more and more natural language processings The method using deep learning gradually is turned to from traditional way, is also obviously improved in effect.Compare in these tasks Typically have: machine translation, text generation, emotional semantic classification, intelligent answer etc..
In existing text similarity analysis task, commonly used in the base neural net to input text vector feature extraction There are two types of networks, and one is convolutional neural networks, another kind is Recognition with Recurrent Neural Network, while also including being based on both neural networks Other improved complicated neural networks.
Different neural networks has respective advantage in terms of feature extraction, but simultaneously also without other neural network institutes Some advantages, existing method only simply simply merge various neural networks to extract more features, such net Network is computationally intensive, and the feature of extraction is also limited, causes the accuracy of analysis task not high.
Summary of the invention
It is an object of the present invention to provide a kind of statement similarity method of discrimination and equipment.
According to an aspect of the invention, there is provided a kind of statement similarity method of discrimination, this method comprises:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or The sequence expression of word is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, to obtain The feature of the Bilstm neural network of each sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and Total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification exported according to the softmax classifier As a result judge whether two sentences of the input are similar.
Further, in the above method, the expression of the sequence of the word of each sentence or word is converted into corresponding matrix sequence Column, comprising:
According to preparatory trained word or term vector model, the expression of the sequence of the word of each sentence or word is converted into pair The matrix sequence answered.
Further, in the above method, the data digging method includes:
Method, the method for editing distance or the tf-idf feature combination wordev character representation sentence of n meta-model similarity The method of text calculating cos similarity.
Further, in the above method, total eigenmatrix is inputted into full articulamentum, comprising:
Reduce the dimension of total eigenmatrix in the method in maximum and average pond, and total after dimension by reducing Eigenmatrix inputs full articulamentum.
According to another aspect of the present invention, a kind of statement similarity discriminating device is additionally provided, which includes:
First device indicates for input two sentences to be split into the sequence of corresponding word or word respectively, will be every The sequence expression of the word or word of a sentence is converted into corresponding matrix sequence;
Second device, for the corresponding matrix sequence of obtained each sentence to be inputted three layers of quick stack Bilstm mind Through network, to obtain the feature of the Bilstm neural network of each sentence;
3rd device, for from the corresponding matrix sequence of each sentence, extracting two sentences using data digging method Similarity feature;
4th device, the feature and similarity feature of the Bilstm neural network for will extract are spliced to obtain total Eigenmatrix, and the total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
5th device, for the output of the full articulamentum to be inputted softmax classifier, according to the softmax points The classification results of class device output judge whether two sentences of the input are similar.
Further, in above equipment, the first device is used for according to preparatory trained word or term vector model, The expression of the sequence of the word of each sentence or word is converted into corresponding matrix sequence.
Further, in above equipment, the data digging method includes:
Method, the method for editing distance or the tf-idf feature combination wordev character representation sentence of n meta-model similarity The method of text calculating cos similarity.
Further, in above equipment, the 4th device is described total for reducing in the method in maximum and average pond Eigenmatrix dimension, and input full articulamentum for the total eigenmatrix after dimension is reduced.
According to another aspect of the present invention, a kind of equipment based on calculating is additionally provided, wherein include:
Processor;And
It is arranged to the memory of storage computer executable instructions, the executable instruction makes the place when executed Manage device:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or The sequence expression of word is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, to obtain The feature of the Bilstm neural network of each sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and Total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification exported according to the softmax classifier As a result judge whether two sentences of the input are similar.
According to another aspect of the present invention, a kind of computer readable storage medium is additionally provided, computer is stored thereon with Executable instruction, wherein the computer executable instructions make processor when being executed by processor:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or The sequence expression of word is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, to obtain The feature of the Bilstm neural network of each sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and Total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification exported according to the softmax classifier As a result judge whether two sentences of the input are similar.
Compared with prior art, the present invention proposes the feature for extracting data mining technology and stack convolutional neural networks The method of fusion can be improved the accuracy of semantic similarity judgement.
Detailed description of the invention
By reading a detailed description of non-restrictive embodiments in the light of the attached drawings below, of the invention other Feature, objects and advantages will become more apparent upon:
Fig. 1 shows the statement similarity method of discrimination of one aspect according to the present invention and the schematic diagram of equipment.
The same or similar appended drawing reference represents the same or similar component in attached drawing.
Specific embodiment
Present invention is further described in detail with reference to the accompanying drawing.
In a typical configuration of this application, terminal, the equipment of service network and trusted party include one or more Processor (CPU), input/output interface, network interface and memory.
Memory may include the non-volatile memory in computer-readable medium, random access memory (RAM) and/or The forms such as Nonvolatile memory, such as read-only memory (ROM) or flash memory (flashRAM).Memory is showing for computer-readable medium Example.
Computer-readable medium includes permanent and non-permanent, removable and non-removable media can be by any method Or technology come realize information store.Information can be computer readable instructions, data structure, the module of program or other data. The example of the storage medium of computer includes, but are not limited to phase change memory (PRAM), static random access memory (SRAM), moves State random access memory (DRAM), other kinds of random access memory (RAM), read-only memory (ROM), electric erasable Programmable read only memory (EEPROM), flash memory or other memory techniques, read-only disc read only memory (CD-ROM) (CD-ROM), Digital versatile disc (DVD) or other optical storage, magnetic cassettes, magnetic tape disk storage or other magnetic storage devices or Any other non-transmission medium, can be used for storage can be accessed by a computing device information.As defined in this article, computer Readable medium does not include non-temporary computer readable media (transitory media), such as the data-signal and carrier wave of modulation.
As shown in Figure 1, the present invention provides a kind of statement similarity method of discrimination, comprising:
Step S1 indicates the sequence that two sentences of input split into corresponding word or word respectively, by each sentence Word or word sequence expression be converted into corresponding matrix sequence;
Here, first having to the sequence that discrete character string is switched to a vector row;
The sequence of the word of each sentence or word can be indicated to convert according to preparatory trained word or term vector model At corresponding matrix sequence;
By step S1, the natural language sentences text conversion that can enter text at corresponding matrix sequence, so as to The processing of subsequent step;
Step S2, the corresponding matrix sequence of each sentence that step S1 is obtained input three layers of quick stack Bilstm mind Through network, to obtain the feature of the Bilstm neural network of each sentence;
Here, step S2 utilizes three layers of quick stack Bilstm neural network, the front and back language of each sentence is sufficiently extracted Adopted relationship characteristic;
Step S3, step S2 is arranged side by side with step S3, using data digging method from the corresponding matrix sequence of each sentence, Extract the similarity feature of two sentences;
Here, the corresponding matrix sequence of each sentence obtained according to step S1, and two are extracted using data digging method The similarity feature of a sentence;
The data digging method includes: the method for n meta-model similarity, the method for editing distance or tf-idf feature knot Close the method that wordev character representation sentence text calculates cos similarity;
Step S4 carries out the feature for the Bilstm neural network extracted in step S2 and step S3 and similarity feature Splicing obtains total eigenmatrix, and total eigenmatrix is inputted full articulamentum, to obtain the defeated of the full articulamentum Out;
Here, total eigenmatrix is inputted full articulamentum, may include:
Reduce the dimension of total eigenmatrix in the method in maximum and average pond, and total after dimension by reducing Eigenmatrix inputs full articulamentum;
The output of the full articulamentum is inputted softmax classifier by step S5, defeated according to the softmax classifier Classification results out judge whether two sentences of the input are similar.
Here, the present invention proposes the side for the Fusion Features for extracting data mining technology and stack convolutional neural networks Method can be improved the accuracy of semantic similarity judgement.
According to another aspect of the present invention, a kind of statement similarity discriminating device is additionally provided, which includes:
First device indicates for input two sentences to be split into the sequence of corresponding word or word respectively, will be every The sequence expression of the word or word of a sentence is converted into corresponding matrix sequence;
Second device, for the corresponding matrix sequence of obtained each sentence to be inputted three layers of quick stack Bilstm mind Through network, to obtain the feature of the Bilstm neural network of each sentence;
3rd device, for from the corresponding matrix sequence of each sentence, extracting two sentences using data digging method Similarity feature;
4th device, the feature and similarity feature of the Bilstm neural network for will extract are spliced to obtain total Eigenmatrix, and the total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
5th device, for the output of the full articulamentum to be inputted softmax classifier, according to the softmax points The classification results of class device output judge whether two sentences of the input are similar.
Further, in above equipment, the first device is used for according to preparatory trained word or term vector model, The expression of the sequence of the word of each sentence or word is converted into corresponding matrix sequence.
Further, in above equipment, the data digging method includes:
Method, the method for editing distance or the tf-idf feature combination wordev character representation sentence of n meta-model similarity The method of text calculating cos similarity.
Further, in above equipment, the 4th device is described total for reducing in the method in maximum and average pond Eigenmatrix dimension, and input full articulamentum for the total eigenmatrix after dimension is reduced.
According to another aspect of the present invention, a kind of equipment based on calculating is additionally provided, wherein include:
Processor;And
It is arranged to the memory of storage computer executable instructions, the executable instruction makes the place when executed Manage device:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or The sequence expression of word is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, to obtain The feature of the Bilstm neural network of each sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and Total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification exported according to the softmax classifier As a result judge whether two sentences of the input are similar.
According to another aspect of the present invention, a kind of computer readable storage medium is additionally provided, computer is stored thereon with Executable instruction, wherein the computer executable instructions make processor when being executed by processor:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or The sequence expression of word is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, to obtain The feature of the Bilstm neural network of each sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and Total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification exported according to the softmax classifier As a result judge whether two sentences of the input are similar.
The detailed content of each equipment and storage medium embodiment of the invention, for details, reference can be made to the correspondences of each method embodiment Part, here, repeating no more.
Obviously, those skilled in the art can carry out various modification and variations without departing from the essence of the application to the application Mind and range.In this way, if these modifications and variations of the application belong to the range of the claim of this application and its equivalent technologies Within, then the application is also intended to include these modifications and variations.
It should be noted that the present invention can be carried out in the assembly of software and/or software and hardware, for example, can adopt With specific integrated circuit (ASIC), general purpose computer or any other realized similar to hardware device.In one embodiment In, software program of the invention can be executed to implement the above steps or functions by processor.Similarly, of the invention Software program (including relevant data structure) can be stored in computer readable recording medium, for example, RAM memory, Magnetic or optical driver or floppy disc and similar devices.In addition, some of the steps or functions of the present invention may be implemented in hardware, example Such as, as the circuit cooperated with processor thereby executing each step or function.
In addition, a part of the invention can be applied to computer program product, such as computer program instructions, when its quilt When computer executes, by the operation of the computer, it can call or provide according to the method for the present invention and/or technical solution. And the program instruction of method of the invention is called, it is possibly stored in fixed or moveable recording medium, and/or pass through Broadcast or the data flow in other signal-bearing mediums and transmitted, and/or be stored according to described program instruction operation In the working storage of computer equipment.Here, according to one embodiment of present invention including a device, which includes using Memory in storage computer program instructions and processor for executing program instructions, wherein when the computer program refers to When enabling by processor execution, method and/or skill of the device operation based on aforementioned multiple embodiments according to the present invention are triggered Art scheme.
It is obvious to a person skilled in the art that invention is not limited to the details of the above exemplary embodiments, Er Qie In the case where without departing substantially from spirit or essential attributes of the invention, the present invention can be realized in other specific forms.Therefore, no matter From the point of view of which point, the present embodiments are to be considered as illustrative and not restrictive, and the scope of the present invention is by appended power Benefit requires rather than above description limits, it is intended that all by what is fallen within the meaning and scope of the equivalent elements of the claims Variation is included in the present invention.Any reference signs in the claims should not be construed as limiting the involved claims.This Outside, it is clear that one word of " comprising " does not exclude other units or steps, and odd number is not excluded for plural number.That states in device claim is multiple Unit or device can also be implemented through software or hardware by a unit or device.The first, the second equal words are used to table Show title, and does not indicate any particular order.

Claims (10)

1. a kind of statement similarity method of discrimination, wherein this method comprises:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or word Sequence expression is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, it is each to obtain The feature of the Bilstm neural network of sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and by institute It states total eigenmatrix and inputs full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification results exported according to the softmax classifier Judge whether two sentences of the input are similar.
2. according to the method described in claim 1, wherein, the expression of the sequence of the word of each sentence or word is converted into corresponding Matrix sequence, comprising:
According to preparatory trained word or term vector model, the expression of the sequence of the word of each sentence or word is converted into corresponding Matrix sequence.
3. according to the method described in claim 1, wherein, the data digging method includes:
Method, the method for editing distance or the tf-idf feature combination wordev character representation sentence text of n meta-model similarity The method for calculating cos similarity.
4. according to the method described in claim 1, wherein, total eigenmatrix is inputted full articulamentum, comprising:
It reduces the dimension of total eigenmatrix in the method in maximum and average pond, and total feature after dimension will be reduced The full articulamentum of Input matrix.
5. a kind of statement similarity discriminating device, wherein the equipment includes:
First device is indicated for input two sentences to be split into the sequence of corresponding word or word respectively, by each sentence The sequence expression of the word or word of son is converted into corresponding matrix sequence;
Second device, for the corresponding matrix sequence of obtained each sentence to be inputted three layers of quick stack Bilstm nerve net Network, to obtain the feature of the Bilstm neural network of each sentence;
3rd device, for from the corresponding matrix sequence of each sentence, extracting the phase of two sentences using data digging method Like degree feature;
4th device, the feature and similarity feature of the Bilstm neural network for will extract are spliced to obtain total spy Matrix is levied, and total eigenmatrix is inputted into full articulamentum, to obtain the output of the full articulamentum;
5th device, for the output of the full articulamentum to be inputted softmax classifier, according to the softmax classifier The classification results of output judge whether two sentences of the input are similar.
6. equipment according to claim 5, wherein the first device, for according to preparatory trained word or word to Model is measured, the expression of the sequence of the word of each sentence or word is converted into corresponding matrix sequence.
7. equipment according to claim 5, wherein the data digging method includes:
Method, the method for editing distance or the tf-idf feature combination wordev character representation sentence text of n meta-model similarity The method for calculating cos similarity.
8. equipment according to claim 5, wherein the 4th device, for the method drop with maximum and average pond The dimension of low total eigenmatrix, and total eigenmatrix after reduction dimension is inputted into full articulamentum.
9. a kind of equipment based on calculating, wherein include:
Processor;And
It is arranged to the memory of storage computer executable instructions, the executable instruction makes the processing when executed Device:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or word Sequence expression is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, it is each to obtain The feature of the Bilstm neural network of sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and by institute It states total eigenmatrix and inputs full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification results exported according to the softmax classifier Judge whether two sentences of the input are similar.
10. a kind of computer readable storage medium, is stored thereon with computer executable instructions, wherein the computer is executable Instruction makes the processor when being executed by processor:
The sequence that two sentences of input split into corresponding word perhaps word respectively is indicated the word of each sentence or word Sequence expression is converted into corresponding matrix sequence;
The corresponding matrix sequence of obtained each sentence is inputted into three layers of quick stack Bilstm neural network, it is each to obtain The feature of the Bilstm neural network of sentence;
Using data digging method from the corresponding matrix sequence of each sentence, the similarity feature of two sentences is extracted;
The feature for the Bilstm neural network extracted and similarity feature are spliced to obtain total eigenmatrix, and by institute It states total eigenmatrix and inputs full articulamentum, to obtain the output of the full articulamentum;
The output of the full articulamentum is inputted into softmax classifier, the classification results exported according to the softmax classifier Judge whether two sentences of the input are similar.
CN201811088787.6A 2018-09-18 2018-09-18 Statement similarity method of discrimination and equipment Pending CN109522921A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811088787.6A CN109522921A (en) 2018-09-18 2018-09-18 Statement similarity method of discrimination and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811088787.6A CN109522921A (en) 2018-09-18 2018-09-18 Statement similarity method of discrimination and equipment

Publications (1)

Publication Number Publication Date
CN109522921A true CN109522921A (en) 2019-03-26

Family

ID=65771041

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811088787.6A Pending CN109522921A (en) 2018-09-18 2018-09-18 Statement similarity method of discrimination and equipment

Country Status (1)

Country Link
CN (1) CN109522921A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110134852A (en) * 2019-05-06 2019-08-16 北京四维图新科技股份有限公司 A kind of De-weight method of document, equipment and readable medium
CN114781496A (en) * 2022-04-01 2022-07-22 北京百度网讯科技有限公司 Optimizing sampling method and device and electronic equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105469104A (en) * 2015-11-03 2016-04-06 小米科技有限责任公司 Text information similarity calculating method, device and server
CN107291699A (en) * 2017-07-04 2017-10-24 湖南星汉数智科技有限公司 A kind of sentence semantic similarity computational methods
US20180114097A1 (en) * 2015-10-06 2018-04-26 Adobe Systems Incorporated Font Attributes for Font Recognition and Similarity
CN108182275A (en) * 2018-01-24 2018-06-19 上海互教教育科技有限公司 A kind of mathematics variant training topic supplying system and correlating method
CN108509408A (en) * 2017-02-27 2018-09-07 芋头科技(杭州)有限公司 A kind of sentence similarity judgment method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180114097A1 (en) * 2015-10-06 2018-04-26 Adobe Systems Incorporated Font Attributes for Font Recognition and Similarity
CN105469104A (en) * 2015-11-03 2016-04-06 小米科技有限责任公司 Text information similarity calculating method, device and server
CN108509408A (en) * 2017-02-27 2018-09-07 芋头科技(杭州)有限公司 A kind of sentence similarity judgment method
CN107291699A (en) * 2017-07-04 2017-10-24 湖南星汉数智科技有限公司 A kind of sentence semantic similarity computational methods
CN108182275A (en) * 2018-01-24 2018-06-19 上海互教教育科技有限公司 A kind of mathematics variant training topic supplying system and correlating method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
谢晓超: "基于深度神经网络的句子相似度研究与实现", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110134852A (en) * 2019-05-06 2019-08-16 北京四维图新科技股份有限公司 A kind of De-weight method of document, equipment and readable medium
CN110134852B (en) * 2019-05-06 2021-05-28 北京四维图新科技股份有限公司 Document duplicate removal method and device and readable medium
CN114781496A (en) * 2022-04-01 2022-07-22 北京百度网讯科技有限公司 Optimizing sampling method and device and electronic equipment
CN114781496B (en) * 2022-04-01 2023-11-07 北京百度网讯科技有限公司 Optimizing sampling method and device and electronic equipment

Similar Documents

Publication Publication Date Title
CN112685565B (en) Text classification method based on multi-mode information fusion and related equipment thereof
EP3926531A1 (en) Method and system for visio-linguistic understanding using contextual language model reasoners
EP3637295B1 (en) Risky address identification method and apparatus, and electronic device
US10740678B2 (en) Concept hierarchies
KR101657495B1 (en) Image recognition method using deep learning analysis modular systems
CN104735468B (en) A kind of method and system that image is synthesized to new video based on semantic analysis
CN111159412B (en) Classification method, classification device, electronic equipment and readable storage medium
CN111885000B (en) Network attack detection method, system and device based on graph neural network
CN106599317B (en) Test data processing method, device and the terminal of question answering system
JP2018190188A (en) Summary creating device, summary creating method and computer program
US10108661B2 (en) Using synthetic events to identify complex relation lookups
CN111143551A (en) Text preprocessing method, classification method, device and equipment
US20140236991A1 (en) Parsing and Rendering Structured Images
CN112199473A (en) Multi-turn dialogue method and device in knowledge question-answering system
US20200364216A1 (en) Method, apparatus and storage medium for updating model parameter
US20180018392A1 (en) Topic identification based on functional summarization
CN110674297B (en) Public opinion text classification model construction method, public opinion text classification device and public opinion text classification equipment
CN112784009A (en) Subject term mining method and device, electronic equipment and storage medium
CN110851797A (en) Block chain-based work creation method and device and electronic equipment
US10558760B2 (en) Unsupervised template extraction
CN110489559A (en) A kind of file classification method, device and storage medium
CN109522921A (en) Statement similarity method of discrimination and equipment
CN109299276A (en) One kind converting the text to word insertion, file classification method and device
US10229156B2 (en) Using priority scores for iterative precision reduction in structured lookups for questions
CN112765976A (en) Text similarity calculation method, device and equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190326