[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN110362809B - Text analysis method and device - Google Patents

Text analysis method and device Download PDF

Info

Publication number
CN110362809B
CN110362809B CN201810252468.8A CN201810252468A CN110362809B CN 110362809 B CN110362809 B CN 110362809B CN 201810252468 A CN201810252468 A CN 201810252468A CN 110362809 B CN110362809 B CN 110362809B
Authority
CN
China
Prior art keywords
result
module
text
processing
feature information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810252468.8A
Other languages
Chinese (zh)
Other versions
CN110362809A (en
Inventor
茅越
蔡龙军
李明
沈一
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Youku Culture Technology Beijing Co ltd
Original Assignee
Alibaba China Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alibaba China Co Ltd filed Critical Alibaba China Co Ltd
Priority to CN201810252468.8A priority Critical patent/CN110362809B/en
Publication of CN110362809A publication Critical patent/CN110362809A/en
Application granted granted Critical
Publication of CN110362809B publication Critical patent/CN110362809B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Machine Translation (AREA)

Abstract

The disclosure relates to a text analysis method and device. The method comprises the following steps: acquiring characteristic information corresponding to a plurality of word segments of a text to be analyzed; and inputting the characteristic information into an analysis model for processing to obtain a text analysis result of the text to be analyzed, wherein the analysis model comprises an attention mechanism module, a convolution module, a relation module, a pooling module and a splicing output module. According to the embodiment of the disclosure, the feature information corresponding to a plurality of participles of the text to be analyzed can be acquired, the feature information is input into the analysis model to be processed so as to acquire the text analysis result, and the text analysis is realized by utilizing the analysis model comprising the attention mechanism module, the convolution module, the relation module, the pooling module and the concatenation output module, so that the accuracy of the text analysis result is improved.

Description

Text analysis method and device
Technical Field
The present disclosure relates to the field of computers, and in particular, to a text analysis method and apparatus.
Background
With the continuous popularization of social networks and mobile internet, the cost for people to publish information is lower and lower, and more users are willing to share own opinions and comments on people, events and products on the internet. These comments reflect the opinion and emotional tendency of people to things, and are of great significance for public opinion analysis and big data-based prediction. Therefore, the comment text of the user needs to be analyzed to determine information such as the viewpoint and emotional tendency of the user. However, the accuracy of the analysis result of analyzing the text in the related art is low.
Disclosure of Invention
In view of this, the present disclosure provides a text analysis method, which can accurately obtain a text analysis result.
According to an aspect of the present disclosure, there is provided a text analysis method including: acquiring characteristic information corresponding to a plurality of word segments of a text to be analyzed; and inputting the characteristic information into an analysis model for processing to obtain a text analysis result of the text to be analyzed, wherein the analysis model comprises an attention mechanism module, a convolution module, a relation module, a pooling module and a splicing output module.
In a possible implementation manner, inputting the feature information into an analysis model for processing, and obtaining a text analysis result of the text to be analyzed, includes:
inputting the characteristic information into the attention mechanism module for processing to obtain an expansion processing result;
inputting the extended processing result into the convolution module for processing to obtain a convolution result;
inputting the convolution result into the relation module and the pooling module respectively for processing, and obtaining a relation result and a pooling result respectively;
and inputting the relation result and the pooling result into a splicing output module for processing to obtain a text analysis result of the text to be analyzed.
In one possible implementation manner, obtaining feature information corresponding to a plurality of word segments of a text to be analyzed includes:
vectorizing the multiple participles of the text to be analyzed respectively to obtain multiple vector information corresponding to the multiple participles;
and determining the characteristic information of the word segmentation according to the vector information.
In one possible implementation, the feature information includes N feature information, N >1,
inputting the feature information into the attention mechanism module for processing, and acquiring an expansion processing result, wherein the processing method comprises the following steps:
for a first feature information in the N feature information, acquiring N-1 weights of N-1 feature information except the first feature information in the N feature information relative to the first feature information, wherein the first feature information is any one of the N feature information;
carrying out weighted summation processing on the N-1 characteristic information according to the N-1 weights, and determining expanded vector information corresponding to the first characteristic information;
splicing the N pieces of feature information and the expansion vector information corresponding to the N pieces of feature information to determine an expansion processing result,
Wherein the sum of the N-1 weights is 1.
In one possible implementation, the stitching output module comprises a plurality of fully-connected layers and a softmax processing layer,
inputting the relationship result and the pooling result into a stitching output module for processing, and acquiring a text analysis result of the text to be analyzed, wherein the method comprises the following steps:
performing vector splicing processing on the relation result and the pooling result to obtain spliced vector information;
and sequentially inputting the spliced vector information into the full connection layers and the softmax processing layer for processing to obtain a text analysis result of the text to be analyzed.
In one possible implementation, the method further includes:
acquiring training characteristic information corresponding to a plurality of participles of a sample text;
inputting the training characteristic information into an initial analysis model for processing to obtain a training analysis result of the sample text, wherein the initial analysis model comprises an initial attention mechanism module, an initial convolution module, an initial relation module, an initial pooling module and an initial splicing output module;
determining the model loss of the initial analysis model according to the training analysis result and the labeling result of the sample text;
According to the model loss, adjusting the parameter weight in the initial analysis model, and determining the adjusted analysis model;
and determining the adjusted analysis model as a final analysis model under the condition that the model loss meets the training condition.
In one possible implementation, the convolution module includes a convolutional neural network, the relationship module includes a relationship network, and the pooling module includes a maximum pooling layer.
According to another aspect of the present disclosure, there is provided a text analysis apparatus including:
the device comprises a characteristic acquisition unit, a feature analysis unit and a feature analysis unit, wherein the characteristic acquisition unit is used for acquiring characteristic information corresponding to a plurality of word segments of a text to be analyzed;
a result obtaining unit, configured to input the feature information into an analysis model for processing, and obtain a text analysis result of the text to be analyzed,
the analysis model comprises an attention mechanism module, a convolution module, a relation module, a pooling module and a splicing output module.
In one possible implementation, the result obtaining unit includes:
a first result obtaining subunit, configured to input the feature information into the attention mechanism module for processing, and obtain an extended processing result;
A second result obtaining subunit, configured to input the extended processing result into the convolution module for processing, and obtain a convolution result;
a third result obtaining subunit, configured to input the convolution result into the relationship module and the pooling module, respectively, for processing, and obtain a relationship result and a pooling result, respectively;
and the fourth result obtaining subunit is configured to input the relationship result and the pooling result into a stitching output module for processing, and obtain a text analysis result of the text to be analyzed.
In one possible implementation, the feature obtaining unit includes:
the vectorization subunit is configured to perform vectorization processing on the multiple word segments of the text to be analyzed, and acquire multiple pieces of vector information corresponding to the multiple word segments;
and the characteristic determining subunit is used for determining the characteristic information of the multiple word segmentations according to the multiple vector information.
In one possible implementation, the feature information includes N feature information, N >1,
wherein the first result obtaining subunit includes:
a weight obtaining subunit, configured to obtain, for a first feature information of the N feature information, N-1 weights, with respect to the first feature information, of N-1 feature information of the N feature information except the first feature information, where the first feature information is any one of the N feature information;
The extended information determining subunit is configured to perform weighted summation processing on the N-1 pieces of feature information according to the N-1 weights, and determine extended vector information corresponding to the first feature information;
a first splicing subunit, configured to splice the N pieces of feature information and extension vector information corresponding to the N pieces of feature information, and determine an extension processing result,
wherein the sum of the N-1 weights is 1.
In one possible implementation, the stitching output module comprises a plurality of fully connected layers and a softmax processing layer,
wherein the fourth result obtaining subunit includes:
the second splicing subunit is used for carrying out vector splicing processing on the relationship result and the pooling result to obtain spliced vector information;
and the information processing subunit is used for sequentially inputting the spliced vector information into the full connection layers and the softmax processing layer for processing to obtain a text analysis result of the text to be analyzed.
In one possible implementation, the apparatus further includes:
the training feature obtaining unit is used for obtaining training feature information corresponding to a plurality of word segments of the sample text;
A training result obtaining unit, configured to input the training feature information into an initial analysis model for processing, and obtain a training analysis result of the sample text, where the initial analysis model includes an initial attention mechanism module, an initial convolution module, an initial relationship module, an initial pooling module, and an initial concatenation output module;
a loss determining unit, configured to determine a model loss of the initial analysis model according to the training analysis result and the labeling result of the sample text;
the model adjusting unit is used for adjusting the parameter weight in the initial analysis model according to the model loss and determining an adjusted analysis model;
and a model determining unit, configured to determine the adjusted analysis model as a final analysis model when the model loss satisfies a training condition.
In one possible implementation, the convolution module includes a convolutional neural network, the relationship module includes a relationship network, and the pooling module includes a maximum pooling layer.
According to another aspect of the present disclosure, there is provided a viewpoint extracting apparatus including: a processor; a memory for storing processor-executable instructions; wherein the processor is configured to perform the above method.
According to another aspect of the present disclosure, there is provided a non-transitory computer readable storage medium having stored thereon computer program instructions, wherein the computer program instructions, when executed by a processor, implement the above-described viewpoint extraction method.
According to the embodiment of the disclosure, the feature information corresponding to a plurality of participles of the text to be analyzed can be acquired, the feature information is input into the analysis model to be processed so as to acquire the text analysis result, and the text analysis is realized by utilizing the analysis model comprising the attention mechanism module, the convolution module, the relation module, the pooling module and the splicing output module, so that the accuracy of the text analysis result is improved.
Other features and aspects of the present disclosure will become apparent from the following detailed description of exemplary embodiments, which proceeds with reference to the accompanying drawings.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate exemplary embodiments, features, and aspects of the disclosure and, together with the description, serve to explain the principles of the disclosure.
FIG. 1 is a flow diagram illustrating a method of text analysis in accordance with an exemplary embodiment.
Fig. 2 is a flowchart illustrating a step S11 of a text analysis method according to an exemplary embodiment.
FIG. 3 is a schematic diagram illustrating an analytical model of a method of text analysis in accordance with an exemplary embodiment.
Fig. 4 is a flowchart illustrating a step S12 of a text analysis method according to an exemplary embodiment.
FIG. 5 is a schematic diagram illustrating the processing of an attention mechanism module of a text analysis method in accordance with an exemplary embodiment.
FIG. 6 is a flowchart illustrating a method of text analysis in accordance with an exemplary embodiment.
Fig. 7 is a block diagram illustrating a text analysis apparatus according to an example embodiment.
Fig. 8 is a block diagram illustrating a text analysis apparatus according to an example embodiment.
Fig. 9 is a block diagram illustrating a text analysis device according to an example embodiment.
Detailed Description
Various exemplary embodiments, features and aspects of the present disclosure will be described in detail below with reference to the accompanying drawings. In the drawings, like reference numbers can indicate functionally identical or similar elements. While the various aspects of the embodiments are presented in drawings, the drawings are not necessarily drawn to scale unless specifically indicated.
The word "exemplary" is used exclusively herein to mean "serving as an example, embodiment, or illustration. Any embodiment described herein as "exemplary" is not necessarily to be construed as preferred or advantageous over other embodiments.
Furthermore, in the following detailed description, numerous specific details are set forth in order to provide a better understanding of the present disclosure. It will be understood by those skilled in the art that the present disclosure may be practiced without some of these specific details. In some instances, methods, means, elements and circuits that are well known to those skilled in the art have not been described in detail so as not to obscure the present disclosure.
FIG. 1 is a flow diagram illustrating a method of text analysis in accordance with an exemplary embodiment. The method can be applied to a server. As shown in fig. 1, a text analysis method according to an embodiment of the present disclosure includes:
in step S11, obtaining feature information corresponding to a plurality of segmented words of a text to be analyzed;
in step S12, the feature information is input into an analysis model for processing, and a text analysis result of the text to be analyzed is obtained,
the analysis model comprises an attention mechanism module, a convolution module, a relation module, a pooling module and a splicing output module.
According to the embodiment of the disclosure, the feature information corresponding to a plurality of word segments of the text to be analyzed can be acquired, the feature information is input into the analysis model to be processed so as to acquire the text analysis result, and the text analysis is realized by utilizing the analysis model comprising the attention mechanism module, the convolution module, the relation module, the pooling module and the splicing output module, so that the accuracy of the text analysis result is improved. According to the embodiment of the disclosure, the business personnel can be helped to know the comment angle, the commendatory and derogatory attitude and the like of the comment information (the text to be analyzed) of a certain object, and the value of the comment information (the text to be analyzed) is fully mined.
For example, the text to be analyzed may include comment text of a user for a certain object. The object may refer to any object capable of comment analysis, and may be, for example, video, audio, news, a character, an event, a product, or the like.
In one possible implementation, the comment text of the user may be preprocessed before being participled to improve the accuracy and efficiency of the analysis. Wherein, the preprocessing of the comment text may include: and deleting specified characters in the comment text (for example, deleting forwarding characters in comments such as microblogs), converting traditional characters in the comment text into simplified characters, and the like. After preprocessing, the text to be analyzed can be determined.
In a possible implementation manner, a word segmentation manner of the related art may be adopted to perform word segmentation processing on the text to be analyzed. For example, a new word phrase may be extracted from all the comment texts for an object, and the new word phrase may be used as a segmentation dictionary for the object. The word segmentation dictionary can be used for carrying out word segmentation on the text to be analyzed, so that a plurality of word segments of the text to be analyzed are obtained. The number of the participles is smaller than or equal to the number N of the characteristic information which can be processed by the analysis model, namely the number of the participles is smaller than or equal to N. The present disclosure does not limit the specific manner in which the plurality of segments of the text to be analyzed are obtained.
Fig. 2 is a flowchart illustrating a step S11 of a text analysis method according to an exemplary embodiment. As shown in fig. 2, in one possible implementation, step S11 may include:
in step S111, performing vectorization processing on the multiple participles of the text to be analyzed, respectively, to obtain multiple vector information corresponding to the multiple participles;
in step S112, the feature information is determined according to the plurality of vector information.
For example, a pre-trained mapping model (e.g., google word2vector model, etc.) may be used to convert (map) a plurality of segmented words of a text to be analyzed into a plurality of vector information, i.e., a plurality of real row vectors, respectively. When the number of the word segments of the text to be analyzed is less than N, the remaining positions may be filled with zeros, so that the total number of the vector information is N. The obtained N vector information may be determined as N feature information. In this way, N pieces of feature information to be input into the analysis model for processing can be obtained.
FIG. 3 is a diagram illustrating an analytical model of a method of text analysis, according to an exemplary embodiment. As shown in fig. 3, the analysis model includes an attention mechanism module 31, a convolution module 32, a relationship module 33, a pooling module 34, and a stitching output module 35.
Fig. 4 is a flowchart illustrating a step S12 of a text analysis method according to an exemplary embodiment. As shown in fig. 4, in one possible implementation, step S12 may include:
in step S121, inputting the feature information into the attention mechanism module for processing, and obtaining an expansion processing result;
in step S122, the extended processing result is input into the convolution module for processing, and a convolution result is obtained;
in step S123, the convolution results are respectively input into the relationship module and the pooling module for processing, and a relationship result and a pooling result are respectively obtained;
in step S124, the relationship result and the pooling result are input into a concatenation output module for processing, and a text analysis result of the text to be analyzed is obtained.
For example, human vision focuses attention on an area when observing an object (e.g., an object or a character), and thus an "attention mechanism" may be introduced to process word segmentation (feature information) of a text. In embodiments according to the present disclosure, the characteristic information may be input to attention mechanism module 31 for processing.
In one possible implementation manner, the feature information includes N feature information, where N >1, where step S121 may include:
aiming at first feature information in the N pieces of feature information, acquiring N-1 weights of N-1 pieces of feature information except the first feature information in the N pieces of feature information relative to the first feature information, wherein the first feature information is any one piece of feature information in the N pieces of feature information;
carrying out weighted summation processing on the N-1 characteristic information according to the N-1 weights, and determining expanded vector information corresponding to the first characteristic information;
and splicing the N pieces of feature information and the expansion vector information corresponding to the N pieces of feature information, and determining the expansion processing result, wherein the sum of the N-1 weights is 1.
In one possible implementation, N feature information (vectors) may be set to x1、x2、x3、…、xN. To is directed atAny one of feature information (first feature information) xi(i is more than or equal to 1 and less than or equal to N), and first characteristic information x except the N characteristic information can be obtained according to the attention mechanismiThe other N-1 pieces of feature information are relative to the first feature information xiN-1 weights. Wherein the sum of the N-1 weights is 1.
The N-1 feature information is subjected to weighted summation processing according to the N-1 weights, and expansion vector information g corresponding to the first feature information xi can be obtainedi. As shown in equation (1):
Figure BDA0001608142060000101
in the formula (1), xjCan represent the jth feature information in N-1 feature information, wherein j is more than or equal to 1 and less than or equal to N, and j is not equal to i, alphai,jCan represent xjRelative to xiThe weight of (c).
In one possible implementation, the weight α may be determined by the softmax function of equation (2)i,jThe value of (c).
Figure BDA0001608142060000102
Wherein, the function a can be a multilayer perceptron, p is more than or equal to 1 and less than or equal to N, and p is not equal to i. For different (x)i,xj) For a, a (x) can be calculated using the same function ai,xj). It should be understood that the function a can be selected by those skilled in the art according to actual needs, and the disclosure is not limited thereto.
Expanding vector information g obtained by formula (1)iAnd the first characteristic information xiSplicing to obtain an extended vector x'i
x′i=concat(xi,gi) (3)
FIG. 5 is a schematic diagram illustrating the processing of an attention mechanism module of a text analysis method in accordance with an exemplary embodiment. As shown in fig. 5, the value of N is, for example, 7, and the input feature information (vector) is x1、x2、x3、…、x7. For example, for the first characteristic information x3X can be obtained according to formula (2) of the attention mechanism 1、x2、x4、x5、x6、x7With respect to the first characteristic information x3Weight of alpha3,1、α3,2、α3,4、α3,5、α3,6、α3,7. X according to formula (1)1、x2、x4、x5、x6、x7Performing weighted summation to obtain expanded vector information g3. Will expand the vector information g3And the first characteristic information x3Splicing is carried out to obtain an extended vector x'3(g3+x3)。
Thus, for each feature information xiThe above processing is performed to obtain N spread vectors. These N expansion vectors may be determined as the result of the expansion process of the attention mechanism module. In this way, the accuracy of the text analysis can be improved.
In one possible implementation manner, the expansion processing result may be input to the convolution module 32 for processing in step S122, and a convolution result is obtained. The convolution module 32 may include one or more convolutional neural networks, among others. The convolutional neural network can effectively capture the context information of the local part of the sentence.
For example, for the expansion processing result (N expansion vectors), each vector is a real row vector with k dimensions, i.e., with a length of k (k >1), and the N expansion vectors form a matrix with N rows and k columns. The N rows and k columns of the matrix may be input to the convolution module 32 for processing.
D convolution cores with different weights and (h, k) can be adopted in the convolution module 32 to perform convolution operation on the matrixes of the N rows and the k columns respectively so as to extract the local information of the h continuous word segmentations. After a plurality of convolution operations, d column vectors with dimension of N-h +1 can be obtained to form a real number matrix (convolution result) with rows of N-h +1 and d columns. Each column in the real number matrix may correspond to a result of each convolution kernel operation, and each row may correspond to local information of a text to be analyzed.
In one possible implementation, the convolution module 32 may include a plurality of convolution neural networks, and the plurality of convolution neural networks respectively perform convolution processing on the extended processing result by using different convolution kernels (h, k), so as to obtain a plurality of real number matrices as the convolution result. For example, convolution kernels of h 2, 3, and 4 are used, respectively. In this way, local information of different sizes (consecutive h word segments) of the text to be analyzed can be obtained, so that the local information of different sizes can be analyzed.
It should be understood that, those skilled in the art can select the convolutional neural network according to actual needs, and set parameters such as the number of weights of the convolutional neural network and the size of the convolutional kernel, which is not limited by the present disclosure.
In a possible implementation manner, the convolution result may be input to the relationship module 33 for processing in step S123, and a relationship result may be obtained. The relationship module 33 may include one or more Relationship Networks (RNs). The relationship network can be used for capturing the long-distance dependency relationship between the word segments of the text to be analyzed and extracting the relationship information between any two local information.
For example, the convolution results may be input into the relationship module 33 for processing. Let M be N-h +1, the convolution result may be one or more real matrices of M rows and d columns. For each matrix of real numbers, each row thereof (i.e., M d-dimensional vectors of real numbers o 1、o2、…、oM) Local information of the text to be analyzed may be represented. In the relation module 33, a multi-layer perceptron b can be used to express the relation between any two local information, i.e. a relation vector b (o)q,ol) Wherein 1 is not more than q<l is less than or equal to M. For all M (M-1)/2 relation vectors b (o)q,ol) Averaging, and inputting the result into another multi-layer perceptron f for processing to obtain a relation vector r. As shown in equation (4):
Figure BDA0001608142060000121
in the case that the convolution result is one or more real number matrices, the relationship module 33 may include one or more relationship networks that respectively process the convolution result, thereby obtaining one or more relationship vectors r and using the one or more relationship vectors r as the final relationship result.
It should be understood that those skilled in the art can select the relationship network and the multi-layer perceptrons b and f according to actual needs, and the disclosure is not limited thereto. In this way, the relationship result processed by the relationship module 33 can be acquired.
In a possible implementation manner, in step S123, the convolution result may be further input to the pooling module 34 for processing, and a pooling result is obtained. Wherein pooling module 34 may include a maximum pooling layer.
For example, the convolution results may be input to pooling module 34 for processing, and pooling module 34 may, for example, include a maximum pooling layer. The convolution result may be one or more real number matrices of M rows and d columns, and each column of the matrix may represent an operation result of each convolution kernel. The maximum value of each column of the matrix can be respectively obtained, so that d maximum values are obtained. A d-dimensional real vector c of d maxima may be used as the pooling result.
In a possible implementation manner, in step S124, the relationship result r and the pooling result c may be input into the concatenation output module 35 for processing, so as to obtain a text analysis result of the text to be analyzed.
In one possible implementation, the concatenation output module 35 may include a plurality of full connection layers and softmax processing layers, wherein the step S124 may include:
performing vector splicing processing on the relation result and the pooling result to obtain spliced vector information;
and sequentially inputting the spliced vector information into the full connection layers and the softmax processing layer for processing, and acquiring a text analysis result of the text to be analyzed.
For example, the relationship result r and the pooling result c may be concatenated to obtain concatenated vector information (the length is the sum of the lengths of the relationship result r and the pooling result c). And sequentially inputting the spliced vector information into the plurality of full-connection layers and the softmax processing layer for processing, so as to obtain a text analysis result of the text to be analyzed. It should be understood that the fully connected layer and the softmax processing layer can be selected by those skilled in the art according to actual needs, and the disclosure is not limited thereto.
According to the embodiment of the disclosure, before the feature information is processed by the analysis model to obtain the text analysis result of the text to be analyzed, the initial analysis model may be trained.
FIG. 6 is a flow diagram illustrating a method of text analysis in accordance with an exemplary embodiment. As shown in fig. 6, in one possible implementation, the method further includes:
in step S13, training feature information corresponding to a plurality of participles of the sample text is obtained;
in step S14, inputting the training feature information into an initial analysis model for processing, and obtaining a training analysis result of the sample text, where the initial analysis model includes an initial attention mechanism module, an initial convolution module, an initial relationship module, an initial pooling module, and an initial concatenation output module;
in step S15, determining a model loss of the initial analysis model according to the training analysis result and the labeling result of the sample text;
in step S16, according to the model loss, adjusting the parameter weight in the initial analysis model, and determining an adjusted analysis model;
in step S17, when the model loss satisfies the training condition, the adjusted analysis model is determined as the final analysis model.
For example, existing comment texts may be manually analyzed and the analysis results (i.e., the labeling results of the sample texts) may be labeled to form a training set. The method comprises the steps of preprocessing a sample text aiming at any sample text in a training set, and performing word segmentation processing on the sample text by adopting a word segmentation mode of a related technology to obtain a plurality of word segments of the sample text. The number of the participles is smaller than or equal to the number N of the characteristic information which can be processed by the analysis model, namely the number of the participles is smaller than or equal to N.
In one possible implementation, a pre-trained mapping model (e.g., google word2vector model, etc.) may be used to map a plurality of segmented words of the sample text into a plurality of vector information, respectively. When the number of word segments is less than N, the remaining positions may be filled with zeros so that the total number of vector information is N, and the obtained N vector information is determined as training feature information (N feature information) of the sample text.
In a possible implementation manner, the training feature information may be input into an initial analysis model for processing, and a training analysis result of the sample text is obtained, where the initial analysis model includes an initial attention mechanism module, an initial convolution module, an initial relationship module, an initial pooling module, and an initial concatenation output module. The structure and form of each module of the initial analysis model may be as described above, and are not described herein again.
In one possible implementation, the model loss of the initial analysis model is determined according to the training analysis result and the labeling result of the sample text. The specific type of the loss function of the model loss can be selected by those skilled in the art according to actual situations, and the present disclosure is not limited thereto.
In a possible implementation manner, according to the model loss of the initial analysis model, the parameter weight in the initial analysis model may be adjusted, and the adjusted analysis model is determined. For example, a back Propagation algorithm, such as a bptt (back Propagation Through time) algorithm, may be employed to gradient the parameter weights of the initial analysis model based on the model loss, and adjust the parameter weights in the initial analysis model based on the gradient.
In one possible implementation, the model adjustment process of steps S14-S16 described above may be repeated multiple times. The training condition may be preset, and the training condition may include a set number of iterative training times and/or a set convergence condition. When the model loss meets the training condition, the analysis model after the last adjustment can be considered to meet the precision requirement, and the adjusted analysis model can be determined as the final analysis model.
By the method, the analysis model meeting the training condition can be obtained by training according to the training characteristic information of the sample text and the initial analysis model, so that the analysis model can accurately extract viewpoints and emotional tendencies in the text to be analyzed.
According to the embodiment of the disclosure, the feature information corresponding to a plurality of participles of the text to be analyzed can be acquired, the feature information is input into the analysis model to be processed so as to acquire the text analysis result, and the text analysis is realized by utilizing the analysis model comprising the attention mechanism module, the convolution module, the relation module, the pooling module and the splicing output module, so that the accuracy of the text analysis result is improved. According to the embodiment of the disclosure, the business personnel can be helped to know the comment angle, the commendatory and derogatory attitude and the like of the comment information (the text to be analyzed) of a certain object, and the value of the comment information (the text to be analyzed) is fully mined.
Fig. 7 is a block diagram illustrating a text analysis apparatus according to an example embodiment. As shown in fig. 7, the text analysis device includes:
a feature acquisition unit 71 configured to acquire feature information corresponding to a plurality of segmented words of a text to be analyzed;
a result obtaining unit 72, configured to input the feature information into an analysis model for processing, obtain a text analysis result of the text to be analyzed,
the analysis model comprises an attention mechanism module, a convolution module, a relation module, a pooling module and a splicing output module.
Fig. 8 is a block diagram illustrating a text analysis apparatus according to an example embodiment. As shown in fig. 8, in one possible implementation, the result obtaining unit 72 may include:
a first result obtaining subunit 721, configured to input the feature information into the attention mechanism module for processing, and obtain an extended processing result;
a second result obtaining subunit 722, configured to input the extended processing result into the convolution module for processing, so as to obtain a convolution result;
a third result obtaining subunit 723, configured to input the convolution result into the relationship module and the pooling module, respectively, for processing, and obtain a relationship result and a pooling result, respectively;
a fourth result obtaining subunit 724, configured to input the relationship result and the pooling result into a concatenation output module for processing, and obtain a text analysis result of the text to be analyzed.
As shown in fig. 8, in one possible implementation, the feature obtaining unit 71 may include:
a vectorization subunit 711, configured to perform vectorization processing on the multiple segmented words of the text to be analyzed, respectively, to obtain multiple pieces of vector information corresponding to the multiple segmented words;
A feature determining subunit 712, configured to determine feature information of the multiple word segments according to the multiple vector information.
In one possible implementation manner, the feature information includes N feature information, N >1, where the first result obtaining subunit includes:
a weight obtaining subunit, configured to obtain, for a first feature information of the N feature information, N-1 weights, with respect to the first feature information, of N-1 feature information of the N feature information except the first feature information, where the first feature information is any one of the N feature information;
the extended information determining subunit is configured to perform weighted summation processing on the N-1 pieces of feature information according to the N-1 weights, and determine extended vector information corresponding to the first feature information;
a first splicing subunit, configured to perform splicing processing on the N pieces of feature information and expansion vector information corresponding to the N pieces of feature information, and determine an expansion processing result,
wherein the sum of the N-1 weights is 1.
In one possible implementation manner, the concatenation output module includes a plurality of fully-connected layers and a softmax processing layer, wherein the fourth result obtaining subunit includes:
The second splicing subunit is used for carrying out vector splicing processing on the relation result and the pooling result to obtain spliced vector information;
and the information processing subunit is used for sequentially inputting the spliced vector information into the full connection layers and the softmax processing layer for processing to obtain a text analysis result of the text to be analyzed.
As shown in fig. 8, in one possible implementation, the apparatus further includes:
a training feature obtaining unit 73, configured to obtain training feature information corresponding to a plurality of word segments of the sample text;
a training result obtaining unit 74, configured to input the training feature information into an initial analysis model for processing, and obtain a training analysis result of the sample text, where the initial analysis model includes an initial attention mechanism module, an initial convolution module, an initial relationship module, an initial pooling module, and an initial concatenation output module;
a loss determining unit 75, configured to determine a model loss of the initial analysis model according to the training analysis result and the labeling result of the sample text;
a model adjusting unit 76, configured to adjust a parameter weight in the initial analysis model according to the model loss, and determine an adjusted analysis model;
A model determining unit 77, configured to determine the adjusted analysis model as a final analysis model when the model loss satisfies the training condition.
In one possible implementation, the convolution module includes a convolutional neural network, the relationship module includes a relationship network, and the pooling module includes a maximum pooling layer.
Fig. 9 is a block diagram illustrating a text analysis device 1900 according to an example embodiment. For example, the apparatus 1900 may be provided as a server. Referring to fig. 9, the device 1900 includes a processing component 1922 further including one or more processors and memory resources, represented by memory 1932, for storing instructions, e.g., applications, executable by the processing component 1922. The application programs stored in memory 1932 may include one or more modules that each correspond to a set of instructions. Further, the processing component 1922 is configured to execute instructions to perform the above-described method.
The device 1900 may also include a power component 1926 configured to perform power management of the device 1900, a wired or wireless network interface 1950 configured to connect the device 1900 to a network, and an input/output (I/O) interface 1958. The device 1900 may operate based on an operating system stored in memory 1932, such as Windows Server, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM, or the like.
In an exemplary embodiment, a non-transitory computer readable storage medium, such as a memory 1932, is also provided that includes computer program instructions executable by the processing component 1922 of the apparatus 1900 to perform the methods described above.
The present disclosure may be systems, methods, and/or computer program products. The computer program product may include a computer-readable storage medium having computer-readable program instructions embodied thereon for causing a processor to implement various aspects of the present disclosure.
The computer readable storage medium may be a tangible device that can hold and store the instructions for use by the instruction execution device. The computer readable storage medium may be, for example, but not limited to, an electronic memory device, a magnetic memory device, an optical memory device, an electromagnetic memory device, a semiconductor memory device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), a Static Random Access Memory (SRAM), a portable compact disc read-only memory (CD-ROM), a Digital Versatile Disc (DVD), a memory stick, a floppy disk, a mechanical coding device, such as punch cards or in-groove projection structures having instructions stored thereon, and any suitable combination of the foregoing. Computer-readable storage media as used herein is not to be construed as transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission medium (e.g., optical pulses through a fiber optic cable), or electrical signals transmitted through electrical wires.
The computer-readable program instructions described herein may be downloaded from a computer-readable storage medium to a respective computing/processing device, or to an external computer or external storage device via a network, such as the internet, a local area network, a wide area network, and/or a wireless network. The network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. The network adapter card or network interface in each computing/processing device receives computer-readable program instructions from the network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in the respective computing/processing device.
The computer program instructions for carrying out operations of the present disclosure may be assembler instructions, Instruction Set Architecture (ISA) instructions, machine-related instructions, microcode, firmware instructions, state setting data, or source or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider). In some embodiments, aspects of the disclosure are implemented by personalizing an electronic circuit, such as a programmable logic circuit, a Field Programmable Gate Array (FPGA), or a Programmable Logic Array (PLA), with state information of computer-readable program instructions, which can execute the computer-readable program instructions.
Various aspects of the present disclosure are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the disclosure. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
These computer-readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable medium storing the instructions comprises an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer, other programmable apparatus or other devices implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The foregoing description of the embodiments of the present disclosure has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein was chosen in order to best explain the principles of the embodiments, the practical application, or technical improvements to the technology in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein.

Claims (14)

1. A method of text analysis, comprising:
acquiring characteristic information corresponding to a plurality of word segments of a text to be analyzed;
inputting the characteristic information into an analysis model for processing to obtain a text analysis result of the text to be analyzed,
wherein the analysis model comprises an attention mechanism module, a convolution module, a relation module, a pooling module and a splicing output module, the relation module comprises a relation network, the relation network is used for capturing the long-distance dependency relationship among participles,
inputting the characteristic information into the attention mechanism module for processing to obtain an expansion processing result,
Inputting the expansion processing result into the convolution module for processing to obtain a convolution result,
inputting the convolution result into the relation module and the pooling module respectively for processing to obtain a relation result and a pooling result respectively,
and inputting the relation result and the pooling result into a splicing output module for processing to obtain a text analysis result of the text to be analyzed.
2. The method of claim 1, wherein obtaining feature information corresponding to a plurality of tokens of a text to be analyzed comprises:
vectorizing the multiple participles of the text to be analyzed respectively to obtain multiple vector information corresponding to the multiple participles;
and determining the characteristic information of the word segmentation according to the vector information.
3. The method of claim 1, wherein the feature information comprises N feature information, N >1,
inputting the feature information into the attention mechanism module for processing, and acquiring an expansion processing result, wherein the processing method comprises the following steps:
for a first feature information in the N feature information, acquiring N-1 weights of N-1 feature information except the first feature information in the N feature information relative to the first feature information, wherein the first feature information is any one of the N feature information;
Carrying out weighted summation processing on the N-1 characteristic information according to the N-1 weights, and determining expanded vector information corresponding to the first characteristic information;
splicing the N pieces of feature information and the expansion vector information corresponding to the N pieces of feature information to determine an expansion processing result,
wherein the sum of the N-1 weights is 1.
4. The method of claim 1, wherein the stitching output module comprises a plurality of fully-connected layers and softmax processing layers,
inputting the relationship result and the pooling result into a stitching output module for processing, and acquiring a text analysis result of the text to be analyzed, wherein the method comprises the following steps:
performing vector splicing processing on the relation result and the pooling result to obtain spliced vector information;
and sequentially inputting the spliced vector information into the full connection layers and the softmax processing layer for processing, and acquiring a text analysis result of the text to be analyzed.
5. The method of claim 1, further comprising:
acquiring training characteristic information corresponding to a plurality of participles of a sample text;
Inputting the training characteristic information into an initial analysis model for processing to obtain a training analysis result of the sample text, wherein the initial analysis model comprises an initial attention mechanism module, an initial convolution module, an initial relation module, an initial pooling module and an initial splicing output module;
determining the model loss of the initial analysis model according to the training analysis result and the labeling result of the sample text;
according to the model loss, adjusting the parameter weight in the initial analysis model, and determining an adjusted analysis model;
and determining the adjusted analysis model as a final analysis model under the condition that the model loss meets the training condition.
6. The method of any one of claims 1 to 5, wherein the convolution module comprises a convolutional neural network and the pooling module comprises a maximum pooling layer.
7. A text analysis apparatus, comprising:
the device comprises a characteristic acquisition unit, a feature analysis unit and a feature analysis unit, wherein the characteristic acquisition unit is used for acquiring characteristic information corresponding to a plurality of word segments of a text to be analyzed;
a result obtaining unit, configured to input the feature information into an analysis model for processing, and obtain a text analysis result of the text to be analyzed,
Wherein the analysis model comprises an attention mechanism module, a convolution module, a relation module, a pooling module and a splicing output module, the relation module comprises a relation network, the relation network is used for capturing the long-distance dependency relationship among the participles,
a first result obtaining subunit, configured to input the feature information into the attention mechanism module for processing, and obtain an extended processing result,
a second result obtaining subunit, configured to input the extended processing result into the convolution module for processing, so as to obtain a convolution result,
a third result obtaining subunit, configured to input the convolution result into the relationship module and the pooling module, respectively, for processing, and obtain a relationship result and a pooling result, respectively,
and the fourth result obtaining subunit is configured to input the relationship result and the pooling result into a stitching output module for processing, and obtain a text analysis result of the text to be analyzed.
8. The apparatus according to claim 7, wherein the feature acquisition unit includes:
the vectorization subunit is configured to perform vectorization processing on the multiple word segments of the text to be analyzed, and acquire multiple pieces of vector information corresponding to the multiple word segments;
And the characteristic determining subunit is used for determining the characteristic information of the multiple word segmentations according to the multiple vector information.
9. The apparatus of claim 7, wherein the feature information comprises N feature information, N >1,
wherein the first result obtaining subunit includes:
a weight obtaining subunit, configured to obtain, for a first feature information of the N feature information, N-1 weights, with respect to the first feature information, of N-1 feature information of the N feature information except the first feature information, where the first feature information is any one of the N feature information;
the extended information determining subunit is configured to perform weighted summation processing on the N-1 pieces of feature information according to the N-1 pieces of weights, and determine extended vector information corresponding to the first feature information;
a first splicing subunit, configured to splice the N pieces of feature information and extension vector information corresponding to the N pieces of feature information, and determine an extension processing result,
wherein the sum of the N-1 weights is 1.
10. The apparatus of claim 7, wherein the stitching output module comprises a plurality of fully-connected layers and softmax processing layers,
Wherein the fourth result obtaining subunit includes:
the second splicing subunit is used for carrying out vector splicing processing on the relationship result and the pooling result to obtain spliced vector information;
and the information processing subunit is used for sequentially inputting the spliced vector information into the full connection layers and the softmax processing layer for processing to obtain a text analysis result of the text to be analyzed.
11. The apparatus of claim 7, further comprising:
the training feature acquisition unit is used for acquiring training feature information corresponding to a plurality of participles of the sample text;
a training result obtaining unit, configured to input the training feature information into an initial analysis model for processing, and obtain a training analysis result of the sample text, where the initial analysis model includes an initial attention mechanism module, an initial convolution module, an initial relationship module, an initial pooling module, and an initial concatenation output module;
the loss determining unit is used for determining the model loss of the initial analysis model according to the training analysis result and the labeling result of the sample text;
the model adjusting unit is used for adjusting the parameter weight in the initial analysis model according to the model loss and determining an adjusted analysis model;
And the model determining unit is used for determining the adjusted analysis model as the final analysis model under the condition that the model loss meets the training condition.
12. The apparatus of any one of claims 7 to 11, wherein the convolution module comprises a convolutional neural network and the pooling module comprises a maximum pooling layer.
13. A text analysis apparatus, comprising:
a processor;
a memory for storing processor-executable instructions;
wherein the processor is configured to perform the method of any one of claims 1 to 6.
14. A non-transitory computer readable storage medium having computer program instructions stored thereon, wherein the computer program instructions, when executed by a processor, implement the method of any of claims 1 to 6.
CN201810252468.8A 2018-03-26 2018-03-26 Text analysis method and device Active CN110362809B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810252468.8A CN110362809B (en) 2018-03-26 2018-03-26 Text analysis method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810252468.8A CN110362809B (en) 2018-03-26 2018-03-26 Text analysis method and device

Publications (2)

Publication Number Publication Date
CN110362809A CN110362809A (en) 2019-10-22
CN110362809B true CN110362809B (en) 2022-06-14

Family

ID=68212092

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810252468.8A Active CN110362809B (en) 2018-03-26 2018-03-26 Text analysis method and device

Country Status (1)

Country Link
CN (1) CN110362809B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106372058A (en) * 2016-08-29 2017-02-01 中译语通科技(北京)有限公司 Short text emotion factor extraction method and device based on deep learning
CN106599933A (en) * 2016-12-26 2017-04-26 哈尔滨工业大学 Text emotion classification method based on the joint deep learning model
CN107180247A (en) * 2017-05-19 2017-09-19 中国人民解放军国防科学技术大学 Relation grader and its method based on selective attention convolutional neural networks
CN107341145A (en) * 2017-06-21 2017-11-10 华中科技大学 A kind of user feeling analysis method based on deep learning
CN107609009A (en) * 2017-07-26 2018-01-19 北京大学深圳研究院 Text emotion analysis method, device, storage medium and computer equipment

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8060357B2 (en) * 2006-01-27 2011-11-15 Xerox Corporation Linguistic user interface
US8346795B2 (en) * 2010-03-10 2013-01-01 Xerox Corporation System and method for guiding entity-based searching
US9965705B2 (en) * 2015-11-03 2018-05-08 Baidu Usa Llc Systems and methods for attention-based configurable convolutional neural networks (ABC-CNN) for visual question answering
CN106951438A (en) * 2017-02-13 2017-07-14 北京航空航天大学 A kind of event extraction system and method towards open field
CN107092596B (en) * 2017-04-24 2020-08-04 重庆邮电大学 Text emotion analysis method based on attention CNNs and CCR
CN107247702A (en) * 2017-05-05 2017-10-13 桂林电子科技大学 A kind of text emotion analysis and processing method and system
CN107391483A (en) * 2017-07-13 2017-11-24 武汉大学 A kind of comment on commodity data sensibility classification method based on convolutional neural networks
CN107391709A (en) * 2017-07-28 2017-11-24 深圳市唯特视科技有限公司 A kind of method that image captions generation is carried out based on new attention model
CN107515855B (en) * 2017-08-18 2021-02-12 武汉红茶数据技术有限公司 Microblog emotion analysis method and system combined with emoticons
CN107491531B (en) * 2017-08-18 2019-05-17 华南师范大学 Chinese network comment sensibility classification method based on integrated study frame
CN107526725B (en) * 2017-09-04 2021-08-24 北京百度网讯科技有限公司 Method and device for generating text based on artificial intelligence
CN107608956B (en) * 2017-09-05 2021-02-19 广东石油化工学院 Reader emotion distribution prediction algorithm based on CNN-GRNN

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106372058A (en) * 2016-08-29 2017-02-01 中译语通科技(北京)有限公司 Short text emotion factor extraction method and device based on deep learning
CN106599933A (en) * 2016-12-26 2017-04-26 哈尔滨工业大学 Text emotion classification method based on the joint deep learning model
CN107180247A (en) * 2017-05-19 2017-09-19 中国人民解放军国防科学技术大学 Relation grader and its method based on selective attention convolutional neural networks
CN107341145A (en) * 2017-06-21 2017-11-10 华中科技大学 A kind of user feeling analysis method based on deep learning
CN107609009A (en) * 2017-07-26 2018-01-19 北京大学深圳研究院 Text emotion analysis method, device, storage medium and computer equipment

Also Published As

Publication number Publication date
CN110362809A (en) 2019-10-22

Similar Documents

Publication Publication Date Title
CN109816039B (en) Cross-modal information retrieval method and device and storage medium
CN108563767B (en) Image retrieval method and device
CN109947919B (en) Method and apparatus for generating text matching model
WO2019080864A1 (en) Semantic encoding method and device for text
JP2022509327A (en) Cross-modal information retrieval method, its device, and storage medium
Shang et al. Llava-prumerge: Adaptive token reduction for efficient large multimodal models
CN108629414B (en) Deep hash learning method and device
CN114020950B (en) Training method, device, equipment and storage medium for image retrieval model
CN110309407A (en) Viewpoint extracting method and device
US11055330B2 (en) Utilizing external knowledge and memory networks in a question-answering system
CN114841142A (en) Text generation method and device, electronic equipment and storage medium
CN111898338B (en) Text generation method and device and electronic equipment
CN112926334A (en) Method and device for determining word expression vector and electronic equipment
CN110852057A (en) Method and device for calculating text similarity
CN110362809B (en) Text analysis method and device
CN110555104B (en) Text analysis method and device
CN115952854B (en) Training method of text desensitization model, text desensitization method and application
CN110362808B (en) Text analysis method and device
CN110362810B (en) Text analysis method and device
US11763082B2 (en) Accelerating inference of transformer-based models
CN112464654B (en) Keyword generation method and device, electronic equipment and computer readable medium
CN115392234A (en) Text representation method, word representation method, corresponding device, medium and equipment
CN112651231A (en) Spoken language information processing method and device and electronic equipment
CN113378774A (en) Gesture recognition method, device, equipment, storage medium and program product
CN113128180A (en) Text generation method and equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20200512

Address after: 310052 room 508, floor 5, building 4, No. 699, Wangshang Road, Changhe street, Binjiang District, Hangzhou City, Zhejiang Province

Applicant after: Alibaba (China) Co.,Ltd.

Address before: 100080 Beijing Haidian District city Haidian street A Sinosteel International Plaza No. 8 block 5 layer A, C

Applicant before: Youku network technology (Beijing) Co.,Ltd.

CB02 Change of applicant information
CB02 Change of applicant information

Address after: Room 554, 5 / F, building 3, 969 Wenyi West Road, Wuchang Street, Yuhang District, Hangzhou City, Zhejiang Province

Applicant after: Alibaba (China) Co.,Ltd.

Address before: 310052 room 508, 5th floor, building 4, No. 699 Wangshang Road, Changhe street, Binjiang District, Hangzhou City, Zhejiang Province

Applicant before: Alibaba (China) Co.,Ltd.

GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20240624

Address after: 101400 Room 201, 9 Fengxiang East Street, Yangsong Town, Huairou District, Beijing

Patentee after: Youku Culture Technology (Beijing) Co.,Ltd.

Country or region after: China

Address before: Room 554, 5 / F, building 3, 969 Wenyi West Road, Wuchang Street, Yuhang District, Hangzhou City, Zhejiang Province

Patentee before: Alibaba (China) Co.,Ltd.

Country or region before: China