[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN112214594A - Text briefing generation method and device, electronic equipment and readable storage medium - Google Patents

Text briefing generation method and device, electronic equipment and readable storage medium Download PDF

Info

Publication number
CN112214594A
CN112214594A CN202011054465.7A CN202011054465A CN112214594A CN 112214594 A CN112214594 A CN 112214594A CN 202011054465 A CN202011054465 A CN 202011054465A CN 112214594 A CN112214594 A CN 112214594A
Authority
CN
China
Prior art keywords
briefing
label
vector
category
brief report
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202011054465.7A
Other languages
Chinese (zh)
Inventor
徐君妍
陈先丽
刘屹
孙敬萍
沈志勇
高宏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Merchants Finance Technology Co Ltd
Original Assignee
China Merchants Finance Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Merchants Finance Technology Co Ltd filed Critical China Merchants Finance Technology Co Ltd
Priority to CN202011054465.7A priority Critical patent/CN112214594A/en
Publication of CN112214594A publication Critical patent/CN112214594A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/34Browsing; Visualisation therefor
    • G06F16/345Summarisation for human users
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/18Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/151Transformation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Operations Research (AREA)
  • Probability & Statistics with Applications (AREA)
  • Algebra (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention relates to an artificial intelligence technology, and discloses a text briefing generation method, which comprises the following steps: the method comprises the steps of training a pre-constructed original label classification network by utilizing a first brief report set and a first category label set to obtain a standard label classification network, executing label classification operation on a second brief report set by utilizing the standard label classification network to obtain a second category label set, summarizing the first category label set and the second category label set to obtain a category label set, dividing the first brief report set and the second brief report set by utilizing the category label set to obtain a category brief report set, generating a brief report template by utilizing the category brief report set, receiving a brief report text to be generated, executing named entity recognition on the brief report text to be generated to obtain an entity set, and executing combined operation on the entity set and the brief report template to obtain a brief report. The invention also provides a text briefing generation device, electronic equipment and a computer readable storage medium. The invention can save the computing resource and improve the accuracy of the generation of the briefing.

Description

Text briefing generation method and device, electronic equipment and readable storage medium
Technical Field
The invention relates to the technical field of artificial intelligence, in particular to a text briefing generation method and device, electronic equipment and a computer readable storage medium.
Background
The brief report is the short information of the transmitted information, also called as 'brief message', and the current text brief report generation method mainly includes an extraction brief report generation method and a generation brief report generation method.
The extraction type briefing generating method comprises the steps of firstly carrying out text vectorization, calculating the similarity between each group of text vectors, storing the similarity into a matrix, and utilizing the matrix to construct a diagram structure to generate a briefing; the generation type briefing generation method inputs vectorized texts into a deep neural network of an encoder-decoder structure, iteratively trains internal parameters of the deep neural network, and then outputs vectors by using the network and decodes the vectors to obtain the briefing.
Both methods can generate the briefs, but the extraction-type briefs generation method only depends on simple mathematical calculation due to vector similarity, the randomness of similarity results is high, the briefs quality needs to be improved, the extraction-type briefs generation method needs to depend on a large amount of training data to train a neural network, the labeling of early-stage training data and the like not only wastes time and labor, but also consumes excessive computer resources to train a model based on a large amount of training data.
Disclosure of Invention
The invention provides a method and a device for generating a text briefing, electronic equipment and a computer readable storage medium, and mainly aims to save computing resources and improve the accuracy of briefing generation.
In order to achieve the above object, the method for generating a text briefing provided by the present invention comprises:
receiving a brief report set, and classifying the brief report set according to a preset proportion to obtain a first brief report set and a second brief report set;
receiving a first class label set corresponding to the first brief report set, and training a pre-constructed original label classification network by using the first brief report set and the first class label set to obtain a standard label classification network;
performing label classification operation on the second brief report set by using the standard label classification network to obtain a second class label set, summarizing the first class label set and the second class label set to obtain a class label set corresponding to the brief report set;
dividing the briefing set into different categories by using the category label set to obtain a category briefing set, and generating a briefing template by using the category briefing set;
receiving a to-be-generated brief report text, executing named entity recognition on the to-be-generated brief report text to obtain an entity set, and executing combined operation on the entity set and the brief report template to obtain a brief report.
Optionally, the training a pre-constructed original label classification network by using the first briefing set and the first class label set to obtain a standard label classification network includes:
respectively executing vectorization operation on the first briefing set and the first category label set by using a vector conversion network in the original label classification network to obtain a first briefing vector set and a first category label vector set;
performing a combination operation on the first briefing vector set and the first class label vector set to obtain a training vector set;
calculating a probability distribution value set of the training vector set by using a probability distribution function in the original label classification network;
and calculating a minimized target value of the probability distribution value set, and when the minimized target value is greater than a preset threshold value, adjusting internal parameters of the original label classification network until the minimized target value is less than or equal to the preset threshold value, so as to obtain the standard label classification network.
Optionally, the calculating a set of probability distribution values of the training vector set by using a probability distribution function in the original label classification network includes:
respectively constructing a first vector set and a second vector set according to the training vector set;
calculating a set of class prototypes for the first set of vectors;
and solving to obtain the probability distribution value set by taking the class prototype set and the second vector set as function parameters of the probability distribution function.
Optionally, the computing a set of class archetypes for the first set of vectors comprises:
calculating a class prototype set of the first vector set by adopting the following method:
Figure BDA0002710522960000021
wherein, ckClass prototype representing class label k, S (k) being said set of training vectors, fθ(xi) A vector representation function, x, representing the ith training vector in the first vector set in the vector transformation networkiRepresenting the ith training vector, y, in the first set of vectorsiAnd representing a category label vector corresponding to the ith training vector in the first vector set.
Optionally, said solving a set of probability distribution values of the class prototype set and the second vector set using the probability distribution function includes:
solving a set of probability distribution values of the class prototype set and the second vector set by using the following probability distribution functions:
Figure BDA0002710522960000031
wherein x isjRepresenting the jth training vector, p, in said second set of vectorsθ(y=k|xj) A probability distribution value, f, representing that the jth training vector in the second vector set belongs to a class label kθ(xj) Representing a vector representation function of a jth training vector of said second set of vectors in said vector transformation network, ckA class prototype representing a class label k.
Optionally, the generating a presentation template by using the category presentation set includes:
performing attribute replacement operation and part-of-speech tagging operation on the category briefing set to obtain a primary briefing set;
and sequencing the sentences of the primary briefing set to obtain the briefing template.
Optionally, the preset ratio is 4: 6.
In order to solve the above problem, the present invention further provides a text briefing generating apparatus, including:
the briefing dividing module is used for receiving a briefing set, classifying the briefing set according to a preset proportion, and obtaining a first briefing set and a second briefing set;
the classification network training module is used for receiving a first class label set corresponding to the first brief report set, and training a pre-constructed original label classification network by using the first brief report set and the first class label set to obtain a standard label classification network;
a category label identification module, configured to perform a label classification operation on the second brief report set by using the standard label classification network to obtain a second category label set, and summarize the first category label set and the second category label set to obtain a category label set corresponding to the brief report set;
the briefing generating module is used for dividing the briefing set into different categories by using the category label set to obtain a category briefing set, generating a briefing template by using the category briefing set, receiving a briefing text to be generated, performing named entity identification on the briefing text to be generated to obtain an entity set, and performing combined operation on the entity set and the briefing template to obtain the briefing.
In order to solve the above problem, the present invention also provides an electronic device, including:
at least one processor; and the number of the first and second groups,
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to cause the at least one processor to perform the method of generating a text presentation described above.
In order to solve the above problem, the present invention further provides a computer-readable storage medium including a storage data area and a storage program area, the storage data area storing created data, the storage program area storing a computer program; wherein the computer program, when executed by a processor, implements a method of generating a text presentation as described in any of the above.
The embodiment of the invention divides the brief report set into two to obtain a first brief report set and a second brief report set, trains a label classification network by utilizing the first brief report set and the corresponding first class label set, performs label classification on the second brief report set by utilizing the trained label classification network to obtain a second class label set, and summarizes the first class label set and the second class label set to obtain the class label set corresponding to the brief report set. Because the embodiment of the invention only uses part of the brief report set training model and classifies the rest brief report sets based on the trained model, compared with relying on the model training using all the brief report training data, the embodiment of the invention not only reduces the label of the brief report training data in the early stage, but also reduces the model calculation amount and saves the computer resources because of the reduction of the training data. Therefore, the method, the device and the computer readable storage medium for generating the text briefing can save computing resources and improve the accuracy of briefing generation.
Drawings
Fig. 1 is a schematic flow chart of a method for generating a text presentation according to an embodiment of the present invention;
fig. 2 is a schematic detailed flowchart of S2 in the text briefing generation method according to an embodiment of the present invention;
fig. 3 is a detailed flowchart of S3 in the text briefing generation method according to an embodiment of the present invention;
fig. 4 is a schematic block diagram of an apparatus for generating a text presentation according to an embodiment of the present invention;
fig. 5 is a schematic internal structural diagram of an electronic device implementing a method for generating a text presentation according to an embodiment of the present invention;
the implementation, functional features and advantages of the objects of the present invention will be further explained with reference to the accompanying drawings.
Detailed Description
It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The embodiment of the application provides a text briefing generation method. The execution subject of the text briefing generation method includes, but is not limited to, at least one of electronic devices such as a server and a terminal, which can be configured to execute the method provided by the embodiment of the present application. In other words, the text presentation generation method may be executed by software or hardware installed in the terminal device or the server device, and the software may be a block chain platform. The server includes but is not limited to: a single server, a server cluster, a cloud server or a cloud server cluster, and the like.
The invention provides a text briefing generation method. Fig. 1 is a schematic flow chart of a method for generating a text presentation according to an embodiment of the present invention. In this embodiment, the method for generating a text presentation includes:
s1, receiving the briefing sets, and classifying the briefing sets according to a preset proportion to obtain a first briefing set and a second briefing set.
The briefing set is a set of briefing under a specific scene, wherein the specific scene comprises a health field briefing scene, an enterprise analysis briefing scene and the like. Taking a health field briefing scene as an example, under the current broad background of the prevalence of the novel coronavirus pneumonia, how to efficiently and accurately extract brief information from massive and various articles to form a briefing for a user to refer is very important, so that the embodiment of the invention obtains a briefing set related to the novel coronavirus pneumonia from a website, a database and the like by using a crawler or a user collection and arrangement mode and the like.
The preset ratio is generally specified by a user, and in a preferred embodiment of the present invention, the preset ratio is 4:6, wherein the data amount of the first presentation set accounts for 40% of the presentation set, and the data amount of the second presentation set accounts for 60% of the presentation set.
S2, receiving a first category label set corresponding to the first brief report set, and training a pre-constructed original label classification network by using the first brief report set and the first category label set to obtain a standard label classification network.
Preferably, the category labels in the first category label set are determined according to specific scenes, and for a new coronary pneumonia brief report scene, the category labels include category 1 "suspected or confirmed by oneself", category 2 "whether suspected or confirmed by oneself is touched, category 3" fever and discomfort symptom by oneself ", and the like. The first category tag set is generally determined by an expert in a specific scene, and corresponds to the first brief report set in a many-to-one or one-to-one manner, for example, brief reports a exist in the first brief report set, the first sentence of the brief report a is category 1, and the second sentence and the third sentence are category 2.
Further, the original label classification network comprises a vector transformation network and a probability distribution function. In detail, referring to fig. 2, the training of the pre-constructed original tag classification network by using the first briefing set and the first category tag set to obtain the standard tag classification network includes:
s21, executing vectorization operation on the first briefing set and the first category label set by using a vector conversion network in the original label classification network to obtain a first briefing vector set and a first category label vector set;
s22, combining the first briefing vector set and the first class label vector set to obtain a training vector set;
in a preferred embodiment of the present invention, the vector transformation network adopts a Bert deep learning network, and after vectorization operation is performed by the Bert deep learning network, the first presentation vector set { x } is obtained1,x2,x3,...,xnAnd the first category label vector set { y }1,y2,y3,...,ykAnd for the first presentation vector set { x }1,x2,x3,...,xnAnd the first category label vector set { y }1,y2,y3,...,ykPerforming a combining operation to obtain the training vector set s (k) { (x)1,y1),(x2,y2),...,(xn,yk)}。
To explain further, in the training vector set s (k) { (x)1,y1),(x2,y2),...,(xn,yk) In the method, the presentation vector and the category label vector are in one-to-many or one-to-one relationship, so each presentation vector xnMay correspond to a plurality of class label vectors, so the training vectors in the training vector set may also be
Figure BDA0002710522960000061
In the form of (1).
S23, calculating a probability distribution value set of the training vector set by using a probability distribution function in the original label classification network;
further, the S23 includes: respectively constructing a first vector set and a second vector set according to the training vector set; and calculating a class prototype set of the first vector set, and solving the probability distribution value sets of the class prototype set and the second vector set.
In detail, the constructing and obtaining a first vector set and a second vector set according to the training vector set respectively includes: and selecting training vectors from the training vector set according to a first preset number to obtain the first vector set, and selecting training vectors of a second preset number from the training vector set to obtain the second vector set.
In detail, the first preset number and the second preset number are generally empirical values and are not greater than the data amount of the training vector set.
In a preferred embodiment of the present invention, the class prototype refers to training vectors belonging to the same class label number, and the average value in the training vector set is calculated as follows:
Figure BDA0002710522960000071
wherein, ckClass prototype representing class label k, S (k) being said set of training vectors, fθ(xi) Representing the ith training vector in said first set of vectors, a vector representation function in said vector transformation network, xiRepresenting the ith training vector, y, in the first set of vectorsiAnd representing a category label vector corresponding to the ith training vector in the first vector set.
In a preferred embodiment of the present invention, the probability distribution value sets of the class prototype set and the second vector set are solved by using the following probability distribution functions:
Figure BDA0002710522960000072
wherein x isjRepresenting the jth training vector, p, in said second set of vectorsθ(y=k|xj) A probability distribution value, f, representing that the jth training vector in the second vector set belongs to a class label kθ(xj) Representing the jth training vector in said second vector set, a vector representation function in said vector transformation network, ckA class prototype representing a class label k.
And S24, calculating a minimized target value of the probability distribution value set, adjusting internal parameters of the original label classification network when the minimized target value is larger than a preset threshold value, and returning to S21 until the minimized target value is smaller than or equal to the preset threshold value, so as to obtain the standard label classification network.
In the preferred embodiment of the present invention, a gradient descent algorithm is used to calculate the minimum target value of the probability distribution value set until a preset threshold requirement is met, so as to obtain the standard label classification network.
S3, performing label classification operation on the second brief report set by using the standard label classification network to obtain a second class label set, summarizing the first class label set and the second class label set to obtain a class label set corresponding to the brief report set.
In a preferred embodiment of the present invention, a standard label classification network is used to perform vectorization operations on the second briefing set, and the probability distribution function is used to solve the class label corresponding to each briefing, so as to obtain the second class label set by summarization.
As can be seen from the above, in the preferred embodiment of the present invention, the first presentation set accounts for 40% of the presentation set, the second presentation set accounts for 60% of the presentation set, and the first category tag set corresponding to the first presentation set and the second category tag set corresponding to the second presentation set are summarized, so as to obtain the category tag set corresponding to the presentation set.
And S4, dividing the briefing set into different categories by using the category label set to obtain a category briefing set, and generating a briefing template by using the category briefing set.
As stated in S2, the category labels correspond to the presentations in a many-to-one or one-to-one correspondence relationship, and if the first sentence of the presentation a is the category 1, and the second sentence and the third sentence are the category 2, the category label set and the presentation set have a many-to-one or one-to-one correspondence relationship, so that each word or the corresponding category label of each segment in each presentation in the presentation set is divided, and the category presentation set is obtained by summarizing.
Further, referring to fig. 3, the generating a presentation template by using the category presentation set includes:
s41, performing attribute replacement operation and part-of-speech tagging operation on the category brief report set to obtain a primary brief report set;
and S42, carrying out sentence sequencing on the primary briefing set to obtain the briefing template.
In detail, the attribute replacement operation adopts a multi-mode trie tree matching method, for example, specific names appearing in the category brief report set are uniformly replaced by 'name' attributes, telephone numbers in a digital form are uniformly replaced by 'telephone', addresses are uniformly replaced by 'cities', for example, 'three persons in innovation centers have cough and running nose symptoms, the current isolation at home is that three persons in the city are 17731897168', and after the attribute replacement operation is executed, the following steps are performed: the < department > employee < name > has symptoms of fever and running nose, and the < name > telephone is a < telephone number > isolated at home at present.
The part-of-speech tagging operation adopts a natural language processing method such as HanLP, wherein w represents punctuation marks, n is a noun, v is a verb, and t is a time word, for example, a word expression form (a < department > employee < name > has symptoms of fever and running nose, and is isolated at home) is converted into a primary brief including (wnwnwnwvvwnnwtvv) part-of-speech arrangement.
The sequence of the sentences is also important for readability of the presentation, and in the preferred embodiment of the invention, the WER value of each primary presentation is calculated by using a WER method (Word Error Rate), and the presentation template is obtained by sequencing the sentences of the primary presentation set according to the WER value of each primary presentation.
S5, receiving the to-be-generated brief report text, executing named entity recognition on the to-be-generated brief report text to obtain an entity set, and executing combined operation on the entity set and the brief report template to obtain the brief report.
In a preferred embodiment of the present invention, the named entity recognition may adopt a bidirectional cyclic neural network BiLSTM, a conditional random field CRF model, etc., and further, according to a specific scenario of the brief report, a corresponding entity set is obtained through the named entity recognition, for example, a specific scenario of the novel coronavirus pneumonia, and the entity set includes a name, an organization, a department, a telephone, an address, a body temperature, symptoms, etc.
Further, according to the part-of-speech arrangement and keyword information corresponding to the entity set, a combination operation is executed with the briefing template, such that the entity with the part-of-speech arrangement (wnwnwvvwnwtvv) exists in the entity set
Figure BDA0002710522960000091
And the part-of-speech arrangement of the template is the same as that of a certain template of the presentation template, and the keywords in the presentation set corresponding to the presentation template have a matching relationship with the keywords of the presentation text to be generated, the entity sets are sequentially filled into the presentation template to obtain the presentation corresponding to the presentation text to be generated.
The embodiment of the invention divides the brief report set into two to obtain a first brief report set and a second brief report set, trains a label classification network by utilizing the first brief report set and the corresponding first class label set, performs label classification on the second brief report set by utilizing the trained label classification network to obtain a second class label set, and summarizes the first class label set and the second class label set to obtain the class label set corresponding to the brief report set. Because the embodiment of the invention only uses part of the brief report set training model and classifies the rest brief report sets based on the trained model, compared with relying on the model training using all the brief report training data, the embodiment of the invention not only reduces the label of the brief report training data in the early stage, but also reduces the model calculation amount and saves the computer resources because of the reduction of the training data. Therefore, the method, the device and the computer readable storage medium for generating the text briefing can save computing resources and improve the accuracy of briefing generation.
Fig. 4 is a schematic block diagram of a text presentation generating apparatus according to the present invention.
The apparatus 100 for generating a text presentation according to the present invention may be installed in an electronic device. According to the realized functions, the text briefing generating device can comprise a briefing dividing module 101, a classification network training module 102, a class label identification module 103 and a briefing generating module 104. A module according to the present invention, which may also be referred to as a unit, refers to a series of computer program segments that can be executed by a processor of an electronic device and that can perform a fixed function, and that are stored in a memory of the electronic device.
In the present embodiment, the functions regarding the respective modules/units are as follows:
the briefing dividing module 101 is configured to receive a briefing set, classify the briefing set according to a preset proportion, and obtain a first briefing set and a second briefing set;
the classification network training module 102 is configured to receive a first class label set corresponding to the first brief report set, train a pre-constructed original label classification network by using the first brief report set and the first class label set, and obtain a standard label classification network;
the category label identification module 103 is configured to perform a label classification operation on the second presentation set by using the standard label classification network to obtain a second category label set, and summarize the first category label set and the second category label set to obtain a category label set corresponding to the presentation set;
the briefing generating module 104 is configured to divide the briefing set into different categories by using the category tag set to obtain a category briefing set, generate a briefing template by using the category briefing set, receive a briefing text to be generated, perform named entity identification on the briefing text to be generated to obtain an entity set, and perform a combining operation on the entity set and the briefing template to obtain a briefing.
In detail, the specific implementation of each module of the driving violation detection device is as follows:
the briefing dividing module 101 is configured to receive a briefing set, classify the briefing set according to a preset proportion, and obtain a first briefing set and a second briefing set.
The briefing set is a set of briefing under a specific scene, wherein the specific scene comprises a health field briefing scene, an enterprise analysis briefing scene and the like. Taking a health field briefing scene as an example, under the current broad background of the prevalence of the novel coronavirus pneumonia, how to efficiently and accurately extract brief information from massive and various articles to form a briefing for a user to refer is very important, so that the embodiment of the invention obtains a briefing set related to the novel coronavirus pneumonia from a website, a database and the like by using a crawler or a user collection and arrangement mode and the like.
The preset ratio is generally specified by a user, and in a preferred embodiment of the present invention, the preset ratio is 4:6, wherein the data amount of the first presentation set accounts for 40% of the presentation set, and the data amount of the second presentation set accounts for 60% of the presentation set.
The classification network training module 102 is configured to receive a first class label set corresponding to the first presentation set, train a pre-constructed original label classification network by using the first presentation set and the first class label set, and obtain a standard label classification network.
Preferably, the category labels in the first category label set are determined according to specific scenes, and for a new coronary pneumonia brief report scene, the category labels include category 1 "suspected or confirmed by oneself", category 2 "whether suspected or confirmed by oneself is touched, category 3" fever and discomfort symptom by oneself ", and the like. The first category tag set is generally determined by an expert in a specific scene, and corresponds to the first brief report set in a many-to-one or one-to-one manner, for example, brief reports a exist in the first brief report set, the first sentence of the brief report a is category 1, and the second sentence and the third sentence are category 2.
Further, the original label classification network comprises a vector transformation network and a probability distribution function. In detail, the training of the pre-constructed original label classification network by using the first briefing set and the first class label set to obtain a standard label classification network includes:
step A: utilizing a vector conversion network in the original label classification network to perform vectorization operation on the first briefing set and the first category label set to obtain a first briefing vector set and a first category label vector set;
and B: combining the first briefing vector set and the first class label vector set to obtain a training vector set;
and C: calculating a probability distribution value set of the training vector set by using a probability distribution function in the original label classification network;
step D: and B, calculating a minimized target value of the probability distribution value set, adjusting internal parameters of the original label classification network when the minimized target value is larger than a preset threshold value, and returning to the step A until the minimized target value is smaller than or equal to the preset threshold value to obtain the standard label classification network.
In a preferred embodiment of the present invention, the vector transformation network adopts a Bert deep learning network, and after vectorization operation is performed by the Bert deep learning network, the first presentation vector set { x } is obtained1,x2,x3,...,xnAnd the first category label vector set { y }1,y2,y3,...,ykAnd for the first presentation vector set { x }1,x2,x3,...,xnAnd the first category label vector set { y }1,y2,y3,...,ykPerforming a combining operation to obtain the training vector set s (k) { (x)1,y1),(x2,y2),...,(xn,yk)}。
To explain further, in the training vector set s (k) { (x)1,y1),(x2,y2),...,(xn,yk) In the method, the presentation vector and the category label vector are in one-to-many or one-to-one relationship, so each presentation vector xnMay correspond to a plurality of class label vectors, so the training vectors in the training vector set may also be
Figure BDA0002710522960000111
In the form of (1).
Further, the calculating a set of probability distribution values of the training vector set using the probability distribution function in the original label classification network includes: respectively constructing a first vector set and a second vector set according to the training vector set; and calculating a class prototype set of the first vector set, and solving the probability distribution value sets of the class prototype set and the second vector set.
In detail, the constructing and obtaining a first vector set and a second vector set according to the training vector set respectively includes: and selecting training vectors from the training vector set according to a first preset number to obtain the first vector set, and selecting training vectors of a second preset number from the training vector set to obtain the second vector set.
In detail, the first preset number and the second preset number are generally empirical values and are not greater than the data amount of the training vector set.
In a preferred embodiment of the present invention, the class prototype refers to training vectors belonging to the same class label number, and the average value in the training vector set is calculated as follows:
Figure BDA0002710522960000122
wherein, ckClass prototype representing class label k, S (k) being said set of training vectors, fθ(xi) Representing the ith training vector in said first set of vectors, a vector representation function in said vector transformation network, xiRepresenting the ith training vector, y, in the first set of vectorsiAnd representing a category label vector corresponding to the ith training vector in the first vector set.
In a preferred embodiment of the present invention, the probability distribution value sets of the class prototype set and the second vector set are solved by using the following probability distribution functions:
Figure BDA0002710522960000121
wherein x isjRepresenting the jth training vector, p, in said second set of vectorsθ(y=k|xj) A probability distribution value, f, representing that the jth training vector in the second vector set belongs to a class label kθ(xj) Representing the jth training vector in said second vector set, a vector representation function in said vector transformation network, ckA class prototype representing a class label k.
In the preferred embodiment of the present invention, a gradient descent algorithm is used to calculate the minimum target value of the probability distribution value set until a preset threshold requirement is met, so as to obtain the standard label classification network.
The category label identification module 103 is configured to perform a label classification operation on the second presentation set by using the standard label classification network to obtain a second category label set, and summarize the first category label set and the second category label set to obtain a category label set corresponding to the presentation set.
In a preferred embodiment of the present invention, a standard label classification network is used to perform vectorization operations on the second briefing set, and the probability distribution function is used to solve the class label corresponding to each briefing, so as to obtain the second class label set by summarization.
As can be seen from the above, in the preferred embodiment of the present invention, the first presentation set accounts for 40% of the presentation set, the second presentation set accounts for 60% of the presentation set, and the first category tag set corresponding to the first presentation set and the second category tag set corresponding to the second presentation set are summarized, so as to obtain the category tag set corresponding to the presentation set.
The briefing generating module 104 is configured to divide the briefing set into different categories by using the category tag set to obtain a category briefing set, generate a briefing template by using the category briefing set, receive a briefing text to be generated, perform named entity identification on the briefing text to be generated to obtain an entity set, and perform a combining operation on the entity set and the briefing template to obtain a briefing.
The category labels and the briefs are in a many-to-one or one-to-one corresponding relationship, if the first sentence of the briefs A is the category 1, and the second sentence and the third sentence are the category 2, each sentence or each section of the corresponding category label in each briefs in the briefs set is divided through the many-to-one or one-to-one corresponding relationship between the category label set and the briefs set, and the category briefs set is obtained through summarization.
Further, the generating a presentation template by using the category presentation set includes: performing attribute replacement operation and part-of-speech tagging operation on the category briefing set to obtain a primary briefing set; and sequencing the sentences of the primary briefing set to obtain the briefing template.
In detail, the attribute replacement operation adopts a multi-mode trie tree matching method, for example, specific names appearing in the category brief report set are uniformly replaced by 'name' attributes, telephone numbers in a digital form are uniformly replaced by 'telephone', addresses are uniformly replaced by 'cities', for example, 'three persons in innovation centers have cough and running nose symptoms, the current isolation at home is that three persons in the city are 17731897168', and after the attribute replacement operation is executed, the following steps are performed: the < department > employee < name > has symptoms of fever and running nose, and the < name > telephone is a < telephone number > isolated at home at present.
The part-of-speech tagging operation adopts a natural language processing method such as HanLP, wherein w represents punctuation marks, n is a noun, v is a verb, and t is a time word, for example, a word expression form (a < department > employee < name > has symptoms of fever and running nose, and is isolated at home) is converted into a primary brief including (wnwnwnwvvwnnwtvv) part-of-speech arrangement.
The sequence of the sentences is also important for readability of the presentation, and in the preferred embodiment of the invention, the WER value of each primary presentation is calculated by using a WER method (Word Error Rate), and the presentation template is obtained by sequencing the sentences of the primary presentation set according to the WER value of each primary presentation.
In a preferred embodiment of the present invention, the named entity recognition may adopt a bidirectional cyclic neural network BiLSTM, a conditional random field CRF model, etc., and further, according to a specific scenario of the brief report, a corresponding entity set is obtained through the named entity recognition, for example, a specific scenario of the novel coronavirus pneumonia, and the entity set includes a name, an organization, a department, a telephone, an address, a body temperature, symptoms, etc.
Further, according to the part-of-speech arrangement and keyword information corresponding to the entity set, a combination operation is executed with the briefing template, such that the entity with the part-of-speech arrangement (wnwnwvvwnwtvv) exists in the entity set
Figure BDA0002710522960000141
And the part-of-speech arrangement of the template is the same as that of a certain template of the presentation template, and the keywords in the presentation set corresponding to the presentation template have a matching relationship with the keywords of the presentation text to be generated, the entity sets are sequentially filled into the presentation template to obtain the presentation corresponding to the presentation text to be generated.
Fig. 5 is a schematic structural diagram of an electronic device implementing the method for generating a text presentation according to the present invention.
The electronic device 1 may comprise a processor 10, a memory 11 and a bus, and may further comprise a computer program, such as a text briefing generating program 12, stored in the memory 11 and executable on the processor 10.
The memory 11 includes at least one type of readable storage medium, which includes flash memory, removable hard disk, multimedia card, card-type memory (e.g., SD or DX memory, etc.), magnetic memory, magnetic disk, optical disk, etc. The memory 11 may in some embodiments be an internal storage unit of the electronic device 1, such as a removable hard disk of the electronic device 1. The memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in mobile hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), and the like, which are provided on the electronic device 1. Further, the memory 11 may also include both an internal storage unit and an external storage device of the electronic device 1. The memory 11 may be used not only to store application software installed in the electronic device 1 and various types of data, such as codes of the text briefing generating program 12, etc., but also to temporarily store data that has been output or is to be output.
The processor 10 may be composed of an integrated circuit in some embodiments, for example, a single packaged integrated circuit, or may be composed of a plurality of integrated circuits packaged with the same or different functions, including one or more Central Processing Units (CPUs), microprocessors, digital Processing chips, graphics processors, and combinations of various control chips. The processor 10 is a Control Unit (Control Unit) of the electronic device, connects various components of the electronic device by using various interfaces and lines, and executes various functions and processes data of the electronic device 1 by running or executing programs or modules (for example, executing a text presentation generation program, etc.) stored in the memory 11 and calling data stored in the memory 11.
The bus may be a Peripheral Component Interconnect (PCI) bus, an Extended Industry Standard Architecture (EISA) bus, or the like. The bus may be divided into an address bus, a data bus, a control bus, etc. The bus is arranged to enable connection communication between the memory 11 and at least one processor 10 or the like.
Fig. 5 only shows an electronic device with components, and it will be understood by a person skilled in the art that the structure shown in fig. 5 does not constitute a limitation of the electronic device 1, and may comprise fewer or more components than shown, or a combination of certain components, or a different arrangement of components.
For example, although not shown, the electronic device 1 may further include a power supply (such as a battery) for supplying power to each component, and preferably, the power supply may be logically connected to the at least one processor 10 through a power management device, so as to implement functions of charge management, discharge management, power consumption management, and the like through the power management device. The power supply may also include any component of one or more dc or ac power sources, recharging devices, power failure detection circuitry, power converters or inverters, power status indicators, and the like. The electronic device 1 may further include various sensors, a bluetooth module, a Wi-Fi module, and the like, which are not described herein again.
Further, the electronic device 1 may further include a network interface, and optionally, the network interface may include a wired interface and/or a wireless interface (such as a WI-FI interface, a bluetooth interface, etc.), which are generally used for establishing a communication connection between the electronic device 1 and other electronic devices.
Optionally, the electronic device 1 may further comprise a user interface, which may be a Display (Display), an input unit (such as a Keyboard), and optionally a standard wired interface, a wireless interface. Alternatively, in some embodiments, the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch device, or the like. The display, which may also be referred to as a display screen or display unit, is suitable for displaying information processed in the electronic device 1 and for displaying a visualized user interface, among other things.
It is to be understood that the described embodiments are for purposes of illustration only and that the scope of the appended claims is not limited to such structures.
The text briefing generating program 12 stored in the memory 11 of the electronic device 1 is a combination of instructions that, when executed in the processor 10, enable:
receiving a brief report set, and classifying the brief report set according to a preset proportion to obtain a first brief report set and a second brief report set;
receiving a first class label set corresponding to the first brief report set, and training a pre-constructed original label classification network by using the first brief report set and the first class label set to obtain a standard label classification network;
performing label classification operation on the second brief report set by using the standard label classification network to obtain a second class label set, summarizing the first class label set and the second class label set to obtain a class label set corresponding to the brief report set;
dividing the briefing set into different categories by using the category label set to obtain a category briefing set, and generating a briefing template by using the category briefing set;
receiving a to-be-generated brief report text, executing named entity recognition on the to-be-generated brief report text to obtain an entity set, and executing combined operation on the entity set and the brief report template to obtain a brief report.
Further, the integrated modules/units of the electronic device 1, if implemented in the form of software functional units and sold or used as separate products, may be stored in a computer readable storage medium. The computer-readable medium may include: any entity or device capable of carrying said computer program code, recording medium, U-disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM).
Further, the computer usable storage medium may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function, and the like; the storage data area may store data created according to the use of the blockchain node, and the like.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus, device and method can be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, the division of the modules is only one logical functional division, and other divisions may be realized in practice.
The modules described as separate parts may or may not be physically separate, and parts displayed as modules may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment.
In addition, functional modules in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, or in a form of hardware plus a software functional module.
It will be evident to those skilled in the art that the invention is not limited to the details of the foregoing illustrative embodiments, and that the present invention may be embodied in other specific forms without departing from the spirit or essential attributes thereof.
The present embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims rather than by the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein. Any accompanying claims should not be construed as limiting the claim concerned.
The block chain is a novel application mode of computer technologies such as distributed data storage, point-to-point transmission, a consensus mechanism, an encryption algorithm and the like. A block chain (Blockchain), which is essentially a decentralized database, is a series of data blocks associated by using a cryptographic method, and each data block contains information of a batch of network transactions, so as to verify the validity (anti-counterfeiting) of the information and generate a next block. The blockchain may include a blockchain underlying platform, a platform product service layer, an application service layer, and the like.
Furthermore, it is obvious that the word "comprising" does not exclude other elements or steps, and the singular does not exclude the plural. A plurality of units or means recited in the system claims may also be implemented by one unit or means in software or hardware. The terms second, etc. are used to denote names, but not any particular order.
Finally, it should be noted that the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting, and although the present invention is described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made on the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention.

Claims (10)

1. A method for generating a text presentation, the method comprising:
receiving a brief report set, and classifying the brief report set according to a preset proportion to obtain a first brief report set and a second brief report set;
receiving a first class label set corresponding to the first brief report set, and training a pre-constructed original label classification network by using the first brief report set and the first class label set to obtain a standard label classification network;
performing label classification operation on the second brief report set by using the standard label classification network to obtain a second class label set, summarizing the first class label set and the second class label set to obtain a class label set corresponding to the brief report set;
dividing the briefing set into different categories by using the category label set to obtain a category briefing set, and generating a briefing template by using the category briefing set;
receiving a to-be-generated brief report text, executing named entity recognition on the to-be-generated brief report text to obtain an entity set, and executing combined operation on the entity set and the brief report template to obtain a brief report.
2. The method of claim 1, wherein the training of a pre-constructed original tag classification network using the first presentation set and the first category tag set to obtain a standard tag classification network comprises:
respectively executing vectorization operation on the first briefing set and the first category label set by using a vector conversion network in the original label classification network to obtain a first briefing vector set and a first category label vector set;
performing a combination operation on the first briefing vector set and the first class label vector set to obtain a training vector set;
calculating a probability distribution value set of the training vector set by using a probability distribution function in the original label classification network;
and calculating a minimized target value of the probability distribution value set, and when the minimized target value is greater than a preset threshold value, adjusting internal parameters of the original label classification network until the minimized target value is less than or equal to the preset threshold value, so as to obtain the standard label classification network.
3. The method of generating a text presentation of claim 2 wherein the computing a set of probability distribution values for the set of training vectors using the probability distribution functions in the original label classification network comprises:
respectively constructing a first vector set and a second vector set according to the training vector set;
calculating a set of class prototypes for the first set of vectors;
and solving to obtain the probability distribution value set by taking the class prototype set and the second vector set as function parameters of the probability distribution function.
4. The method of generating a text presentation of claim 3, wherein said computing a set of class prototypes for the first set of vectors comprises:
calculating a class prototype set of the first vector set by adopting the following method:
Figure FDA0002710522950000021
wherein, ckClass prototypes representing class labels k, S (k) being the training directionVolume set, fθ(xi) A vector representation function, x, representing the ith training vector in the first vector set in the vector transformation networkiRepresenting the ith training vector, y, in the first set of vectorsiAnd representing a category label vector corresponding to the ith training vector in the first vector set.
5. The method of generating a text presentation according to claim 4, wherein said solving a set of probability distribution values for said class prototype set and said second vector set using said probability distribution function comprises:
solving a set of probability distribution values of the class prototype set and the second vector set by using the following probability distribution functions:
Figure FDA0002710522950000022
wherein x isjRepresenting the jth training vector, p, in said second set of vectorsθ(y=k|xj) A probability distribution value, f, representing that the jth training vector in the second vector set belongs to a class label kθ(xj) Representing a vector representation function of a jth training vector of said second set of vectors in said vector transformation network, ckA class prototype representing a class label k.
6. The method of generating a text presentation as recited in claim 1, wherein the generating a presentation template using the set of category presentations comprises:
performing attribute replacement operation and part-of-speech tagging operation on the category briefing set to obtain a primary briefing set;
and sequencing the sentences of the primary briefing set to obtain the briefing template.
7. The method for generating a text presentation according to any one of claims 1 to 6, wherein the preset ratio is 4: 6.
8. An apparatus for generating a text presentation, the apparatus comprising:
the briefing dividing module is used for receiving a briefing set, classifying the briefing set according to a preset proportion, and obtaining a first briefing set and a second briefing set;
the classification network training module is used for receiving a first class label set corresponding to the first brief report set, and training a pre-constructed original label classification network by using the first brief report set and the first class label set to obtain a standard label classification network;
a category label identification module, configured to perform a label classification operation on the second brief report set by using the standard label classification network to obtain a second category label set, and summarize the first category label set and the second category label set to obtain a category label set corresponding to the brief report set;
the briefing generating module is used for dividing the briefing set into different categories by using the category label set to obtain a category briefing set, generating a briefing template by using the category briefing set, receiving a briefing text to be generated, performing named entity identification on the briefing text to be generated to obtain an entity set, and performing combined operation on the entity set and the briefing template to obtain the briefing.
9. An electronic device, characterized in that the electronic device comprises:
at least one processor; and the number of the first and second groups,
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform a method of generating a text presentation according to any one of claims 1 to 7.
10. A computer-readable storage medium comprising a storage data area and a storage program area, wherein the storage data area stores created data, and the storage program area stores a computer program; wherein the computer program, when executed by a processor, implements a method of generating a text briefing as claimed in any one of claims 1 to 7.
CN202011054465.7A 2020-09-30 2020-09-30 Text briefing generation method and device, electronic equipment and readable storage medium Withdrawn CN112214594A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011054465.7A CN112214594A (en) 2020-09-30 2020-09-30 Text briefing generation method and device, electronic equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011054465.7A CN112214594A (en) 2020-09-30 2020-09-30 Text briefing generation method and device, electronic equipment and readable storage medium

Publications (1)

Publication Number Publication Date
CN112214594A true CN112214594A (en) 2021-01-12

Family

ID=74052156

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011054465.7A Withdrawn CN112214594A (en) 2020-09-30 2020-09-30 Text briefing generation method and device, electronic equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN112214594A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114022709A (en) * 2021-11-04 2022-02-08 内蒙古呼和浩特市立信电气技术有限责任公司 Control method and system based on stockpile layered three-dimensional real-time model

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114022709A (en) * 2021-11-04 2022-02-08 内蒙古呼和浩特市立信电气技术有限责任公司 Control method and system based on stockpile layered three-dimensional real-time model
CN114022709B (en) * 2021-11-04 2024-11-01 内蒙古呼和浩特市立信电气技术有限责任公司 Control method and system based on material pile layering three-dimensional real-time model

Similar Documents

Publication Publication Date Title
CN111639153A (en) Query method and device based on legal knowledge graph, electronic equipment and medium
CN113704429A (en) Semi-supervised learning-based intention identification method, device, equipment and medium
CN112883190A (en) Text classification method and device, electronic equipment and storage medium
CN115002200A (en) User portrait based message pushing method, device, equipment and storage medium
CN112559687A (en) Question identification and query method and device, electronic equipment and storage medium
CN113821622A (en) Answer retrieval method and device based on artificial intelligence, electronic equipment and medium
CN113378970A (en) Sentence similarity detection method and device, electronic equipment and storage medium
CN112988963A (en) User intention prediction method, device, equipment and medium based on multi-process node
CN112885423A (en) Disease label detection method and device, electronic equipment and storage medium
CN112528013A (en) Text abstract extraction method and device, electronic equipment and storage medium
CN114491047A (en) Multi-label text classification method and device, electronic equipment and storage medium
CN113887941A (en) Business process generation method and device, electronic equipment and medium
CN113807973A (en) Text error correction method and device, electronic equipment and computer readable storage medium
CN114595321A (en) Question marking method and device, electronic equipment and storage medium
CN114880449A (en) Reply generation method and device of intelligent question answering, electronic equipment and storage medium
CN114840684A (en) Map construction method, device and equipment based on medical entity and storage medium
CN114003704A (en) Method and device for creating designated tag guest group, electronic equipment and storage medium
CN113434650A (en) Question and answer pair expansion method and device, electronic equipment and readable storage medium
CN112214594A (en) Text briefing generation method and device, electronic equipment and readable storage medium
CN112347739A (en) Application rule analysis method and device, electronic equipment and storage medium
CN112214602A (en) Text classification method and device based on humor, electronic equipment and storage medium
CN116521867A (en) Text clustering method and device, electronic equipment and storage medium
CN115146064A (en) Intention recognition model optimization method, device, equipment and storage medium
CN115221274A (en) Text emotion classification method and device, electronic equipment and storage medium
CN114385815A (en) News screening method, device, equipment and storage medium based on business requirements

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20210112