CN112633087A - Automatic journaling method and device based on picture analysis for IBC system - Google Patents
Automatic journaling method and device based on picture analysis for IBC system Download PDFInfo
- Publication number
- CN112633087A CN112633087A CN202011430028.0A CN202011430028A CN112633087A CN 112633087 A CN112633087 A CN 112633087A CN 202011430028 A CN202011430028 A CN 202011430028A CN 112633087 A CN112633087 A CN 112633087A
- Authority
- CN
- China
- Prior art keywords
- video
- keywords
- recognition
- event
- script
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 40
- 239000000463 material Substances 0.000 claims abstract description 81
- 230000009471 action Effects 0.000 claims abstract description 14
- 230000008569 process Effects 0.000 claims abstract description 12
- 239000010454 slate Substances 0.000 claims description 2
- 241000053208 Porcellio laevis Species 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 12
- 238000004590 computer program Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 3
- PCHJSUWPFVWCPO-UHFFFAOYSA-N gold Chemical compound [Au] PCHJSUWPFVWCPO-UHFFFAOYSA-N 0.000 description 2
- 239000010931 gold Substances 0.000 description 2
- 229910052737 gold Inorganic materials 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
- G06V20/42—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items of sport video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/2228—Video assist systems used in motion picture production, e.g. video cameras connected to viewfinders of motion picture cameras or related video signal processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Television Signal Processing For Recording (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The embodiment of the invention provides an automatic script recording method and device based on picture analysis of an IBC system, in particular to a method and device for identifying and processing video and audio materials recorded by an event according to a script recording starting event to obtain the time positions of a target character and a keyword in the video and audio materials; traversing the video and audio material segments with the time position as the core and preset duration, and intelligently comparing the video and audio material segments based on the target character and the keywords to obtain an effective video picture comprising the target character and the action keywords; and the script recording system for driving the IBC system performs script recording processing on the effective video pictures based on the keywords. The script recording processing can be realized without manual operation in the process, so that the problem that a large amount of human resources are consumed in manual script recording is solved.
Description
Technical Field
The invention relates to the technical field of televisions, in particular to an automatic script recording method and device based on picture analysis of an IBC system.
Background
In the IBC system, the main task of the script is to record the scene information of the video and audio signals, and record the position of the picture by relative time, absolute time or position time code, that is, the detailed information of each scene shot in the scene, such as scene switching, shooting method, character action, caption, transition scene, etc., and the detailed data are recorded in detail and precisely, and correspond to the recorded material one by one, so as to provide accurate retrieval data and material for the later editing.
The inventor of the present application has found that in current practice of events, the time-keeping system is implemented by manual operation, that is, the positions of the shots in the sports game are recorded in real time by manual means and correspond to the recorded materials, so that the operator needs to be familiar with the match rules of various game items and needs to be trained for a long time and operated on the spot, thereby consuming a large amount of human resources.
Disclosure of Invention
In order to solve the above problems, the present invention provides an automatic journaling method and apparatus based on picture analysis for an IBC system, so as to solve the problem that a large amount of human resources are consumed for manual journaling.
In view of this, the present invention discloses an automatic journaling method based on picture analysis for an IBC system, which includes the steps of:
responding to a preset script starting event, and identifying the video and audio material recorded by the event based on a pre-trained identification model to obtain the time positions of the target character and the keywords in the video and audio material;
traversing video and audio material segments with preset duration and taking the time position as a core, and intelligently comparing the video and audio material segments based on the target character and the keyword to obtain an effective video picture comprising the target character and the keyword;
and a script recording module for driving the IBC system carries out script recording processing on the effective video pictures based on the keywords.
Optionally, the script start time is an input event of a script start request of a user or a real-time preview event of the video and audio material.
Optionally, the recognition model is a comprehensive recognition model based on image recognition and auxiliary recognition, where:
the image recognition comprises face recognition, action recognition and/or object recognition;
the auxiliary recognition comprises voice recognition and/or subtitle recognition.
Optionally, the keywords include a script event keyword, a lens level description keyword, and an event scene description keyword.
Optionally, the method further comprises the steps of:
and performing model training by using the keywords and the labeled video and audio materials to obtain the identification model.
There is also provided a picture analysis based automatic journaling apparatus of an IBC system, the automatic journaling apparatus comprising:
the material identification module is configured to respond to a preset script starting event and identify and process video and audio materials recorded by the event based on a pre-trained identification model to obtain the time positions of target characters and keywords in the video and audio materials;
the comparison processing module is configured to traverse video and audio material segments with preset duration and taking the time position as a core, and intelligently compare the video and audio material segments based on the target character and the keywords to obtain an effective video picture comprising the target character and the keywords;
a script execution module configured to drive a script module of the IBC system to script the active video frames based on the keywords.
Optionally, the script start time is an input event of a script start request of a user or a real-time preview event of the video and audio material.
Optionally, the recognition model is a comprehensive recognition model based on image recognition and auxiliary recognition, where:
the image recognition comprises face recognition, action recognition and/or object recognition;
the auxiliary recognition comprises voice recognition and/or subtitle recognition.
Optionally, the keywords include a script event keyword, a lens level description keyword, and an event scene description keyword.
Optionally, the method further comprises the steps of:
and performing model training by using the keywords and the labeled video and audio materials to obtain the identification model.
The technical scheme can be seen that the invention provides an automatic script recording method and device based on picture analysis for an IBC system, and specifically relates to a method and device for identifying and processing video and audio materials recorded by an event according to a script recording starting event to obtain time positions of target characters and keywords in the video and audio materials; traversing the video and audio material segments with the time position as the core and preset duration, and intelligently comparing the video and audio material segments based on the target character and the keywords to obtain an effective video picture comprising the target character and the action keywords; and the script recording system for driving the IBC system performs script recording processing on the effective video pictures based on the keywords. The script recording processing can be realized without manual operation in the process, so that the problem that a large amount of human resources are consumed in manual script recording is solved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
Fig. 1 is a flowchart of an automatic picture-recognition-based journaling method of an IBC system according to an embodiment of the present disclosure;
FIG. 2 is a diagram of time key words of illustrative sentences in the present application;
FIG. 3 is a schematic illustration of a job interface in the present application;
FIG. 4 is a flowchart of an alternative embodiment of an IBC system for automatic picture-recognition-based journaling;
FIG. 5 is a block diagram of an automatic picture-recognition-based journaling apparatus of an IBC system according to an embodiment of the present application;
fig. 6 is a block diagram of an automatic picture recognition-based journaling apparatus of another IBC system according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example one
Fig. 1 is a flowchart of an automatic picture-recognition-based journaling method of an IBC system according to an embodiment of the present disclosure.
The IBC system provides an IBCloud media data sharing system, which includes a receiving module, a script module, and a content search module.
The recording module is used for recording video and audio signals of events corresponding to the script characters, and specifically, recording the event signals by taking a recording task as a unit and adopting a 'simultaneous folding and coding' mode to form video and audio materials and material metadata information files; the script recording module is used for creating script recording tasks and selecting the script recording tasks to carry out script recording operation; the content retrieval module is used for retrieving the event figure materials, the script recording event pictures and the material segments;
referring to fig. 1, the automatic journaling method specifically includes the following steps:
and S1, identifying the video and audio materials according to the script start events.
The script start event here may be an input event of the user when inputting the corresponding script start request, or may be a real-time preview event generated by real-time previewing the corresponding video and audio material on the terminal of the IBC. And on the basis of the script starting event, identifying the video and audio material recorded in the event according to a pre-trained identification model, and identifying and obtaining the time positions of the target character and the keywords in the video and audio material. The target character may be a target athlete or a target sports team, etc.
The keywords include competition rules, specifications and etiquette of sports events, and event keywords of memorial events are uniformly defined, and also can include description keywords of lens layers such as panorama, overhead shot, slow shot and the like, and description keywords of event scenes such as awarding ceremony, flag raising, gold medal and the like. The pre-enumerated user-defined keywords are shot pictures with high estimated follow-up user retrieval query access rate, are effective or meaningful interesting picture contents, and the script needs to label key frames on the shot pictures and extract effective picture segments according to the key frames for users to preview or use.
An exemplary script time keyword is shown in FIG. 2, taking an ice hockey project as an example. Besides self-defining the script recording event keyword group, the name information of the players and the coach who participate in the recorded event can be pre-recorded, so that the script recording shot can be conveniently and clearly marked. Taking the ice hockey project as an example, the operation interface of a marquee of an ice hockey event is shown in fig. 3.
The recognition model in the scheme is a comprehensive recognition model based on image recognition and auxiliary recognition, and the image recognition means comprises face recognition, action recognition and object recognition, and can also adopt partial means; the auxiliary recognition means comprises voice recognition or subtitle recognition or both.
And S2, determining an effective video picture based on the target person and the keywords.
After the time positions of the target characters and the keywords in the video and audio material are obtained, video and audio material segments with preset time length including the time positions are intercepted, for example, the video and audio material segments with the total length of 20 seconds are respectively extended to the front and the back of the time positions by 10S. And then intelligently comparing the segment based on the target characters and the keywords to obtain an effective video picture of the segment comprising the target characters and the keywords.
And S3, driving the script system to script the effective video picture.
On the basis of obtaining the effective video picture, the IBC system script recording system is driven to script recording the effective video picture, namely, the effective frame of the effective video picture is marked by using a corresponding mark, thereby realizing automatic script recording of the video and audio materials.
In addition, a complete bookmarked event record contains the recorded material information, relative time and position time codes, bookmarked event keywords (there may be multiple), athlete information (there may be multiple), etc.
It can be seen from the foregoing technical solutions that, the present embodiment provides an automatic script recording method based on picture analysis for an IBC system, specifically, a video and audio material recorded in an event is identified according to a script recording start event to obtain time positions of a target character and a keyword in the video and audio material; traversing the video and audio material segments with the time position as the core and preset duration, and intelligently comparing the video and audio material segments based on the target character and the keywords to obtain an effective video picture comprising the target character and the action keywords; and the script recording system for driving the IBC system performs script recording processing on the effective video pictures based on the keywords. The script recording processing can be realized without manual operation in the process, so that the problem that a large amount of human resources are consumed in manual script recording is solved.
In addition, in a specific implementation manner of this embodiment, the following steps are further included, as shown in fig. 4.
And S01, performing model training by using the keywords and the video and audio materials.
The above recognition model is obtained through training, the keywords are the same as the keywords described above, and the description is omitted here, the training process here is realized based on the artificially labeled video and audio material, i.e., the artificially labeled video and audio material is the video and audio material after the script-keeping processing is performed in an artificial way.
And the video and audio materials subjected to automatic script-keeping processing can be used as extension samples to further train the recognition model, so that a more accurate recognition model is obtained.
Example two
Fig. 5 is a block diagram of an automatic picture recognition-based slate device of an IBC system according to an embodiment of the present disclosure.
The IBC system provides an IBCloud media data sharing system, which includes a receiving module, a script module, and a content search module.
The recording module is used for recording video and audio signals of events corresponding to the script characters, and specifically, recording the event signals by taking a recording task as a unit and adopting a 'simultaneous folding and coding' mode to form video and audio materials and material metadata information files; the script recording module is used for creating script recording tasks and selecting the script recording tasks to carry out script recording operation; the content retrieval module is used for retrieving the event figure materials, the script recording event pictures and the material segments;
referring to fig. 5, the automatic journaling apparatus specifically includes a material recognition module 10, a comparison processing module 20, and a journaling execution module 30.
The material identification module is used for identifying the video and audio materials according to the script-keeping starting events.
The script start event here may be an input event of the user when inputting the corresponding script start request, or may be a real-time preview event generated by real-time previewing the corresponding video and audio material on the terminal of the IBC. And on the basis of the script starting event, identifying the video and audio material recorded in the event according to a pre-trained identification model, and identifying and obtaining the time positions of the target character and the keywords in the video and audio material. The target character may be a target athlete or a target sports team, etc.
The keywords include competition rules, specifications and etiquette of sports events, and event keywords of memorial events are uniformly defined, and also can include description keywords of lens layers such as panorama, overhead shot, slow shot and the like, and description keywords of event scenes such as awarding ceremony, flag raising, gold medal and the like. The pre-enumerated user-defined keywords are shot pictures with high estimated follow-up user retrieval query access rate, are effective or meaningful interesting picture contents, and the script needs to label key frames on the shot pictures and extract effective picture segments according to the key frames for users to preview or use.
An exemplary script time keyword is shown in FIG. 2, taking an ice hockey project as an example. Besides self-defining the script recording event keyword group, the name information of the players and the coach who participate in the recorded event can be pre-recorded, so that the script recording shot can be conveniently and clearly marked. Taking the ice hockey project as an example, the operation interface of a marquee of an ice hockey event is shown in fig. 3.
The recognition model in the scheme is a comprehensive recognition model based on image recognition and auxiliary recognition, and the image recognition means comprises face recognition, action recognition and object recognition, and can also adopt partial means; the auxiliary recognition means comprises voice recognition or subtitle recognition or both.
The comparison processing module is used for determining an effective video picture based on the target person and the keywords.
After the time positions of the target characters and the keywords in the video and audio material are obtained, video and audio material segments with preset time length including the time positions are intercepted, for example, the video and audio material segments with the total length of 20 seconds are respectively extended to the front and the back of the time positions by 10S. And then intelligently comparing the segment based on the target characters and the keywords to obtain an effective video picture of the segment comprising the target characters and the keywords.
The script recording execution module is used for driving the script recording system to perform script recording processing on the effective video pictures.
On the basis of obtaining the effective video picture, the IBC system script recording system is driven to script recording the effective video picture, namely, the effective frame of the effective video picture is marked by using a corresponding mark, thereby realizing automatic script recording of the video and audio materials.
In addition, a complete bookmarked event record contains the recorded material information, relative time and position time codes, bookmarked event keywords (there may be multiple), athlete information (there may be multiple), etc.
As can be seen from the foregoing technical solutions, the embodiment provides an automatic script device based on picture analysis for an IBC system, which is specifically configured to perform recognition processing on a video/audio material recorded in an event according to a script start event, so as to obtain time positions of a target character and a keyword in the video/audio material; traversing the video and audio material segments with the time position as the core and preset duration, and intelligently comparing the video and audio material segments based on the target character and the keywords to obtain an effective video picture comprising the target character and the action keywords; and the script recording system for driving the IBC system performs script recording processing on the effective video pictures based on the keywords. The script recording processing can be realized without manual operation in the process, so that the problem that a large amount of human resources are consumed in manual script recording is solved.
In addition, in a specific implementation manner of this embodiment, the apparatus further includes a model training module 40, as shown in fig. 6.
And the model training module is used for performing model training by using the keywords and the video and audio materials.
The above recognition model is obtained through training, the keywords are the same as the keywords described above, and the description is omitted here, the training process here is realized based on the artificially labeled video and audio material, i.e., the artificially labeled video and audio material is the video and audio material after the script-keeping processing is performed in an artificial way.
And the video and audio materials subjected to automatic script-keeping processing can be used as extension samples to further train the recognition model, so that a more accurate recognition model is obtained.
The embodiments in the present specification are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, apparatus, or computer program product. Accordingly, embodiments of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, embodiments of the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
Embodiments of the present invention are described with reference to flowchart illustrations and/or block diagrams of methods, terminal devices (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing terminal to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing terminal, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing terminal to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing terminal to cause a series of operational steps to be performed on the computer or other programmable terminal to produce a computer implemented process such that the instructions which execute on the computer or other programmable terminal provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While preferred embodiments of the present invention have been described, additional variations and modifications of these embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all such alterations and modifications as fall within the scope of the embodiments of the invention.
Finally, it should also be noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or terminal that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or terminal. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or terminal that comprises the element.
The technical solutions provided by the present invention are described in detail above, and the principle and the implementation of the present invention are explained in this document by applying specific examples, and the descriptions of the above examples are only used to help understanding the method and the core idea of the present invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.
Claims (10)
1. An automatic journaling method based on picture analysis for IBC system, characterized in that said automatic journaling method comprises the steps of:
responding to a preset script starting event, and identifying the video and audio material recorded by the event based on a pre-trained identification model to obtain the time positions of the target character and the keywords in the video and audio material;
traversing video and audio material segments with preset duration and taking the time position as a core, and intelligently comparing the video and audio material segments based on the target character and the keyword to obtain an effective video picture comprising the target character and the keyword;
and a script recording module for driving the IBC system carries out script recording processing on the effective video pictures based on the keywords.
2. The automatic journaling method according to claim 1, wherein said journaling start time is an input event of a journaling start request of a user or a live preview event of said audio and video material.
3. The automatic clapping method of claim 1, wherein the recognition model is a comprehensive recognition model based on image recognition and assisted recognition, wherein:
the image recognition comprises face recognition, action recognition and/or object recognition;
the auxiliary recognition comprises voice recognition and/or subtitle recognition.
4. The automatic journaling method according to claim 1, wherein said keywords include a journaling event keyword, a shot level description keyword and an event scene description keyword.
5. The automatic clapping method of any one of claims 1 to 4, further comprising the steps of:
and performing model training by using the keywords and the labeled video and audio materials to obtain the identification model.
6. An automatic picture-analysis-based journaling apparatus of an IBC system, the automatic journaling apparatus comprising:
the material identification module is configured to respond to a preset script starting event and identify and process video and audio materials recorded by the event based on a pre-trained identification model to obtain the time positions of target characters and keywords in the video and audio materials;
the comparison processing module is configured to traverse video and audio material segments with preset duration and taking the time position as a core, and intelligently compare the video and audio material segments based on the target character and the keywords to obtain an effective video picture comprising the target character and the keywords;
a script execution module configured to drive a script module of the IBC system to script the active video frames based on the keywords.
7. The automatic ticker apparatus of claim 6 wherein said ticker start time is an input event of a user's ticker start request or a live preview event of said audio and video material.
8. The automatic clap slater of claim 6 wherein said identification model is a comprehensive identification model based on image recognition and assisted recognition wherein:
the image recognition comprises face recognition, action recognition and/or object recognition;
the auxiliary recognition comprises voice recognition and/or subtitle recognition.
9. The automatic memorandum device according to claim 6, wherein said keywords comprise memorandum event keywords, shot level description keywords, and event scenario description keywords.
10. The automatic clap slate device according to any one of claims 6 to 9, further comprising the steps of:
and performing model training by using the keywords and the labeled video and audio materials to obtain the identification model.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011430028.0A CN112633087A (en) | 2020-12-09 | 2020-12-09 | Automatic journaling method and device based on picture analysis for IBC system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011430028.0A CN112633087A (en) | 2020-12-09 | 2020-12-09 | Automatic journaling method and device based on picture analysis for IBC system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112633087A true CN112633087A (en) | 2021-04-09 |
Family
ID=75309474
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011430028.0A Pending CN112633087A (en) | 2020-12-09 | 2020-12-09 | Automatic journaling method and device based on picture analysis for IBC system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112633087A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117132925A (en) * | 2023-10-26 | 2023-11-28 | 成都索贝数码科技股份有限公司 | Intelligent stadium method and device for sports event |
Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010047266A1 (en) * | 1998-01-16 | 2001-11-29 | Peter Fasciano | Apparatus and method using speech recognition and scripts to capture author and playback synchronized audio and video |
CN101296322A (en) * | 2007-04-27 | 2008-10-29 | 新奥特硅谷视频技术有限责任公司 | Sports event venue recording system |
CN101472082A (en) * | 2007-12-25 | 2009-07-01 | 新奥特(北京)视频技术有限公司 | Log keeping system and method |
CN102196159A (en) * | 2010-03-12 | 2011-09-21 | 新奥特(北京)视频技术有限公司 | Competition material sharing method and system for international broadcasting center (IBC) system |
CN102196189A (en) * | 2010-03-12 | 2011-09-21 | 新奥特(北京)视频技术有限公司 | Method and system for retrieving highlights from large-scale competition international broadcasting center (IBC) system |
CN102196231A (en) * | 2010-03-12 | 2011-09-21 | 新奥特(北京)视频技术有限公司 | Method and system for realizing real-time publication of competition scenes |
CN102393854A (en) * | 2011-09-09 | 2012-03-28 | 杭州海康威视数字技术股份有限公司 | Method and device obtaining audio/video data |
CN102572293A (en) * | 2010-12-16 | 2012-07-11 | 新奥特(北京)视频技术有限公司 | Field recording-based retrieval system |
CN102572294A (en) * | 2010-12-16 | 2012-07-11 | 新奥特(北京)视频技术有限公司 | Field recoding system with ranking function |
KR20130008285A (en) * | 2011-07-12 | 2013-01-22 | (주)마인스코퍼레이션 | Method and apparatus for providing moving pictures by producing in real time |
US20140327779A1 (en) * | 2013-05-01 | 2014-11-06 | Nokia Corporation | Method and apparatus for providing crowdsourced video |
CN106851407A (en) * | 2017-01-24 | 2017-06-13 | 维沃移动通信有限公司 | A kind of control method and terminal of video playback progress |
CN107888843A (en) * | 2017-10-13 | 2018-04-06 | 深圳市迅雷网络技术有限公司 | Sound mixing method, device, storage medium and the terminal device of user's original content |
CN109587552A (en) * | 2018-11-26 | 2019-04-05 | Oppo广东移动通信有限公司 | Video personage sound effect treatment method, device, mobile terminal and storage medium |
US20190191205A1 (en) * | 2017-12-19 | 2019-06-20 | At&T Intellectual Property I, L.P. | Video system with second screen interaction |
CN110012348A (en) * | 2019-06-04 | 2019-07-12 | 成都索贝数码科技股份有限公司 | A kind of automatic collection of choice specimens system and method for race program |
KR20190098775A (en) * | 2018-01-12 | 2019-08-23 | 상명대학교산학협력단 | Artificial intelligence deep-learning based video object recognition system and method |
CN110188241A (en) * | 2019-06-04 | 2019-08-30 | 成都索贝数码科技股份有限公司 | A kind of race intelligence manufacturing system and production method |
-
2020
- 2020-12-09 CN CN202011430028.0A patent/CN112633087A/en active Pending
Patent Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010047266A1 (en) * | 1998-01-16 | 2001-11-29 | Peter Fasciano | Apparatus and method using speech recognition and scripts to capture author and playback synchronized audio and video |
CN101296322A (en) * | 2007-04-27 | 2008-10-29 | 新奥特硅谷视频技术有限责任公司 | Sports event venue recording system |
CN101472082A (en) * | 2007-12-25 | 2009-07-01 | 新奥特(北京)视频技术有限公司 | Log keeping system and method |
CN102196159A (en) * | 2010-03-12 | 2011-09-21 | 新奥特(北京)视频技术有限公司 | Competition material sharing method and system for international broadcasting center (IBC) system |
CN102196189A (en) * | 2010-03-12 | 2011-09-21 | 新奥特(北京)视频技术有限公司 | Method and system for retrieving highlights from large-scale competition international broadcasting center (IBC) system |
CN102196231A (en) * | 2010-03-12 | 2011-09-21 | 新奥特(北京)视频技术有限公司 | Method and system for realizing real-time publication of competition scenes |
CN102572294A (en) * | 2010-12-16 | 2012-07-11 | 新奥特(北京)视频技术有限公司 | Field recoding system with ranking function |
CN102572293A (en) * | 2010-12-16 | 2012-07-11 | 新奥特(北京)视频技术有限公司 | Field recording-based retrieval system |
KR20130008285A (en) * | 2011-07-12 | 2013-01-22 | (주)마인스코퍼레이션 | Method and apparatus for providing moving pictures by producing in real time |
CN102393854A (en) * | 2011-09-09 | 2012-03-28 | 杭州海康威视数字技术股份有限公司 | Method and device obtaining audio/video data |
US20140327779A1 (en) * | 2013-05-01 | 2014-11-06 | Nokia Corporation | Method and apparatus for providing crowdsourced video |
CN106851407A (en) * | 2017-01-24 | 2017-06-13 | 维沃移动通信有限公司 | A kind of control method and terminal of video playback progress |
CN107888843A (en) * | 2017-10-13 | 2018-04-06 | 深圳市迅雷网络技术有限公司 | Sound mixing method, device, storage medium and the terminal device of user's original content |
US20190191205A1 (en) * | 2017-12-19 | 2019-06-20 | At&T Intellectual Property I, L.P. | Video system with second screen interaction |
KR20190098775A (en) * | 2018-01-12 | 2019-08-23 | 상명대학교산학협력단 | Artificial intelligence deep-learning based video object recognition system and method |
CN109587552A (en) * | 2018-11-26 | 2019-04-05 | Oppo广东移动通信有限公司 | Video personage sound effect treatment method, device, mobile terminal and storage medium |
CN110012348A (en) * | 2019-06-04 | 2019-07-12 | 成都索贝数码科技股份有限公司 | A kind of automatic collection of choice specimens system and method for race program |
CN110188241A (en) * | 2019-06-04 | 2019-08-30 | 成都索贝数码科技股份有限公司 | A kind of race intelligence manufacturing system and production method |
Non-Patent Citations (2)
Title |
---|
GURVINDER SINGH等: "Video frame and region duplication forgery detection based on correlation coefficient and coefficient of variation", MULTIMEDIA TOOLS AND APPLICATIONS(SPRINGER LINK), vol. 78, 26 September 2018 (2018-09-26), pages 11527 - 11562, XP036779905, DOI: 10.1007/s11042-018-6585-1 * |
董璐;: "十运会IBC赛事共享系统的实时场记", 广播与电视技术, no. 04, 30 April 2006 (2006-04-30), pages 58 - 61 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117132925A (en) * | 2023-10-26 | 2023-11-28 | 成都索贝数码科技股份有限公司 | Intelligent stadium method and device for sports event |
CN117132925B (en) * | 2023-10-26 | 2024-02-06 | 成都索贝数码科技股份有限公司 | Intelligent stadium method and device for sports event |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021104242A1 (en) | Video processing method, electronic device, and storage medium | |
US7873258B2 (en) | Method and apparatus for reviewing video | |
CN107707931B (en) | Method and device for generating interpretation data according to video data, method and device for synthesizing data and electronic equipment | |
US6799180B1 (en) | Method of processing signals and apparatus for signal processing | |
RU2440606C2 (en) | Method and apparatus for automatic generation of summary of plurality of images | |
JP4250301B2 (en) | Method and system for editing video sequences | |
US8121462B2 (en) | Video edition device and method | |
CN110121116A (en) | Video generation method and device | |
JP2001028722A (en) | Moving picture management device and moving picture management system | |
KR20090093904A (en) | Apparatus and method for scene variation robust multimedia image analysis, and system for multimedia editing based on objects | |
CN110677735A (en) | Video positioning method and device | |
CN101112090B (en) | Video content reproduction supporting method, video content reproduction supporting system, and information delivery server | |
CN112633087A (en) | Automatic journaling method and device based on picture analysis for IBC system | |
JP3934780B2 (en) | Broadcast program management apparatus, broadcast program management method, and recording medium recording broadcast program management processing program | |
JP2001285780A (en) | Image editing method and device and storage medium therefor | |
WO2024109813A1 (en) | Video processing method and apparatus | |
CN109688430A (en) | A kind of court trial file playback method, system and storage medium | |
CN113992973A (en) | Video abstract generation method and device, electronic equipment and storage medium | |
CN117319765A (en) | Video processing method, device, computing equipment and computer storage medium | |
KR101783872B1 (en) | Video Search System and Method thereof | |
JP2005039354A (en) | Metadata input method and editing system | |
CN101833978A (en) | Character signal-triggered court trial video real-time indexing method | |
Cobârzan et al. | How do users search with basic html5 video players? | |
TWI823018B (en) | Method of video graph developing | |
Bhat et al. | Cricket Video Summarization Using Deep Learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |