[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN1245697C - Method of proceeding video frequency searching through video frequency segment - Google Patents

Method of proceeding video frequency searching through video frequency segment Download PDF

Info

Publication number
CN1245697C
CN1245697C CN 03149998 CN03149998A CN1245697C CN 1245697 C CN1245697 C CN 1245697C CN 03149998 CN03149998 CN 03149998 CN 03149998 A CN03149998 A CN 03149998A CN 1245697 C CN1245697 C CN 1245697C
Authority
CN
China
Prior art keywords
similar
fragment
camera lens
video
similarity
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN 03149998
Other languages
Chinese (zh)
Other versions
CN1514644A (en
Inventor
彭宇新
董庆杰
郭宗明
肖建国
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BEIDA FANGZHENG TECHN INST Co Ltd BEIJING
Inst Of Computer Science & Technology Peking University
Original Assignee
BEIDA FANGZHENG TECHN INST Co Ltd BEIJING
Inst Of Computer Science & Technology Peking University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BEIDA FANGZHENG TECHN INST Co Ltd BEIJING, Inst Of Computer Science & Technology Peking University filed Critical BEIDA FANGZHENG TECHN INST Co Ltd BEIJING
Priority to CN 03149998 priority Critical patent/CN1245697C/en
Publication of CN1514644A publication Critical patent/CN1514644A/en
Application granted granted Critical
Publication of CN1245697C publication Critical patent/CN1245697C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The present invention relates to a novel method for retrieving videos by video segments, which belongs to the technical field of video retrieval. The existing video segment retrieving method based on contents has the defects of low retrieval precision and low retrieval speed. In view of the defects existing in the prior art, the present invention solves the problem of two segment retrieval difficulty in automatically acquiring similar segments and evaluating the similarity of the two fragments while the optimally matched frame of the graph theory gets down. A weighting bipartite graph model for retrieving segments is established; a similar segment is primarily obtained by inspecting the continuity of similar lenses; the maximum weight match of the segments and the enquiry segments is calculated by using an optimally matched Kuhn_Munkres method; the match is used for judging whether the two segments are similar and determining the boundary of the similar segments; the maximum weight is used for representing the similarity of the two similar fragments. The experimental result indicates that compared with the existing method, the present invention can obtain a high precision ratio and a high recall ratio; meanwhile, the present invention has high retrieval speed.

Description

A kind of method of carrying out video frequency searching by video segment
Technical field
The invention belongs to video search technique area, be specifically related to a kind of method of carrying out video frequency searching by video segment.
Background technology
With image and video information process is that the intelligent information processing technology of core is that people make full use of multimedia messages strong instrument is provided.But, appearance along with extensive image and video database, the data volume that infosystem must be handled increases progressively significantly, how these databases are carried out effective index and become one of bottleneck problem in the multimedia messages process field with retrieval, content-based image and video retrieval method have been opened up new approach for people's head it off.Video clip retrieval is based on the main mode of the video frequency searching of content, and it is meant a given query fragment, finds all fragments similar to it in video library.
Existing fragment search method can be divided into two classes: one, as document " A Framework forMeasuring Video Similarity and Its Application to Video Query byExample " [Y.P.Tan, S.R.Kulkarni, and P.J.Ramadge, IEEEInternational Conference on Image Processing, Vol.2, pp.106-110,1999] described, video segment is divided into the two-layer consideration of fragment-frame, and the similarity utilization of fragment is formed the similarity of its frame and is directly measured.The shortcoming of these class methods is to limit similar fragment must observe same time sequencing, and the practical video program is not observed this constraint, because the result of later stage compilation makes similar fragment may have different camera lens orders fully, difference editor as same advertisement, this comparison based on every frame simultaneously also makes retrieval rate slow.Two, the prior art the most approaching with the present invention is that (author is L.Chen to the document of delivering at IEEE International Conference on Multimedia and Expo calendar year 2001 " A Match and Tiling Approach to Content-based VideoRetrieval ", and T.S.Chua, page number 417-420), this documents discloses a class fragment search method, this method is divided into fragment-camera lens-three layers of consideration of frame to video segment, it comprises such several steps: (1) uses MRA (Temporal Multi-ResolutionAnalysis) method detector lens border earlier, to each frame of each camera lens, carry out color coding and texture coding then.Color coding adopts the average μ and the variances sigma coding of Y component, and texture adopts FRACTAL DIMENSION feature (Fractal Dimension, FD) coding; (2) suppose the similar frame of two camera lens inside, similar according to the time sequencing correspondence, therefore calculate the maximum length sequence of two similar frames of camera lens, the similarity of final two camera lenses is expressed as the linear combination of above-mentioned 3 features, determines similar threshold value σ L, judge whether two camera lenses are similar; (3) on this basis, use the way of moving window (Sliding Window), finally find the fragment similar to query fragment.This method can accurately be retrieved and similarity retrieval simultaneously, but its problem is: (1) has only considered the quantity of two similar camera lenses of fragment, and the camera lens of not considering multi-to-multi similar (granularity) is to the influence of overall similarity, therefore, even all camera lenses of fragment Y are only similar with the camera lens of fragment X, it is similar to X that Y also can be considered to; (2) hypothesis of Ti Chuing and being false, i.e. the similar frame of two camera lens inside may not be similar according to the time sequencing correspondence; (3) similarity of camera lens is to judge according to two camera lens the longest similar frame sequences, this comparison based on every frame, and the retrieval rate of fragment is slow.
Summary of the invention
At the existing defective of existing video fragment searching method, the objective of the invention is to propose a kind of method of carrying out video frequency searching by video segment, this method can improve precision ratio, recall ratio and the retrieval rate of content-based video clip retrieval on the basis of existing technology greatly, thereby brings into play the huge effect of video clip retrieval technology in current network information society more fully.
The object of the present invention is achieved like this: a kind ofly carry out the method for video frequency searching by video segment, may further comprise the steps:
(1) at first carrying out shot boundary and detect, is the Video Segmentation among query fragment X and the video library Y camera lens;
(2) represent the content of each camera lens then with the first frame of each camera lens; Similar value Similarity (the x of two camera lenses i, y j) be the first frame result calculated according to two camera lenses, wherein x i, y jRepresent two camera lenses, as Similarity (x i, y j)>T s, just think two camera lens x iAnd y jSimilar, T in the formula sFor the similar threshold values of camera lens,, retrieve among the video database Y camera lens x with query fragment X according to this formula iAll similar camera lens y j
(3) set up the cum rights bipartite graph model that fragment is retrieved,, tentatively be partitioned into the fragment Y similar to query fragment X by investigating the continuity of similar camera lens k
(4) because Y kA plurality of fragment Ys similar may have been comprised to X X, in order further to obtain Y X, and can estimate Y simultaneously XWith the similarity of X, adopt the Optimum Matching method to calculate k bipartite graph G k={ X, Y k, E kMaximum weight matching M, from M, obtain Y X, represent X and Y with this authority simultaneously XSimilarity.
Specifically, when carrying out video clip retrieval, in step (2), utilize the fragment length constraint setting threshold T of Optimum Matching s, remove dissimilar camera lens, threshold value T sBe 0~1, T sBe preferably 0.71.
Again specifically, when carrying out video clip retrieval, adopt following method tentatively to be partitioned into fragment similar among the video library Y: with camera lens y similar among the Y to X to query fragment X jBy rank order from small to large, investigate y then jContinuity, if | y J+1-y j|>2, j=1,2 ..., m-1 then tentatively obtains a fragment Y similar to X k={ y i, y I+1..., y j, i, j ∈ [1, m].
Further, the constraint condition of use fragment length as previously described filters out T sThe dissimilar camera lens that brings, realize by twice constraint:
(1), adopts the preceding length constraint of Optimum Matching method: for fear of unnecessary calculating,, fall, if promptly with regard to direct filtration if the fragment length of Yk does not reach requirement Then ignore, wherein, Y k={ y i, y I+1..., y j, i, j ∈ [1, m], n is the length of query fragment X;
(2), the length constraint after the employing Optimum Matching method: all ω that get the related Optimum Matching M of X Ij>0 y, ordering is Y '={ y from small to large α, y β... y γ, α, beta, gamma ∈ [i, j], i, j ∈ [1, m], if
Figure C0314999800072
Then ignore.
Again specifically, when carrying out video clip retrieval, because Y kA plurality of fragment Ys similar may have been comprised to X X, in order further to obtain Y X, simultaneously, estimate Y XWith the similarity of X, the present invention can calculate k bipartite graph G with the Kuhn_Munkres method k={ X, Y k, E kMaximum weight matching M, from M, obtain Y X, represent X and Y with this authority simultaneously XSimilarity, the Kuhn_Munkres method of specifically calculating Optimum Matching is as follows:
(1) provides initial label l ( x i ) = max j ω ij , l (y j)=0, i, j=1,2..., t, t=max (n, | Y k|), n is the length of query fragment X, Y k={ y i, y i+ 1 ..., y j, i, j ∈ [1, m];
(2) obtain limit collection E l={ (x i, y j) | l (x i)+l (y j)=ω Ij, G l=(X, Y k, E l) and G lIn one the coupling M;
(3) as all nodes of the saturated X of M, then M promptly is the Optimum Matching of G, calculates and finishes, otherwise carry out next step;
(4) in X, look for a M unsaturation point x 0, make A ← { x 0, B ← φ, A, B are two set;
(5) if N Gl(A)=and B, then changeed for (9) step, otherwise carry out next step, wherein, N Gl(A) Y k, be with A in the node set of node adjacency;
(6) look for a node y ∈ N Gl(A)-B;
(7) if y is the M saturation point, then find out the match point z of y, make A ← A ∪ z}, { y} changeed for (5) step, otherwise carries out next step B ← B ∪;
(8) there is one from x 0But the augmenting path P to y makes M ← M E (P), changes for (3) step;
(9) be calculated as follows the α value: a = min x i ∈ A y j ∉ N G l ( A ) { l ( x i ) + l ( y j ) - ω ij } , Revise label:
Figure C0314999800075
Ask E according to l ' L 'And G L '
(10) l ← l ', G l← G L ', changeed for (6) step;
It is function Kuhn_Munkres (X, a Y that the present invention defines the Kuhn_Munkres method k), so, obtain similar fragment Y XWith evaluation Y XBe described below with the method for X similarity:
(1) calls Kuhn_Munkres (X, Y k), obtain G k={ X, Y k, E kMaximum weight matching M and authority ω;
(2) get all ω of the related M of X Ij>0 y, ordering is Y '={ y from small to large α, y β..., y γ, α, beta, gamma ∈ [i, j], in this set, y α, y βPossibility is also discontinuous, i.e. y β-y α>1, according to the definition and the Y of front video segment kMiddle camera lens is all similar in appearance to X, and the present invention gets y αWith y γBetween all camera lenses constitute continuous fragment Y X={ y α, y α+1..., y γ, alpha, gamma ∈ [i, j];
(3) if | Y ' |<L, then Y XIgnore, calculate and finish, otherwise carry out next step;
(4) calculate X and Y XSimilarity Similarity ( X , Y X ) = ω n ;
(5) if | X| 〉=| Y k|, calculate and finish, otherwise carry out next step;
(6) for | X|<| Y k|, at Y kGet Y XAfter, see Y kCan remaining fragment continue to use the Kuhn_Munkres method, if promptly (α-i) 〉=L, Y k← { y i, y I+1, y α-1, change (1); (if j-γ) 〉=L, Y k← { y γ+1, y γ+2, y j, change (1);
Get
Figure C0314999800082
In above-mentioned (4) step, the present invention uses Similarity ( X , Y X ) = ω n With Similarity (X, Y X) normalize between [0,1], this value is big more, shows that two fragments are similar more, relatively more existing method, and this formula has succinctly been estimated two maximum similarities that fragment can reach effectively.
Effect of the present invention is: adopt video fragment searching method of the present invention, can obtain higher precision ratio and recall ratio, also obtained retrieval rate faster simultaneously.
Why the present invention has so significant technique effect, its reason is: as described in the previous technique content, the present invention has taken into full account vision and granulometry that the fragment similarity is judged, proposes to obtain with the Kuhn_Munkres method optimization of graph theory Optimum Matching the similarity of similar fragment and two fragments of evaluation.Because video segment is made up of a set of shots, therefore, retrieving is divided into searching lens in the present invention and fragment is retrieved such two stages: 1, on searching lens, the present invention represents that with the first frame of each camera lens the content of each camera lens, this method can improve the speed of fragment retrieval greatly; Simultaneously, in order to ensure the integrality and the correctness of the fragment that retrieval obtains, the present invention proposes to adopt the similar low threshold value of camera lens to ensure recall ratio and the precision ratio that similar camera lens is high simultaneously with fragment length constraint way of combining.2, in the fragment retrieval, set up the cum rights bipartite graph model of fragment retrieval, tentatively obtain similar one by one fragment by the continuity of investigating similar camera lens, use the Kuhn_Munkres method to try to achieve the maximum weight matching of these fragments and query fragment again, coupling be used to judge two fragments whether the phase Sihe determine the border of similar fragment, authority is used to represent the similarity of two similar fragments.The present invention uses the Optimum Matching of graph theory to solve the video frequency searching problem first, this is because the thought of coupling requires similar camera lens corresponding (granularity) one by one, under this condition, the Optimum Matching of obtaining can objectively reflect two number of shots and two degree that the fragment vision is similar that fragment is similar all sidedly, thereby has avoided the granularity problem that camera lens calculates in the existing method.Experimental result shows, compares with existing method with said function, and no matter be precision ratio, recall ratio, or retrieval rate, the present invention has obtained better effect.
Description of drawings
Fig. 1 is the cum rights bipartite graph model synoptic diagram of fragment retrieval;
Fig. 2 is the result for retrieval of the present invention to tennis tournament;
Fig. 3 is the result for retrieval of the present invention to personage's fragment in " black hole ";
Fig. 4 is the result for retrieval of the present invention to the Talks on Focus head.
Embodiment
The present invention is described in further detail below in conjunction with accompanying drawing.
A kind ofly carry out the method for video frequency searching, may further comprise the steps by video segment:
1, at first using space-time section algorithm (spatio-temporal slice) to carry out shot boundary detects, is the Video Segmentation among query fragment X and the video library Y camera lens, can list of references " Video Partitioning by Temporal Slice Coherency " [C.W.Ngo about the detailed description of space-time section algorithm, T.C.Pong, and R.T.Chin, IEEE Transactions on Circuits and Systems forVideo Technology, Vol.11, No.8, pp.941-953, August, 2001];
2, the content of representing each camera lens then with the first frame of each camera lens; Similar value Similarity (the x of two camera lenses i, y j) be the first frame result calculated according to two camera lenses, wherein x i, y jRepresent two camera lenses, then, setting threshold T of the present invention s=0.71, as Similarity (x i, y j)>T s, just think two camera lens x iAnd y jSimilar, according to this formula, retrieve among the video database Y camera lens x with query fragment X iAll similar camera lens y j
3, set up the cum rights bipartite graph model of fragment retrieval,, tentatively be partitioned into the fragment Y similar to query fragment X by investigating the continuity of similar camera lens kAs shown in Figure 1, the cum rights bipartite graph model synoptic diagram G={X of fragment retrieval, Y, among the E}, vertex set V=X ∪ Y, X={x 1, x 2..., x nThe expression query fragment, Y={y 1, y 2..., y mThe expression video library, limit collection E={e Ij, e IjExpression camera lens x iWith y jSimilar, e IjWeights ω IjExpression camera lens x iWith y jSimilar value.
Suppose query fragment X={x 1, x 2..., x n, video library Y={y 1, y 2..., y m(m>>n), x wherein i, y jThe expression camera lens, so, the similar camera lens corresponding diagram of X and Y can be expressed as the cum rights bipartite graph G={X in the graph theory, Y, E}, wherein, vertex set V=X ∪ Y, limit collection E={e Ij, e IjExpression camera lens x iWith y jSimilar, e IjWeights ω IjExpression camera lens x iWith y jSimilar value, as shown in Figure 1.For video library Y, the camera lens similar to query fragment X is a minority, and a large amount of camera lenses is also dissimilar, is reflected in Fig. 1, i.e. the node y of incidence edge jBe minority, a large amount of node y lThe frontier juncture does not join, therefore, and the y of continuous incidence edge jBetween, generally can be by the y on a plurality of continuous dereferenceds limit lSeparate.Utilize this character, according to the definition that fragment is made up of cinestrip, the present invention is at first with camera lens y similar to X among the Y jBy rank order from small to large, investigate y then jContinuity, if | y J+1-y j|>2, j=1,2 ..., m-1 then tentatively obtains a fragment Y similar to X k={ y i, y I+1..., y j, i, j ∈ [1, m].
Obtaining fragment Y kAfter, the present invention adopts the constraint condition of fragment length to remove the dissimilar camera lens that low threshold value is brought.If promptly | Y k|<L, then ignore.The present invention gets
Figure C0314999800101
N is the length of query fragment X.
4, because Y kA plurality of fragment Ys similar may have been comprised to X X, in order further to obtain Y X, simultaneously, estimate Y XWith the similarity of X, the present invention uses the Kuhn_Munkres method to calculate k bipartite graph G k={ X, Y k, E kMaximum weight matching M, from M, obtain Y X, represent X and Y with this authority simultaneously XSimilarity, the Kuhn_Munkres method of specifically calculating Optimum Matching is as follows:
(1) provides initial label l ( x i ) = max j ω ij ,l(y j)=0,i,j=1,2...,t,t=max(n,|Y k|);
(2) obtain limit collection E l={ (x i, y j) | l (x i)+l (y j)=ω Ij, G l=(X, Y k, E l) and G lIn one the coupling M:
(3) as all nodes of the saturated X of M, then M promptly is the Optimum Matching of G, calculates and finishes, otherwise carry out next step;
(4) in X, look for a M unsaturation point x 0, make A ← { x 0, B ← φ, A, B are two set;
(5) if N Gl(A)=and B, then changeed for (9) step, otherwise carry out next step, wherein, N Gl(A) Y k, be with A in the node set of node adjacency;
(6) look for a node y ∈ N Gl(A)-B;
(7) if y is the M saturation point, then find out the match point z of y, make A ← A ∪ z}, { y} changeed for (5) step, otherwise carries out next step B ← B ∪;
(8) there is one from x 0But the augmenting path P to y makes M ← M E (P), changes for (3) step;
(9) be calculated as follows the α value: a = min x i ∈ A y j ∉ N G l ( A ) { l ( x i ) + l ( y j ) - ω ij } , Revise label:
Ask E according to l ' L 'And G L '
(10), l ← l ', G l← G L 'Changeed for (6) step.
It is function Kuhn_Munkres (X, a Y that the present invention defines the Kuhn_Munkres method k), so, obtain similar fragment Y XWith evaluation Y XBe described below with the method for X similarity:
(1) calls Kuhn_Munkres (X, Y k), obtain G k={ X, Y k, E kMaximum weight matching M and authority ω;
(2) get all ω of the related M of X Ij>0 y, ordering is Y '={ y from small to large α, y β... y γ, α, beta, gamma ∈ [i, j], in this set, y α, y βPossibility is also discontinuous, i.e. y β-y α>1, according to the definition and the Y of front video segment kMiddle camera lens is all similar in appearance to X, and the present invention gets y αWith y γBetween all camera lenses constitute continuous fragment Y X={ y α, y α+1..., y γ, alpha, gamma ∈ [i, j];
(3) if | Y ' |<L, then Y XIgnore (length constraint after the Kuhn_Munkres method), calculate and finish, otherwise carry out next step;
(4) calculate X and Y XSimilarity Similarity ( X , Y X ) = ω n ;
(5) if | X| 〉=| Y k|, calculate and finish, otherwise carry out next step;
(6) for | X|<| Y k|, at Y kGet Y XAfter, see Y kCan remaining fragment continue to use the Kuhn_Munkres method, if promptly (α-i) 〉=L, Y k← { y i, y I+1, y α-1, change (1); (if j-γ) 〉=L, Y x← { y γ+1, y γ+2..., y j, change (1).
Get In above-mentioned (4) step, the present invention uses Similarity ( X , Y X ) = ω n With Similarity (X, Y X) normalize between [0,1], this value is big more, shows that two fragments are similar more.Relatively more existing method, this formula has succinctly been estimated two maximum similarities that fragment can reach effectively.
With experimental result the outstanding representation of the present invention in video clip retrieval is described below.Experimental data is several days programs from television recording, this video database is very challenging, always have 3 hours 11 minutes, 4714 camera lenses, 286936 two field pictures, comprised advertisement, news, physical culture, the various types of programs of film, the same video fragment of repetition has been arranged here, as the head of news, advertisement etc.; The similar video segments that a lot of repetitions are also arranged is as same advertisement of the different tennis tournaments in the sports cast, different time length and editor etc.In order to verify validity of the present invention, we have used existing method as the experiment contrast, mainly contain such two reasons: 1, existing method is present given experimental data the best way, also is up-to-date a kind of method; 2, consistent with function of the present invention, can in video library, be partitioned into similar fragment automatically, arrange these similar fragments from high to low by similarity then.In video clip retrieval, except precision ratio and recall ratio, retrieval rate also is a very important index.Relatively more existing method, the present invention has tested the index of retrieval rate first, has compared the retrieval rate of two kinds of methods, and the test machine of use is PIII Dual CPU 1G Hz, internal memory 256M.
Fig. 2, Fig. 3 and Fig. 4 are result for retrieval of the present invention: top delegation is the query fragment that the user submits to, demonstration be the first frame of its each camera lens, be below the retrieval the result, successively arrange according to the order that similarity is successively decreased.First row that retrieves promptly is a query fragment itself, and that yes is the highest for its similarity, and the order that remaining fragment is successively decreased according to similarity is successively arranged.Can see that the fragment that retrieves is all similar with query fragment.Wherein, Fig. 2 is the fragment about tennis tournament; Fig. 3 is about personage's scene in the TV play " black hole "; Fig. 4 is the head of Talks on Focus.Concrete experimental result provides at table 1 and table 2 respectively, it is the same substantially with query fragment accurately to retrieve the fragment of indicating to retrieve in the table 1, have same camera lens and frame sequence, the fragment that similarity retrieval indicates to retrieve in the table 2 only has identical semanteme with query fragment, as tennis tournament, utilize and low-level features such as color are difficult.
Table 1 video segment is the experimental result of retrieval accurately
Query fragment Frame number The present invention Existing method
Precision ratio Recall ratio Speed (second) Precision ratio Recall ratio Speed (second)
1, the head of news 832 100% 100% 9 75% 100% 230
2, football news 715 100% 100% 11 100% 100% 196
3, Huiyuan's advertisement 367 100% 100% 21 33.3% 100% 97
4, bright advertisement 374 100% 100% 11 100% 100% 101
5, good fortune advertisement near the house 432 100% 100% 13 100% 100% 116
On average 544 100% 100% 13 81.7% 100% 148
The experimental result of table 2 video segment similarity retrieval
Query fragment Frame number The present invention Existing method
Precision ratio Recall ratio Speed (second) Precision ratio Recall ratio Speed (second)
1, tennis tournament 507 100% 50% 5 100% 50% 140
2, the doctor gives emergency treatment to a patient 1806 100% 50% 13 50% 50% 507
3, TCL advertisement 374 100% 100% 12 85.7% 100% 100
4, melatonin advertisement 374 100% 75% 17 100% 100% 100
5, Amoisonic's advertisement 374 80% 100% 15 100% 50% 99
On average 687 96% 75% 12 87.1% 70% 189
No matter can see from table 1 and table 2, be precision ratio, or recall ratio, and the present invention is better than existing method.Main cause is that existing method only calculates the quantity of two similar camera lenses of fragment, and the present invention has considered the corresponding relation of similar camera lens.In addition, great advantage of the present invention is embodied on the retrieval rate of fragment, according to our experiment, basically be to equal the time that similar camera lens is judged total retrieval time, existing method adopts the way that compares frame by frame in chronological order, and the present invention only need compare the first frame of each camera lens, so the present invention is far away faster than existing method, fast order of magnitude can satisfy the demand of real-time retrieval basically.

Claims (6)

1, a kind ofly carry out the method for video frequency searching, may further comprise the steps by video segment:
(1) at first carrying out shot boundary and detect, is the Video Segmentation in query fragment and the video library camera lens;
(2) represent the content of each camera lens then with the first frame of each camera lens; Similar value Similarity (the x of two camera lenses i, y j) be the first frame result calculated according to two camera lenses, wherein x i, y jRepresent two camera lenses, as Similarity (x i, y j)>T s, just think two camera lens x iAnd y jSimilar, T in the formula sFor the similar threshold value of camera lens,, retrieve among the video database Y camera lens x with query fragment X according to this formula iAll similar camera lens y j
(3) set up the cum rights bipartite graph model that fragment is retrieved,, tentatively be partitioned into the fragment Y similar to query fragment X by investigating the continuity of similar camera lens k
(4) because Y kA plurality of fragment Ys similar may have been comprised to X X, in order further to obtain Y X, and can estimate Y simultaneously XWith the similarity of X, adopt the Optimum Matching method to calculate k bipartite graph G k={ X, Y k, E kMaximum weight matching M, from M, obtain Y X, represent X and Y with this authority simultaneously XSimilarity.
2, as claimed in claim 1ly a kind ofly carry out the method for video frequency searching, it is characterized in that: in step (2), utilize the fragment length constraint setting threshold T of Optimum Matching by video segment s, remove dissimilar camera lens, threshold value T sBe 0~1.
3, as claimed in claim 2ly a kind ofly carry out the method for video frequency searching, it is characterized in that: threshold value T by video segment s=0.71.
4, as claimed in claim 1ly a kind ofly carry out the method for video frequency searching, it is characterized in that in step (3), adopt following method tentatively to be partitioned into fragment similar among the video library Y: camera lens y similar among the Y to X to query fragment X by video segment jBy rank order from small to large, investigate y then jContinuity, if | y J+1-y j|>2, j=1,2 ..., m-1 then tentatively obtains a fragment Y similar to X k={ y i, y I+1..., y j, i, j ∈ [1, m].
5, as claimed in claim 2ly a kind ofly carry out the method for video frequency searching, it is characterized in that using the constraint condition of fragment length to filter out T by video segment sThe dissimilar camera lens that brings, realize by twice constraint:
(1), adopts the preceding length constraint of Optimum Matching method: for fear of unnecessary calculating, if Y kFragment length do not reach requirement, fall with regard to direct filtration, if promptly
Figure C031499980002C1
Then ignore, wherein, Y k={ y i, y I+1..., y j, i, j ∈ [1, m], n is the length of query fragment X;
(2), the length constraint after the employing Optimum Matching method: all ω that get the related Optimum Matching M of X Ij>0 y, ordering is Y '={ y from small to large α, y β..., y γ, α, beta, gamma ∈ [i, j], if
Figure C031499980002C2
Then ignore.
6, describedly a kind ofly carry out the method for video frequency searching by video segment as claim 1 or 5, it is characterized in that in the step (4), the Optimum Matching method is the Kuhn_Munkres method, this method is as follows:
(1) provides initial label l ( x i ) = max j ω ij , L (y j)=0, i, j=1,2 ..., t, and t=max (n, | Y k|), n is the length of query fragment X, Y k={ y i, y I+1..., y j, i, j ∈ [1, m];
(2) obtain limit collection E l={ (x i, y j) | l (x i)+l (y j)=ω Ij, G l=(X, Y k, E l) and G lIn one the coupling M;
(3) as all nodes of the saturated X of M, then M promptly is the Optimum Matching of G, calculates and finishes, otherwise carry out next step;
(4) in X, look for a M unsaturation point x 0, make A ← { x 0, B ← φ, A, B are two set;
(5) if N G l ( A ) = B , Then changeed for (9) step, otherwise carry out next step, wherein, N Gl(A) Y k, be with A in the node set of node adjacency;
(6) look for a node y ∈ N Gl(A)-B;
(7) if y is the M saturation point, then find out the match point z of y, make A ← A ∪ z}, { y} changeed for (5) step, otherwise carries out next step B ← B ∪;
(8) there is one from x 0But the augmenting path P to y makes M ← M E (P), changes for (3) step;
(9) be calculated as follows a value: a = min x i ∈ A y j ∉ N G l { l ( x i ) + l ( y j ) - ω } , Revise label:
Figure C031499980003C4
Ask E according to l ' L 'And G L '
(10) l ← l ', G l← G L ', changeed for (6) step;
Definition Kuhn_Munkres method is function Kuhn_Munkres (X, a Y k), so, obtain similar fragment Y XWith evaluation Y XBe described below with the method for X similarity:
(1) calls Kuhn_Munkres (X, Y k), obtain G k={ X, Y k, E kMaximum weight matching M and authority ω;
(2) get all ω of the related M of X Ij>0 y, ordering is Y '={ y from small to large α, y β..., y γ, α, beta, gamma ∈ [i, j], in this set, y α, y βPossibility is also discontinuous, i.e. y β-y α>1, according to the definition and the Y of front video segment kMiddle camera lens is all similar in appearance to X, and the present invention gets y αWith y γBetween all camera lenses constitute continuous fragment Y X={ y α, y α+1..., y γ, alpha, gamma ∈ [i, j], i, j ∈ [1, m];
(3) if | Y ' |<L, then Y XIgnore, calculate and finish, otherwise carry out next step;
(4) calculate X and Y XSimilarity Similarity ( X , Y X ) = ω n ;
(5) if | X| 〉=| Y k|, calculate and finish, otherwise carry out next step;
(6) for | X|<| Y k|, at Y kGet Y XAfter, see Y kCan remaining fragment continue to use the Kuhn_Munkres method, if promptly (α-i) 〉=L, Y k← { y i, y I+1, y α-1, change (1); (if j-γ) 〉=L, Y k← { y γ+1, y γ+2..., y j, change (1);
Get In above-mentioned (4) step, the present invention uses Similarity ( X , Y X ) = ω n With Similarity (X, Y X) normalize between [0,1], this value is big more, shows that two fragments are similar more.
CN 03149998 2003-08-04 2003-08-04 Method of proceeding video frequency searching through video frequency segment Expired - Fee Related CN1245697C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 03149998 CN1245697C (en) 2003-08-04 2003-08-04 Method of proceeding video frequency searching through video frequency segment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 03149998 CN1245697C (en) 2003-08-04 2003-08-04 Method of proceeding video frequency searching through video frequency segment

Publications (2)

Publication Number Publication Date
CN1514644A CN1514644A (en) 2004-07-21
CN1245697C true CN1245697C (en) 2006-03-15

Family

ID=34240515

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 03149998 Expired - Fee Related CN1245697C (en) 2003-08-04 2003-08-04 Method of proceeding video frequency searching through video frequency segment

Country Status (1)

Country Link
CN (1) CN1245697C (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101201822B (en) * 2006-12-11 2010-06-23 南京理工大学 Method for searching visual lens based on contents

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101281520B (en) * 2007-04-05 2010-04-21 中国科学院自动化研究所 An Interactive Sports Video Retrieval Method Based on Unsupervised Learning and Semantic Matching Features
CN101894156A (en) * 2010-07-12 2010-11-24 清华大学 A Video Fingerprint Matching Method Based on Bipartite Graph
CN102591892A (en) * 2011-01-13 2012-07-18 索尼公司 Data segmenting device and method
WO2014000258A1 (en) * 2012-06-29 2014-01-03 中国科学院自动化研究所 Method for automatic positioning of approximately repeated video clips
CN103177099B (en) * 2013-03-20 2017-05-24 深圳先进技术研究院 Video comparison method and video comparison system
CN103475935A (en) * 2013-09-06 2013-12-25 北京锐安科技有限公司 Method and device for retrieving video segments
CN108271049B (en) * 2017-12-01 2020-09-04 阿里巴巴(中国)有限公司 Resource downloading method and device and computer equipment
CN108170791A (en) * 2017-12-27 2018-06-15 四川理工学院 Video image search method
CN115442656B (en) * 2021-06-04 2023-08-15 中国移动通信集团浙江有限公司 Method, device, equipment and storage medium for automatic detection of video title and title
CN115034621A (en) * 2022-06-14 2022-09-09 杭州卓健信息科技股份有限公司 Multidimensional data fusion clinical diagnosis and treatment intelligent teaching management system and method
CN115499707B (en) * 2022-09-22 2024-08-06 上海联屏文化科技有限公司 Video similarity determination method and device

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101201822B (en) * 2006-12-11 2010-06-23 南京理工大学 Method for searching visual lens based on contents

Also Published As

Publication number Publication date
CN1514644A (en) 2004-07-21

Similar Documents

Publication Publication Date Title
CN1245697C (en) Method of proceeding video frequency searching through video frequency segment
Kang Affective content detection using HMMs
CN106162223B (en) A kind of news video segmentation method and device
Lin et al. Automatic video scene extraction by shot grouping
JP5711387B2 (en) Method and apparatus for comparing pictures
CN1311388C (en) Method and apparatus for representing a group of images
US8995823B2 (en) Method and system for content relevance score determination
KR20170128771A (en) Entity-based temporal segmentation of video streams
CN1206847C (en) Video segment searching method based on contents
CN107358141B (en) Data identification method and device
CN101398854A (en) Video fragment searching method and system
CN103984778B (en) A kind of video retrieval method and system
CN103617263A (en) Television advertisement film automatic detection method based on multi-mode characteristics
CN1240014C (en) Method for making video search of scenes based on contents
CN110610500A (en) News Video Adaptive Stripping Method Based on Dynamic Semantic Features
CN101404030B (en) Method and system for periodic structure fragment detection in video
CN103177099A (en) Video comparison method and video comparison system
CN102346768B (en) Method and device for finding video advertisement
CN101615255B (en) Video text multi-frame interfusion method
CN112418269B (en) Social media network event propagation key time prediction method, system and medium
Tong et al. A unified framework for semantic shot representation of sports video
CN106844573B (en) Video abstract acquisition method based on manifold sorting
CN1508755A (en) Sensitive video detection method
CN1252647C (en) Scene-searching method based on contents
CN100507910C (en) A Method of Integrating Color and Motion Features for Shot Retrieval

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20060315

CF01 Termination of patent right due to non-payment of annual fee