CN1477566A - Method for making video search of scenes based on contents - Google Patents

Method for making video search of scenes based on contents Download PDF

Info

Publication number
CN1477566A
CN1477566A CNA031501265A CN03150126A CN1477566A CN 1477566 A CN1477566 A CN 1477566A CN A031501265 A CNA031501265 A CN A031501265A CN 03150126 A CN03150126 A CN 03150126A CN 1477566 A CN1477566 A CN 1477566A
Authority
CN
China
Prior art keywords
sigma
camera lens
similarity
fuzzy
carried out
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CNA031501265A
Other languages
Chinese (zh)
Other versions
CN1240014C (en
Inventor
董庆杰
彭宇新
郭宗明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BEIDA FANGZHENG TECHN INST Co Ltd BEIJING
Inst Of Computer Science & Technology Peking University
Original Assignee
BEIDA FANGZHENG TECHN INST Co Ltd BEIJING
Inst Of Computer Science & Technology Peking University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by BEIDA FANGZHENG TECHN INST Co Ltd BEIJING, Inst Of Computer Science & Technology Peking University filed Critical BEIDA FANGZHENG TECHN INST Co Ltd BEIJING
Priority to CN 03150126 priority Critical patent/CN1240014C/en
Publication of CN1477566A publication Critical patent/CN1477566A/en
Application granted granted Critical
Publication of CN1240014C publication Critical patent/CN1240014C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The present invention relates to a method of making video search of scene based on contents. It uses the fuzzy aggregation analysis method in the scene search. As compared with existent method it can obtain higher accuracy and quick searching speed.

Description

A kind of method of camera lens being carried out Content-based Video Retrieval
Technical field
The invention belongs to video search technique area, be specifically related to a kind of method of camera lens being carried out Content-based Video Retrieval.
Background technology
Along with in the great technical progress that obtains aspect multi-medium data manufacturing, storage and the propagation, digital video has become a part indispensable in people's daily life.The problem that people face no longer is to lack content of multimedia, but how to find own needed information in the vast as the open sea multimedia world.At present, traditional video frequency searching of describing based on keyword is because descriptive power is limited, and subjectivity is strong, reasons such as manual mark, demand that can not the satisfying magnanimity video frequency searching.In order to be convenient for people to seek multi-medium data, begin the nineties in last century, content-based video analysis and retrieval technique become the hot issue of research, and Multimedia Content Description Interface MPEG-7 progressively formulates and optimizes, and have promoted the development of Content-based Video Retrieval technology more.
In the prior art, as document " A New Approach to Retrieval Video by ExampleVideo Clip " [X.M.Liu, Y.T.Zhuang, and Y.H.Pan, ACM Multimedia, pp.41-44,1999] described, the conventional method of video frequency searching is at first to carry out shot boundary to detect, with basic structural unit and the retrieval unit of camera lens as video sequence; Represent the content of this camera lens then at each camera lens internal extraction key frame, go out low-level features such as color and texture from key-frame extraction, be used for the index and the retrieval of camera lens.Like this, just content-based searching lens being converted into CBIR solves.The problem that these class methods exist is that camera lens is an image continuous sequence in time, temporal information and the movable information that is present in the video is not fully utilized.(document author is s.H.Kim and R.-H.Park to the document of delivering at IEEE Trans.Circuits and Systems for Video Technology in 2002 " An efficient algorithm forvideo sequence matching using the modified Hausdorff distance and the directeddivergence " in addition, vol.CSVT-12, no.7, page number 592-595) disperses (Cumulative Directed Divergence) method with the orientation of accumulation and extract key frame, obtain two similarity degrees between the camera lens with improved Hausdorff distance (Modified Hausdorff Distance) method, used the YUV color space histogram when extracting key frame and definition camera lens similarity.Owing to set two threshold values when extracting key frame: the threshold value of similar value between the threshold value of front and back frame similar value and present frame and the previous key frame, these two conditions must be satisfied simultaneously and a key frame could be occurred, will influence the accuracy of key-frame extraction like this, finally will certainly influence the correctness of inquiry; In addition, used the YUV color space of using always in the video as visual signature, it is compared with the hsv color space, and is also little consistent with people's visually-perceptible.
Summary of the invention
At the existing defective of existing searching lens method, the objective of the invention is to propose a kind of method of camera lens being carried out Content-based Video Retrieval, this method can improve the accuracy rate of content-based searching lens on the basis of existing technology greatly, keep very fast retrieval rate simultaneously, thereby bring into play the huge effect of searching lens technology in current network information society more fully.
The object of the present invention is achieved like this: a kind of camera lens is carried out the method for Content-based Video Retrieval, may further comprise the steps:
(1) at first video database is carried out camera lens and cut apart, with the basic structural unit and the retrieval unit of camera lens as video;
(2) similarity between two two field pictures of calculating is set up fuzzy similarity matrix R by following method: when i=j, make r IjBe 1; When i ≠ j, make r IjBe x iWith y jBetween similarity;
(3) utilize the transitive closure method to calculate the equivalent matrice of fuzzy similarity matrix R
(4) threshold value λ is set and determines cut set, the transitive closure matrix of R matrix
Figure A0315012600052
Carry out fuzzy clustering, calculate [ x ] = { y | R ^ ( x , y ) ≥ λ } , Set [x] is the equivalence class of fuzzy clustering, and each frame is similar in each equivalence class set, so we can get in each set arbitrary frame as key frame;
(5) with key frame { r I1, r I2..., r IkExpression camera lens s i, gather with key frame and to measure two similaritys between the camera lens.
Furthermore, in the step (1) video database is carried out the method that camera lens cuts apart and be preferably space-time section algorithm.Calculate x in the step (2) iWith y jBetween similarity can calculate with the friendship of two image histograms: Inter sec t ( x i , y j ) = 1 A ( x i , y j ) Σ h Σ s Σ v min { H i ( h , s , v ) , H j ( h , s , v ) } A ( x i , x j ) = min { Σ h Σ s Σ v H i ( h , s , v ) , Σ h Σ s Σ v H j ( h , s , v ) }
H i(h, s v) are the histograms in hsv color space, and we use H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as color feature value, Intersect (x i, y j) two histogrammic friendships of expression, judge the similarity of two key frames with it, use A (x i, y j) normalize between 0,1.
Further again, in the step (3), calculate the equivalent matrice of fuzzy similarity matrix R The transitive closure method can adopt quadratic method: R → R 2 → ( R 2 ) 2 → · · · → R 2 k = R ^ , Its time complexity is O (n 3Log 2N), if the n value is big especially, will certainly influence total computing time, so adopt the compose operation of the fuzzy clustering optimal algorithm compute matrix that calculates based on figure connected component, recursion is as follows: r ij ( 0 ) = r ij - - - 0 ≤ i , j ≤ n r ij ( k ) = max { r ij ( k - 1 ) , min [ r ik ( k - 1 ) , r kj ( k - 1 ) ] } - - - - 0 ≤ i , j ≤ n ; 0 ≤ k ≤ n
The time complexity T (n) of this algorithm satisfies O (n)≤T (n)≤O (n 2).
In order to realize purpose of the present invention better, when carrying out searching lens, right
Figure A0315012600065
The method of carrying out fuzzy clustering is as follows:
(1) determines n sample X=(X 1..., X n) on fuzzy resembling relation R and a cut set threshold alpha;
(2) transform R as an equivalent matrice by following calculating;
RoR=R 2
R 2oR 2=R 4
... R 2 k o R 2 k = R 2 ( K + 1 )
Up to existing a k to satisfy R 2 k = R 2 ( k + 1 )
In the above-mentioned formula, RoR is the compose operation of fuzzy relation, is under the hypothesis of similar matrix at R, and having proved to have such k to exist, and satisfies k≤log n;
(3) set of computations [ x ] = { y | R ^ ( x , y ) ≥ α } , [x] is fuzzy clustering, and algorithm finishes;
After n sample space carried out fuzzy cluster analysis, obtain several equivalence classes, in each equivalence class, choose a sample as key frame.Measuring similarity between such two camera lenses just becomes the similarity measurement between the key frame set.
In the step (5) of this method, can be camera lens s iAnd s jSimilarity be defined as Sim ( s i , s j ) = 1 2 { M ( s i , s j ) + M ^ ( s i , s j ) } , M represents the maximal value that key frame is similar, The similar second largest value of expression key frame, wherein, M ( s i , s j ) = max p = { 1,2 , . . . } max q = { 1,2 , . . . } { Inter sec t ( r ip , r jq ) } M ^ ( s i , s j ) = max p = { 1,2 , . . . } ^ max q = { 1,2 , . . . } { Inter sec t ( r ip , r jq ) } Inter sec t ( r i , r j ) = 1 A ( r i , r j ) Σ h Σ s Σ v min { H i ( h , s , v ) , H j ( h , s , v ) } A ( r i , r j ) = min { Σ h Σ s Σ v H i ( h , s , v ) , Σ h Σ s Σ v H j ( h , s , v ) } .
Effect of the present invention is: adopt and of the present invention camera lens is carried out the method for Content-based Video Retrieval, can obtain higher accuracy rate, keep very fast retrieval rate simultaneously.
Why the present invention has so significant technique effect, its reason is: the method for utilization fuzzy cluster analysis, the camera lens content is divided into a plurality of equivalence classes, these equivalence classes have well been described the camera lens content change, and the similarity between the camera lens then shows as the similarity between the key frame combination.Similarity measurement has been considered to use the hsv color histogram to represent the shortcoming of key frame between the camera lens: if two key frames have similar color distribution, even their content is different, can think that also these two key frames are similar.So the mean value that uses maximum similar value and second largest similar value adds the robustness of strong algorithms.Contrast and experiment has confirmed that the present invention proposes the validity of method.
Description of drawings
Fig. 1 is the schematic flow sheet that camera lens is carried out the method for Content-based Video Retrieval;
Fig. 2 is 7 semantic category example synoptic diagram of searching lens in the experiment contrast;
Fig. 3 is the result for retrieval synoptic diagram of method of the present invention to the swimming camera lens.
Embodiment
Fig. 1 is an overall framework of the present invention, is the schematic flow sheet of each one step process among the present invention.As shown in Figure 1, a kind of searching lens method based on fuzzy cluster analysis may further comprise the steps:
1, camera lens is cut apart
At first use space-time section algorithm (spatio-temporal slice), video database is carried out camera lens to be cut apart, with basic structural unit and the retrieval unit of camera lens as video, can list of references " Video Partitioning by Temporal Slice Coherency " [C.W.Ngo about the detailed description of space-time section algorithm, T.C.Pong, and R.T.Chin, IEEE Transactions on Circuits andSystems for Video Technology, Vol.11, No.8, pp.941-953, August, 2001].
2, set up fuzzy similarity matrix R
Set up between the camera lens internal image to set up fuzzy similarity matrix R method as follows: when i=j, make r IjBe 1, when i ≠ j, make r IjBe x iWith y jBetween similarity, similarity then adopts following method to calculate: Inter sec t ( x i , y j ) = 1 A ( x i , y j ) Σ h Σ s Σ v min { H i ( h , s , v ) , H j ( h , s , v ) } A ( x i , x j ) = min { Σ h Σ s Σ v H i ( h , s , v ) , Σ h Σ s Σ v H j ( h , s , v ) }
H i(h, s v) are the histograms in hsv color space, and we use H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as color feature value.Intersect (x i, y j) two histogrammic friendships of expression, judge the similarity of two key frames with it, use A (x i, y j) normalize between 0,1.
3, ask the transitive closure of similar matrix R, obtain equivalent matrice
In the present embodiment, ask the transitive closure of similar matrix to adopt quadratic method: R → R 2 → ( R 2 ) 2 → · · · → R 2 k = R ^
Its time complexity is O (n 3Log 2N), if the n value is big especially, will certainly influence total computing time.So adopt the compose operation of the fuzzy clustering optimal algorithm compute matrix that calculates based on figure connected component, recursion is as follows: r ij ( 0 ) = r ij - - - 0 ≤ i , j ≤ n r ij ( k ) = max { r ij ( k - 1 ) , min [ r ik ( k - 1 ) , r kj ( k - 1 ) ] } - - - - 0 ≤ i , j ≤ n ; 0 ≤ k ≤ n .
The time complexity T (n) of this algorithm satisfies O (n)≤T (n)≤O (n 2).
4, threshold value λ is set and determines cut set, the transitive closure matrix of R matrix Carry out fuzzy clustering.
In the present embodiment, concrete grammar is as follows:
(1) determines n sample X=(x 1..., x n) on fuzzy resembling relation R and a cut set threshold alpha;
(2) transform R as an equivalent matrice by following calculating;
RoR=R 2
R 2oR 2=R 4
... R 2 k o R 2 k = R 2 ( K + 1 )
Up to existing a k to satisfy R 2 k = R 2 ( k + 1 )
In the above-mentioned formula, RoR is the compose operation of fuzzy relation, is under the hypothesis of similar matrix at R, and having proved to have such k to exist, and satisfies k≤log n;
(3) set of computations [ x ] = { y | R ^ ( x , y ) ≥ α } , [x] is fuzzy clustering, and algorithm finishes
5, obtain the camera lens key frame with Fuzzy Cluster Analysis method after, carry out searching lens based on these key frames then.On this basis, with key frame { r I1, r I2..., r Ik) the expression camera lens, s iCamera lens s iAnd s jSimilarity be defined as Sim ( s i , s j ) = 1 2 { M ( s i , s j ) + M ^ ( s i , s j ) } , Wherein, M ( s i , s j ) = max p = { 1,2 , . . . } max q = { 1,2 , . . . } { Inter sec t ( r ip , r jq ) } M ^ ( s i , s j ) = max p = { 1,2 , . . . } ^ max q = { 1,2 , . . . } { Inter sec t ( r ip , r jq ) } Inter sec t ( r i , r j ) = 1 A ( r i , r j ) Σ h Σ s Σ v min { H i ( h , s , v ) , H j ( h , s , v ) } A ( r i , r j ) = min { Σ h Σ s Σ v H i ( h , s , v ) , Σ h Σ s Σ v H j ( h , s , v ) } .
Figure A0315012600095
Represent second largest value, use
Figure A0315012600096
Be that its shortcoming is if two key frames have similar color distribution, even their content is different, can think that also these two key frames are similar because this paper uses the hsv color histogram to represent key frame, in order to overcome this defective, use M and
Figure A0315012600097
Mean value add the robustness of strong algorithms.H i(h, s v) are the histograms in hsv color space, this paper H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as color feature value.Intersect (r i, r j) two histogrammic friendships of expression, this paper judges the similarity of two key frames with it.
Following experimental result shows that the present invention has obtained than the better effect of existing method, and retrieval rate is very fast simultaneously, has confirmed the validity of fuzzy cluster analysis algorithm in searching lens.
The experimental data of searching lens is the Asian Games programs in 2002 from television recording, always has 41 minutes, 777 camera lenses, 62132 two field pictures.It comprises multiple sports items, as various ball game, weight lifting, swim and the advertising programme that intercuts etc.We have selected 7 semantic categories as the inquiry camera lens, and they are weight lifting, vollyball, swimming, judo, row the boat, gymnastics, football, as shown in Figure 2.
In order to verify validity of the present invention, we have tested following 3 kinds of methods and have done the experiment contrast:
(1) the first frame of Chang Yong each camera lens of use is done the searching lens algorithm of key frame;
(document author is s.H.Kim and R.-H.Park to the document of delivering at IEEE Trans.Circuits and Systems for Video Technology in (2) 2002 years " An efficient algorithm for video sequence matching using the modifiedHausdorffdistance and the directed divergence ", vo1.CSVT-12, no.7, page number 592-595) the middle algorithm of describing;
(3) use the fuzzy cluster analysis algorithm to obtain key frame and carry out searching lens (only using color characteristic);
Above-mentioned preceding 3 kinds of methods have all only been used color characteristic, and therefore last experimental result can prove the superiority of the disclosed method of the present invention from the measure of shot similarity.Fig. 3 has provided the user interface of experimental arrangement, delegation is the browsing area of inquiry video above the right, and the 1st of each camera lens the key frame is used for representing each camera lens in the display video, the camera lens that the user can therefrom select to want to inquire about is retrieved, and is the Query Result zone below the right.Fig. 3 is the 1st camera lens of delegation above selecting, it is a swimming camera lens, and the first two field picture 022430.bmp represents by this camera lens, the authority of the similarity that calculates according to method of the present invention, arrange Query Result (from left to right, arranging from top to bottom) from big to small.Below, the left side is a simple and easy broadcast phase, double-clicks that section video that the result for retrieval image can be play corresponding camera lens correspondence.
Two kinds of evaluation indexes in the MPEG-7 standardization activity have been adopted in experiment: the average adjusted retrieval order of normalization ANMRR (average normalized modified retrieval rank) and recall level average AR (average recall).AR is similar to traditional recall ratio (recall), and ANMRR compares with traditional precision ratio (precision), not only can reflect correct result for retrieval ratio, and can reflect correct result's arrangement sequence number.The ANMRR value is more little, means that the rank of the correct camera lens that retrieval obtains is forward more; The AR value is big more, and it is big more to mean that in the individual Query Result of preceding K (K is the cutoff value of result for retrieval) similar camera lens accounts for the ratio of all similar camera lenses.Table 1 is AR and the ANMRR comparisons of above-mentioned 3 kinds of methods to 7 semantic camera lens classes.
Table 1 the present invention and the contrast and experiment that has two kinds of methods now
Classification Method 1 Method 2 Method 3
????AR ????ANMRR ????AR ????ANMRR ????AR ????ANMRR
Weight lifting ????0.8824 ????0.3098 ????0.8824 ????0.1539 ????0.9412 ????0.2186
Vollyball ????0.6333 ????0.4974 ????0.7895 ????0.3264 ????0.8556 ????0.3279
Swimming ????0.8400 ????0.2676 ????0.8250 ????0.3164 ????0.9200 ????0.2175
Judo ????0.7000 ????0.4310 ????0.8214 ????0.2393 ????0.8000 ????0.3093
Row the boat ????0.8750 ????0.3407 ????0.6875 ????0.3570 ????0.8125 ????0.2223
Gymnastics ????0.7857 ????0.3445 ????0.9600 ????0.1759 ????0.7857 ????0.2056
Football ????0.5789 ????0.4883 ????0.6889 ????0.2815 ????0.8421 ????0.2614
Average ????0.7565 ????0.3827 ????0.8078 ????0.2642 ????0.8510 ????0.2518
No matter as can be seen from Table 1, adopt method of the present invention, be AR, or ANMRR, all obtained than existing two kinds of better effects of algorithm, confirmed that the present invention is used for the Fuzzy Cluster Analysis method method validity of searching lens.The method of method utilization fuzzy cluster analysis of the present invention is divided into a plurality of equivalence classes to the camera lens content, and these equivalence classes have well been described the camera lens content change, and the similarity between the camera lens then shows as the similarity between the key frame combination.Similarity measurement has been considered to use the hsv color histogram to represent the shortcoming of key frame between the camera lens: if two key frames have similar color distribution, even their content is different, can think that also these two key frames are similar.So the mean value that uses maximum similar value and second largest similar value adds the robustness of strong algorithms.Contrast and experiment has confirmed that the present invention proposes the validity of method.In addition, at CPU 500M PIII, on the PC of 256 MB of memory, algorithm ART of the present invention is 22.557 seconds, and for the video library of 777 camera lenses, the retrieval rate of two kinds of algorithms of the present invention all is very fast.

Claims (6)

1, a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that this method may further comprise the steps:
(1) at first video database is carried out camera lens and cut apart, with the basic structural unit and the retrieval unit of camera lens as video;
(2) similarity between two two field pictures of calculating is set up fuzzy similarity matrix R by following method: when i=j, make r IjBe 1; When i ≠ j, make r IjBe x iWith y j, between similarity;
(3) utilize the transitive closure method to calculate the equivalent matrice of fuzzy similarity matrix R
Figure A0315012600021
(4) threshold value λ is set and determines cut set, the transitive closure matrix of R matrix
Figure A0315012600022
Carry out fuzzy clustering, calculate [ x ] = { y | R ^ ( x , y ) ≥ λ } , Set [x] is the equivalence class of fuzzy clustering, and each frame is similar in each equivalence class set, so we can get in each set arbitrary two field picture as key frame;
(5) with key frame { r I1, r I2..., r IkExpression camera lens s i, key frame is gathered and is measured two similaritys between the camera lens.
2, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: in the step (1), it is space-time section algorithm that video database is carried out the method that camera lens cuts apart.
3, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: in the step (2), calculate x iWith y jBetween similarity can calculate with the friendship of two image histograms: Inter sec t ( x i , y j ) = 1 A ( x i , y j ) Σ h Σ s Σ v min { H i ( h , s , v ) , H j ( h , s , v ) } A ( x i , x j ) = min { Σ h Σ s Σ v H i ( h , s , v ) , Σ h Σ s Σ v H j ( h , s , v ) }
H i(h, s v) are the histograms in hsv color space, and we use H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as face color characteristic value, Intersect (x i, y j) two histogrammic friendships of expression, judge the similarity of two key frames with it, use A (x i, y j) normalize between 0,1.
4, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: in the step (3), calculate the equivalent matrice of fuzzy similarity matrix R The transitive closure method adopt quadratic method: R → R 2 → ( R 2 ) 2 → · · · → R 2 k = R ^ , Its time complexity is O (n 3Log 2N), if the n value is big especially, will certainly influence total computing time, so adopt the compose operation of the fuzzy clustering optimal algorithm compute matrix that calculates based on figure connected component, recursion is as follows: r ij ( 0 ) = r ij - - - 0 ≤ i , j ≤ n r ij ( k ) = max { r ij ( k - 1 ) , min [ r ik ( k - 1 ) , r kj ( k - 1 ) ] } - - - - 0 ≤ i , j ≤ n ; 0 ≤ k ≤ n
The time complexity T (n) of this algorithm satisfies O (n)≤T (n)≤O (n 2).
5, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: right The method of carrying out fuzzy clustering is as follows:
(1) determines n sample x=(x 1..., x n) on fuzzy resembling relation R and a cut set threshold alpha;
(2) transform R as an equivalent matrice by following calculating
RoR=R 2
R 2oR 2=R 4
...? R 2 k o R 2 k = R 2 ( k + 1 )
Up to existing a k to satisfy R 2 k = R 2 ( k + 1 )
In the above-mentioned formula, RoR is the compose operation of fuzzy relation, is under the hypothesis of similar matrix at R, and having proved to have such k to exist, and satisfies k≤log n;
(3) set of computations [ x ] = { y | R ^ ( x , y ) ≥ α } , [x] is fuzzy clustering, and algorithm finishes;
After n sample space carried out fuzzy cluster analysis, obtain several equivalence classes, choose a sample as key frame in each equivalence class, the measuring similarity between such two camera lenses just becomes the similarity measurement between the key frame set.
6, describedly a kind of camera lens is carried out the method for Content-based Video Retrieval as claim 1 or 5, it is characterized in that: can be camera lens s iAnd s jSimilarity be defined as Sim ( s i , s j ) = 1 2 { M ( s i , s j ) + M ^ ( s i , s j ) } , M represents the maximal value that key frame is similar, The similar second largest value of expression key frame, wherein, M ( s i , s j ) = max p = { 1,2 , . . . } max q = { 1,2 , . . . } { Inter sec t ( r ip , r jq ) } M ^ ( s i , s j ) = max p = { 1,2 , . . . } ^ max q = { 1,2 , . . . } { Inter sec t ( r ip , r jq ) } Inter sec t ( r i , r j ) = 1 A ( r i , r j ) Σ h Σ s Σ v min { H i ( h , s , v ) , H j ( h , s , v ) } A ( r i , r j ) = min { Σ h Σ s Σ v H i ( h , s , v ) , Σ h Σ s Σ v H j ( h , s , v ) } .
CN 03150126 2003-07-18 2003-07-18 Method for making video search of scenes based on contents Expired - Fee Related CN1240014C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 03150126 CN1240014C (en) 2003-07-18 2003-07-18 Method for making video search of scenes based on contents

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 03150126 CN1240014C (en) 2003-07-18 2003-07-18 Method for making video search of scenes based on contents

Publications (2)

Publication Number Publication Date
CN1477566A true CN1477566A (en) 2004-02-25
CN1240014C CN1240014C (en) 2006-02-01

Family

ID=34156438

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 03150126 Expired - Fee Related CN1240014C (en) 2003-07-18 2003-07-18 Method for making video search of scenes based on contents

Country Status (1)

Country Link
CN (1) CN1240014C (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100399804C (en) * 2005-02-15 2008-07-02 乐金电子(中国)研究开发中心有限公司 Mobile communication terminal capable of briefly offering activity video and its abstract offering method
CN100573523C (en) * 2006-12-30 2009-12-23 中国科学院计算技术研究所 A kind of image inquiry method based on marking area
CN101211355B (en) * 2006-12-30 2010-05-19 中国科学院计算技术研究所 Image inquiry method based on clustering
CN101968797A (en) * 2010-09-10 2011-02-09 北京大学 Inter-lens context-based video concept labeling method
CN101339615B (en) * 2008-08-11 2011-05-04 北京交通大学 Method of image segmentation based on similar matrix approximation
WO2011140783A1 (en) * 2010-05-14 2011-11-17 中兴通讯股份有限公司 Method and mobile terminal for realizing video preview and retrieval
CN104217000A (en) * 2014-09-12 2014-12-17 黑龙江斯迪克信息科技有限公司 Content-based video retrieval system
CN105100894A (en) * 2014-08-26 2015-11-25 Tcl集团股份有限公司 Automatic face annotation method and system
CN106960211A (en) * 2016-01-11 2017-07-18 北京陌上花科技有限公司 Key frame acquisition methods and device
CN110175267A (en) * 2019-06-04 2019-08-27 黑龙江省七星农场 A kind of agriculture Internet of Things control processing method based on unmanned aerial vehicle remote sensing technology
CN110852289A (en) * 2019-11-16 2020-02-28 公安部交通管理科学研究所 Method for extracting information of vehicle and driver based on mobile video
US11265317B2 (en) 2015-08-05 2022-03-01 Kyndryl, Inc. Security control for an enterprise network

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101201822B (en) * 2006-12-11 2010-06-23 南京理工大学 Method for searching visual lens based on contents

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100399804C (en) * 2005-02-15 2008-07-02 乐金电子(中国)研究开发中心有限公司 Mobile communication terminal capable of briefly offering activity video and its abstract offering method
CN100573523C (en) * 2006-12-30 2009-12-23 中国科学院计算技术研究所 A kind of image inquiry method based on marking area
CN101211355B (en) * 2006-12-30 2010-05-19 中国科学院计算技术研究所 Image inquiry method based on clustering
CN101339615B (en) * 2008-08-11 2011-05-04 北京交通大学 Method of image segmentation based on similar matrix approximation
WO2011140783A1 (en) * 2010-05-14 2011-11-17 中兴通讯股份有限公司 Method and mobile terminal for realizing video preview and retrieval
US8737808B2 (en) 2010-05-14 2014-05-27 Zte Corporation Method and mobile terminal for previewing and retrieving video
CN101968797A (en) * 2010-09-10 2011-02-09 北京大学 Inter-lens context-based video concept labeling method
CN105100894A (en) * 2014-08-26 2015-11-25 Tcl集团股份有限公司 Automatic face annotation method and system
CN105100894B (en) * 2014-08-26 2020-05-05 Tcl科技集团股份有限公司 Face automatic labeling method and system
CN104217000A (en) * 2014-09-12 2014-12-17 黑龙江斯迪克信息科技有限公司 Content-based video retrieval system
US11265317B2 (en) 2015-08-05 2022-03-01 Kyndryl, Inc. Security control for an enterprise network
US11757879B2 (en) 2015-08-05 2023-09-12 Kyndryl, Inc. Security control for an enterprise network
CN106960211A (en) * 2016-01-11 2017-07-18 北京陌上花科技有限公司 Key frame acquisition methods and device
CN106960211B (en) * 2016-01-11 2020-04-14 北京陌上花科技有限公司 Key frame acquisition method and device
CN110175267A (en) * 2019-06-04 2019-08-27 黑龙江省七星农场 A kind of agriculture Internet of Things control processing method based on unmanned aerial vehicle remote sensing technology
CN110852289A (en) * 2019-11-16 2020-02-28 公安部交通管理科学研究所 Method for extracting information of vehicle and driver based on mobile video

Also Published As

Publication number Publication date
CN1240014C (en) 2006-02-01

Similar Documents

Publication Publication Date Title
ElAlami A novel image retrieval model based on the most relevant features
CN108280187B (en) Hierarchical image retrieval method based on depth features of convolutional neural network
Liu et al. A new approach to retrieve video by example video clip
CN1240014C (en) Method for making video search of scenes based on contents
CN101556600B (en) Method for retrieving images in DCT domain
CN102890700A (en) Method for retrieving similar video clips based on sports competition videos
CN107229710A (en) A kind of video analysis method accorded with based on local feature description
CN106777159B (en) Video clip retrieval and positioning method based on content
CN109086830B (en) Typical correlation analysis near-duplicate video detection method based on sample punishment
CN111177435B (en) CBIR method based on improved PQ algorithm
CN100507910C (en) Method of searching lens integrating color and sport characteristics
Priya et al. Optimized content based image retrieval system based on multiple feature fusion algorithm
CN110162654A (en) It is a kind of that image retrieval algorithm is surveyed based on fusion feature and showing for search result optimization
Kumar et al. Automatic feature weight determination using indexing and pseudo-relevance feedback for multi-feature content-based image retrieval
CN1477600A (en) Scene-searching method based on contents
Sun et al. A novel region-based approach to visual concept modeling using web images
Semela et al. KIT at MediaEval 2012-Content-based Genre Classification with Visual Cues.
Goswami et al. RISE: a robust image search engine
Wang et al. Evaluation of global descriptors for large scale image retrieval
Mojsilovic et al. Psychophysical approach to modeling image semantics
Popova et al. Image similarity search approach based on the best features ranking
電子工程學系 Content based image retrieval using MPEG-7 dominant color descriptor
CN102436487A (en) Optical flow method based on video retrieval system
Alqaraleh et al. A Comparison Study on Image Content Based Retrieval Systems
Koskela et al. Semantic concept detection from news videos with self-organizing maps

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20060201

Termination date: 20170718