CN1477566A - Method for making video search of scenes based on contents - Google Patents
Method for making video search of scenes based on contents Download PDFInfo
- Publication number
- CN1477566A CN1477566A CNA031501265A CN03150126A CN1477566A CN 1477566 A CN1477566 A CN 1477566A CN A031501265 A CNA031501265 A CN A031501265A CN 03150126 A CN03150126 A CN 03150126A CN 1477566 A CN1477566 A CN 1477566A
- Authority
- CN
- China
- Prior art keywords
- sigma
- camera lens
- similarity
- fuzzy
- carried out
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 61
- 239000011159 matrix material Substances 0.000 claims description 22
- 238000007621 cluster analysis Methods 0.000 claims description 9
- 238000010606 normalization Methods 0.000 claims description 5
- 238000005259 measurement Methods 0.000 claims description 4
- 238000004458 analytical method Methods 0.000 abstract description 2
- 238000004519 manufacturing process Methods 0.000 abstract description 2
- 238000004220 aggregation Methods 0.000 abstract 1
- 230000002776 aggregation Effects 0.000 abstract 1
- 238000005516 engineering process Methods 0.000 description 7
- 230000000694 effects Effects 0.000 description 6
- 238000002474 experimental method Methods 0.000 description 6
- 230000009182 swimming Effects 0.000 description 4
- 238000009826 distribution Methods 0.000 description 3
- 238000000605 extraction Methods 0.000 description 3
- 208000018375 cerebral sinovenous thrombosis Diseases 0.000 description 2
- 230000002950 deficient Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 238000009825 accumulation Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000001186 cumulative effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 238000003860 storage Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Landscapes
- Image Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The present invention relates to a method of making video search of scene based on contents. It uses the fuzzy aggregation analysis method in the scene search. As compared with existent method it can obtain higher accuracy and quick searching speed.
Description
Technical field
The invention belongs to video search technique area, be specifically related to a kind of method of camera lens being carried out Content-based Video Retrieval.
Background technology
Along with in the great technical progress that obtains aspect multi-medium data manufacturing, storage and the propagation, digital video has become a part indispensable in people's daily life.The problem that people face no longer is to lack content of multimedia, but how to find own needed information in the vast as the open sea multimedia world.At present, traditional video frequency searching of describing based on keyword is because descriptive power is limited, and subjectivity is strong, reasons such as manual mark, demand that can not the satisfying magnanimity video frequency searching.In order to be convenient for people to seek multi-medium data, begin the nineties in last century, content-based video analysis and retrieval technique become the hot issue of research, and Multimedia Content Description Interface MPEG-7 progressively formulates and optimizes, and have promoted the development of Content-based Video Retrieval technology more.
In the prior art, as document " A New Approach to Retrieval Video by ExampleVideo Clip " [X.M.Liu, Y.T.Zhuang, and Y.H.Pan, ACM Multimedia, pp.41-44,1999] described, the conventional method of video frequency searching is at first to carry out shot boundary to detect, with basic structural unit and the retrieval unit of camera lens as video sequence; Represent the content of this camera lens then at each camera lens internal extraction key frame, go out low-level features such as color and texture from key-frame extraction, be used for the index and the retrieval of camera lens.Like this, just content-based searching lens being converted into CBIR solves.The problem that these class methods exist is that camera lens is an image continuous sequence in time, temporal information and the movable information that is present in the video is not fully utilized.(document author is s.H.Kim and R.-H.Park to the document of delivering at IEEE Trans.Circuits and Systems for Video Technology in 2002 " An efficient algorithm forvideo sequence matching using the modified Hausdorff distance and the directeddivergence " in addition, vol.CSVT-12, no.7, page number 592-595) disperses (Cumulative Directed Divergence) method with the orientation of accumulation and extract key frame, obtain two similarity degrees between the camera lens with improved Hausdorff distance (Modified Hausdorff Distance) method, used the YUV color space histogram when extracting key frame and definition camera lens similarity.Owing to set two threshold values when extracting key frame: the threshold value of similar value between the threshold value of front and back frame similar value and present frame and the previous key frame, these two conditions must be satisfied simultaneously and a key frame could be occurred, will influence the accuracy of key-frame extraction like this, finally will certainly influence the correctness of inquiry; In addition, used the YUV color space of using always in the video as visual signature, it is compared with the hsv color space, and is also little consistent with people's visually-perceptible.
Summary of the invention
At the existing defective of existing searching lens method, the objective of the invention is to propose a kind of method of camera lens being carried out Content-based Video Retrieval, this method can improve the accuracy rate of content-based searching lens on the basis of existing technology greatly, keep very fast retrieval rate simultaneously, thereby bring into play the huge effect of searching lens technology in current network information society more fully.
The object of the present invention is achieved like this: a kind of camera lens is carried out the method for Content-based Video Retrieval, may further comprise the steps:
(1) at first video database is carried out camera lens and cut apart, with the basic structural unit and the retrieval unit of camera lens as video;
(2) similarity between two two field pictures of calculating is set up fuzzy similarity matrix R by following method: when i=j, make r
IjBe 1; When i ≠ j, make r
IjBe x
iWith y
jBetween similarity;
(3) utilize the transitive closure method to calculate the equivalent matrice of fuzzy similarity matrix R
(4) threshold value λ is set and determines cut set, the transitive closure matrix of R matrix
Carry out fuzzy clustering, calculate
Set [x] is the equivalence class of fuzzy clustering, and each frame is similar in each equivalence class set, so we can get in each set arbitrary frame as key frame;
(5) with key frame { r
I1, r
I2..., r
IkExpression camera lens s
i, gather with key frame and to measure two similaritys between the camera lens.
Furthermore, in the step (1) video database is carried out the method that camera lens cuts apart and be preferably space-time section algorithm.Calculate x in the step (2)
iWith y
jBetween similarity can calculate with the friendship of two image histograms:
H
i(h, s v) are the histograms in hsv color space, and we use H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as color feature value, Intersect (x
i, y
j) two histogrammic friendships of expression, judge the similarity of two key frames with it, use A (x
i, y
j) normalize between 0,1.
Further again, in the step (3), calculate the equivalent matrice of fuzzy similarity matrix R
The transitive closure method can adopt quadratic method:
Its time complexity is O (n
3Log
2N), if the n value is big especially, will certainly influence total computing time, so adopt the compose operation of the fuzzy clustering optimal algorithm compute matrix that calculates based on figure connected component, recursion is as follows:
The time complexity T (n) of this algorithm satisfies O (n)≤T (n)≤O (n
2).
In order to realize purpose of the present invention better, when carrying out searching lens, right
The method of carrying out fuzzy clustering is as follows:
(1) determines n sample X=(X
1..., X
n) on fuzzy resembling relation R and a cut set threshold alpha;
(2) transform R as an equivalent matrice by following calculating;
RoR=R
2
R
2oR
2=R
4
...
Up to existing a k to satisfy
In the above-mentioned formula, RoR is the compose operation of fuzzy relation, is under the hypothesis of similar matrix at R, and having proved to have such k to exist, and satisfies k≤log n;
(3) set of computations
[x] is fuzzy clustering, and algorithm finishes;
After n sample space carried out fuzzy cluster analysis, obtain several equivalence classes, in each equivalence class, choose a sample as key frame.Measuring similarity between such two camera lenses just becomes the similarity measurement between the key frame set.
In the step (5) of this method, can be camera lens s
iAnd s
jSimilarity be defined as
M represents the maximal value that key frame is similar,
The similar second largest value of expression key frame, wherein,
Effect of the present invention is: adopt and of the present invention camera lens is carried out the method for Content-based Video Retrieval, can obtain higher accuracy rate, keep very fast retrieval rate simultaneously.
Why the present invention has so significant technique effect, its reason is: the method for utilization fuzzy cluster analysis, the camera lens content is divided into a plurality of equivalence classes, these equivalence classes have well been described the camera lens content change, and the similarity between the camera lens then shows as the similarity between the key frame combination.Similarity measurement has been considered to use the hsv color histogram to represent the shortcoming of key frame between the camera lens: if two key frames have similar color distribution, even their content is different, can think that also these two key frames are similar.So the mean value that uses maximum similar value and second largest similar value adds the robustness of strong algorithms.Contrast and experiment has confirmed that the present invention proposes the validity of method.
Description of drawings
Fig. 1 is the schematic flow sheet that camera lens is carried out the method for Content-based Video Retrieval;
Fig. 2 is 7 semantic category example synoptic diagram of searching lens in the experiment contrast;
Fig. 3 is the result for retrieval synoptic diagram of method of the present invention to the swimming camera lens.
Embodiment
Fig. 1 is an overall framework of the present invention, is the schematic flow sheet of each one step process among the present invention.As shown in Figure 1, a kind of searching lens method based on fuzzy cluster analysis may further comprise the steps:
1, camera lens is cut apart
At first use space-time section algorithm (spatio-temporal slice), video database is carried out camera lens to be cut apart, with basic structural unit and the retrieval unit of camera lens as video, can list of references " Video Partitioning by Temporal Slice Coherency " [C.W.Ngo about the detailed description of space-time section algorithm, T.C.Pong, and R.T.Chin, IEEE Transactions on Circuits andSystems for Video Technology, Vol.11, No.8, pp.941-953, August, 2001].
2, set up fuzzy similarity matrix R
Set up between the camera lens internal image to set up fuzzy similarity matrix R method as follows: when i=j, make r
IjBe 1, when i ≠ j, make r
IjBe x
iWith y
jBetween similarity, similarity then adopts following method to calculate:
H
i(h, s v) are the histograms in hsv color space, and we use H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as color feature value.Intersect (x
i, y
j) two histogrammic friendships of expression, judge the similarity of two key frames with it, use A (x
i, y
j) normalize between 0,1.
3, ask the transitive closure of similar matrix R, obtain equivalent matrice
In the present embodiment, ask the transitive closure of similar matrix to adopt quadratic method:
Its time complexity is O (n
3Log
2N), if the n value is big especially, will certainly influence total computing time.So adopt the compose operation of the fuzzy clustering optimal algorithm compute matrix that calculates based on figure connected component, recursion is as follows:
The time complexity T (n) of this algorithm satisfies O (n)≤T (n)≤O (n
2).
4, threshold value λ is set and determines cut set, the transitive closure matrix of R matrix
Carry out fuzzy clustering.
In the present embodiment, concrete grammar is as follows:
(1) determines n sample X=(x
1..., x
n) on fuzzy resembling relation R and a cut set threshold alpha;
(2) transform R as an equivalent matrice by following calculating;
RoR=R
2
R
2oR
2=R
4
...
Up to existing a k to satisfy
In the above-mentioned formula, RoR is the compose operation of fuzzy relation, is under the hypothesis of similar matrix at R, and having proved to have such k to exist, and satisfies k≤log n;
(3) set of computations
[x] is fuzzy clustering, and algorithm finishes
5, obtain the camera lens key frame with Fuzzy Cluster Analysis method after, carry out searching lens based on these key frames then.On this basis, with key frame { r
I1, r
I2..., r
Ik) the expression camera lens, s
iCamera lens s
iAnd s
jSimilarity be defined as
Wherein,
Represent second largest value, use
Be that its shortcoming is if two key frames have similar color distribution, even their content is different, can think that also these two key frames are similar because this paper uses the hsv color histogram to represent key frame, in order to overcome this defective, use M and
Mean value add the robustness of strong algorithms.H
i(h, s v) are the histograms in hsv color space, this paper H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as color feature value.Intersect (r
i, r
j) two histogrammic friendships of expression, this paper judges the similarity of two key frames with it.
Following experimental result shows that the present invention has obtained than the better effect of existing method, and retrieval rate is very fast simultaneously, has confirmed the validity of fuzzy cluster analysis algorithm in searching lens.
The experimental data of searching lens is the Asian Games programs in 2002 from television recording, always has 41 minutes, 777 camera lenses, 62132 two field pictures.It comprises multiple sports items, as various ball game, weight lifting, swim and the advertising programme that intercuts etc.We have selected 7 semantic categories as the inquiry camera lens, and they are weight lifting, vollyball, swimming, judo, row the boat, gymnastics, football, as shown in Figure 2.
In order to verify validity of the present invention, we have tested following 3 kinds of methods and have done the experiment contrast:
(1) the first frame of Chang Yong each camera lens of use is done the searching lens algorithm of key frame;
(document author is s.H.Kim and R.-H.Park to the document of delivering at IEEE Trans.Circuits and Systems for Video Technology in (2) 2002 years " An efficient algorithm for video sequence matching using the modifiedHausdorffdistance and the directed divergence ", vo1.CSVT-12, no.7, page number 592-595) the middle algorithm of describing;
(3) use the fuzzy cluster analysis algorithm to obtain key frame and carry out searching lens (only using color characteristic);
Above-mentioned preceding 3 kinds of methods have all only been used color characteristic, and therefore last experimental result can prove the superiority of the disclosed method of the present invention from the measure of shot similarity.Fig. 3 has provided the user interface of experimental arrangement, delegation is the browsing area of inquiry video above the right, and the 1st of each camera lens the key frame is used for representing each camera lens in the display video, the camera lens that the user can therefrom select to want to inquire about is retrieved, and is the Query Result zone below the right.Fig. 3 is the 1st camera lens of delegation above selecting, it is a swimming camera lens, and the first two field picture 022430.bmp represents by this camera lens, the authority of the similarity that calculates according to method of the present invention, arrange Query Result (from left to right, arranging from top to bottom) from big to small.Below, the left side is a simple and easy broadcast phase, double-clicks that section video that the result for retrieval image can be play corresponding camera lens correspondence.
Two kinds of evaluation indexes in the MPEG-7 standardization activity have been adopted in experiment: the average adjusted retrieval order of normalization ANMRR (average normalized modified retrieval rank) and recall level average AR (average recall).AR is similar to traditional recall ratio (recall), and ANMRR compares with traditional precision ratio (precision), not only can reflect correct result for retrieval ratio, and can reflect correct result's arrangement sequence number.The ANMRR value is more little, means that the rank of the correct camera lens that retrieval obtains is forward more; The AR value is big more, and it is big more to mean that in the individual Query Result of preceding K (K is the cutoff value of result for retrieval) similar camera lens accounts for the ratio of all similar camera lenses.Table 1 is AR and the ANMRR comparisons of above-mentioned 3 kinds of methods to 7 semantic camera lens classes.
Table 1 the present invention and the contrast and experiment that has two kinds of methods now
Classification | Method 1 | Method 2 | Method 3 | |||
????AR | ????ANMRR | ????AR | ????ANMRR | ????AR | ????ANMRR | |
Weight lifting | ????0.8824 | ????0.3098 | ????0.8824 | ????0.1539 | ????0.9412 | ????0.2186 |
Vollyball | ????0.6333 | ????0.4974 | ????0.7895 | ????0.3264 | ????0.8556 | ????0.3279 |
Swimming | ????0.8400 | ????0.2676 | ????0.8250 | ????0.3164 | ????0.9200 | ????0.2175 |
Judo | ????0.7000 | ????0.4310 | ????0.8214 | ????0.2393 | ????0.8000 | ????0.3093 |
Row the boat | ????0.8750 | ????0.3407 | ????0.6875 | ????0.3570 | ????0.8125 | ????0.2223 |
Gymnastics | ????0.7857 | ????0.3445 | ????0.9600 | ????0.1759 | ????0.7857 | ????0.2056 |
Football | ????0.5789 | ????0.4883 | ????0.6889 | ????0.2815 | ????0.8421 | ????0.2614 |
Average | ????0.7565 | ????0.3827 | ????0.8078 | ????0.2642 | ????0.8510 | ????0.2518 |
No matter as can be seen from Table 1, adopt method of the present invention, be AR, or ANMRR, all obtained than existing two kinds of better effects of algorithm, confirmed that the present invention is used for the Fuzzy Cluster Analysis method method validity of searching lens.The method of method utilization fuzzy cluster analysis of the present invention is divided into a plurality of equivalence classes to the camera lens content, and these equivalence classes have well been described the camera lens content change, and the similarity between the camera lens then shows as the similarity between the key frame combination.Similarity measurement has been considered to use the hsv color histogram to represent the shortcoming of key frame between the camera lens: if two key frames have similar color distribution, even their content is different, can think that also these two key frames are similar.So the mean value that uses maximum similar value and second largest similar value adds the robustness of strong algorithms.Contrast and experiment has confirmed that the present invention proposes the validity of method.In addition, at CPU 500M PIII, on the PC of 256 MB of memory, algorithm ART of the present invention is 22.557 seconds, and for the video library of 777 camera lenses, the retrieval rate of two kinds of algorithms of the present invention all is very fast.
Claims (6)
1, a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that this method may further comprise the steps:
(1) at first video database is carried out camera lens and cut apart, with the basic structural unit and the retrieval unit of camera lens as video;
(2) similarity between two two field pictures of calculating is set up fuzzy similarity matrix R by following method: when i=j, make r
IjBe 1; When i ≠ j, make r
IjBe x
iWith y
j, between similarity;
(3) utilize the transitive closure method to calculate the equivalent matrice of fuzzy similarity matrix R
(4) threshold value λ is set and determines cut set, the transitive closure matrix of R matrix
Carry out fuzzy clustering, calculate
Set [x] is the equivalence class of fuzzy clustering, and each frame is similar in each equivalence class set, so we can get in each set arbitrary two field picture as key frame;
(5) with key frame { r
I1, r
I2..., r
IkExpression camera lens s
i, key frame is gathered and is measured two similaritys between the camera lens.
2, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: in the step (1), it is space-time section algorithm that video database is carried out the method that camera lens cuts apart.
3, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: in the step (2), calculate x
iWith y
jBetween similarity can calculate with the friendship of two image histograms:
H
i(h, s v) are the histograms in hsv color space, and we use H, S, the V component is statistic histogram in 18 * 3 * 3 three dimensions, with 162 numerical value after the normalization as face color characteristic value, Intersect (x
i, y
j) two histogrammic friendships of expression, judge the similarity of two key frames with it, use A (x
i, y
j) normalize between 0,1.
4, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: in the step (3), calculate the equivalent matrice of fuzzy similarity matrix R
The transitive closure method adopt quadratic method:
Its time complexity is O (n
3Log
2N), if the n value is big especially, will certainly influence total computing time, so adopt the compose operation of the fuzzy clustering optimal algorithm compute matrix that calculates based on figure connected component, recursion is as follows:
The time complexity T (n) of this algorithm satisfies O (n)≤T (n)≤O (n
2).
5, as claimed in claim 1ly a kind of camera lens is carried out the method for Content-based Video Retrieval, it is characterized in that: right
The method of carrying out fuzzy clustering is as follows:
(1) determines n sample x=(x
1..., x
n) on fuzzy resembling relation R and a cut set threshold alpha;
(2) transform R as an equivalent matrice by following calculating
RoR=R
2
R
2oR
2=R
4
...?
Up to existing a k to satisfy
In the above-mentioned formula, RoR is the compose operation of fuzzy relation, is under the hypothesis of similar matrix at R, and having proved to have such k to exist, and satisfies k≤log n;
(3) set of computations
[x] is fuzzy clustering, and algorithm finishes;
After n sample space carried out fuzzy cluster analysis, obtain several equivalence classes, choose a sample as key frame in each equivalence class, the measuring similarity between such two camera lenses just becomes the similarity measurement between the key frame set.
6, describedly a kind of camera lens is carried out the method for Content-based Video Retrieval as claim 1 or 5, it is characterized in that: can be camera lens s
iAnd s
jSimilarity be defined as
M represents the maximal value that key frame is similar,
The similar second largest value of expression key frame, wherein,
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 03150126 CN1240014C (en) | 2003-07-18 | 2003-07-18 | Method for making video search of scenes based on contents |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 03150126 CN1240014C (en) | 2003-07-18 | 2003-07-18 | Method for making video search of scenes based on contents |
Publications (2)
Publication Number | Publication Date |
---|---|
CN1477566A true CN1477566A (en) | 2004-02-25 |
CN1240014C CN1240014C (en) | 2006-02-01 |
Family
ID=34156438
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN 03150126 Expired - Fee Related CN1240014C (en) | 2003-07-18 | 2003-07-18 | Method for making video search of scenes based on contents |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN1240014C (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN100399804C (en) * | 2005-02-15 | 2008-07-02 | 乐金电子(中国)研究开发中心有限公司 | Mobile communication terminal capable of briefly offering activity video and its abstract offering method |
CN100573523C (en) * | 2006-12-30 | 2009-12-23 | 中国科学院计算技术研究所 | A kind of image inquiry method based on marking area |
CN101211355B (en) * | 2006-12-30 | 2010-05-19 | 中国科学院计算技术研究所 | Image inquiry method based on clustering |
CN101968797A (en) * | 2010-09-10 | 2011-02-09 | 北京大学 | Inter-lens context-based video concept labeling method |
CN101339615B (en) * | 2008-08-11 | 2011-05-04 | 北京交通大学 | Method of image segmentation based on similar matrix approximation |
WO2011140783A1 (en) * | 2010-05-14 | 2011-11-17 | 中兴通讯股份有限公司 | Method and mobile terminal for realizing video preview and retrieval |
CN104217000A (en) * | 2014-09-12 | 2014-12-17 | 黑龙江斯迪克信息科技有限公司 | Content-based video retrieval system |
CN105100894A (en) * | 2014-08-26 | 2015-11-25 | Tcl集团股份有限公司 | Automatic face annotation method and system |
CN106960211A (en) * | 2016-01-11 | 2017-07-18 | 北京陌上花科技有限公司 | Key frame acquisition methods and device |
CN110175267A (en) * | 2019-06-04 | 2019-08-27 | 黑龙江省七星农场 | A kind of agriculture Internet of Things control processing method based on unmanned aerial vehicle remote sensing technology |
CN110852289A (en) * | 2019-11-16 | 2020-02-28 | 公安部交通管理科学研究所 | Method for extracting information of vehicle and driver based on mobile video |
US11265317B2 (en) | 2015-08-05 | 2022-03-01 | Kyndryl, Inc. | Security control for an enterprise network |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101201822B (en) * | 2006-12-11 | 2010-06-23 | 南京理工大学 | Method for searching visual lens based on contents |
-
2003
- 2003-07-18 CN CN 03150126 patent/CN1240014C/en not_active Expired - Fee Related
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN100399804C (en) * | 2005-02-15 | 2008-07-02 | 乐金电子(中国)研究开发中心有限公司 | Mobile communication terminal capable of briefly offering activity video and its abstract offering method |
CN100573523C (en) * | 2006-12-30 | 2009-12-23 | 中国科学院计算技术研究所 | A kind of image inquiry method based on marking area |
CN101211355B (en) * | 2006-12-30 | 2010-05-19 | 中国科学院计算技术研究所 | Image inquiry method based on clustering |
CN101339615B (en) * | 2008-08-11 | 2011-05-04 | 北京交通大学 | Method of image segmentation based on similar matrix approximation |
WO2011140783A1 (en) * | 2010-05-14 | 2011-11-17 | 中兴通讯股份有限公司 | Method and mobile terminal for realizing video preview and retrieval |
US8737808B2 (en) | 2010-05-14 | 2014-05-27 | Zte Corporation | Method and mobile terminal for previewing and retrieving video |
CN101968797A (en) * | 2010-09-10 | 2011-02-09 | 北京大学 | Inter-lens context-based video concept labeling method |
CN105100894A (en) * | 2014-08-26 | 2015-11-25 | Tcl集团股份有限公司 | Automatic face annotation method and system |
CN105100894B (en) * | 2014-08-26 | 2020-05-05 | Tcl科技集团股份有限公司 | Face automatic labeling method and system |
CN104217000A (en) * | 2014-09-12 | 2014-12-17 | 黑龙江斯迪克信息科技有限公司 | Content-based video retrieval system |
US11265317B2 (en) | 2015-08-05 | 2022-03-01 | Kyndryl, Inc. | Security control for an enterprise network |
US11757879B2 (en) | 2015-08-05 | 2023-09-12 | Kyndryl, Inc. | Security control for an enterprise network |
CN106960211A (en) * | 2016-01-11 | 2017-07-18 | 北京陌上花科技有限公司 | Key frame acquisition methods and device |
CN106960211B (en) * | 2016-01-11 | 2020-04-14 | 北京陌上花科技有限公司 | Key frame acquisition method and device |
CN110175267A (en) * | 2019-06-04 | 2019-08-27 | 黑龙江省七星农场 | A kind of agriculture Internet of Things control processing method based on unmanned aerial vehicle remote sensing technology |
CN110852289A (en) * | 2019-11-16 | 2020-02-28 | 公安部交通管理科学研究所 | Method for extracting information of vehicle and driver based on mobile video |
Also Published As
Publication number | Publication date |
---|---|
CN1240014C (en) | 2006-02-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
ElAlami | A novel image retrieval model based on the most relevant features | |
CN108280187B (en) | Hierarchical image retrieval method based on depth features of convolutional neural network | |
Liu et al. | A new approach to retrieve video by example video clip | |
CN1240014C (en) | Method for making video search of scenes based on contents | |
CN101556600B (en) | Method for retrieving images in DCT domain | |
CN102890700A (en) | Method for retrieving similar video clips based on sports competition videos | |
CN107229710A (en) | A kind of video analysis method accorded with based on local feature description | |
CN106777159B (en) | Video clip retrieval and positioning method based on content | |
CN109086830B (en) | Typical correlation analysis near-duplicate video detection method based on sample punishment | |
CN111177435B (en) | CBIR method based on improved PQ algorithm | |
CN100507910C (en) | Method of searching lens integrating color and sport characteristics | |
Priya et al. | Optimized content based image retrieval system based on multiple feature fusion algorithm | |
CN110162654A (en) | It is a kind of that image retrieval algorithm is surveyed based on fusion feature and showing for search result optimization | |
Kumar et al. | Automatic feature weight determination using indexing and pseudo-relevance feedback for multi-feature content-based image retrieval | |
CN1477600A (en) | Scene-searching method based on contents | |
Sun et al. | A novel region-based approach to visual concept modeling using web images | |
Semela et al. | KIT at MediaEval 2012-Content-based Genre Classification with Visual Cues. | |
Goswami et al. | RISE: a robust image search engine | |
Wang et al. | Evaluation of global descriptors for large scale image retrieval | |
Mojsilovic et al. | Psychophysical approach to modeling image semantics | |
Popova et al. | Image similarity search approach based on the best features ranking | |
電子工程學系 | Content based image retrieval using MPEG-7 dominant color descriptor | |
CN102436487A (en) | Optical flow method based on video retrieval system | |
Alqaraleh et al. | A Comparison Study on Image Content Based Retrieval Systems | |
Koskela et al. | Semantic concept detection from news videos with self-organizing maps |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20060201 Termination date: 20170718 |