CN101504775B - Roaming video automatic generation method based on image set - Google Patents
Roaming video automatic generation method based on image set Download PDFInfo
- Publication number
- CN101504775B CN101504775B CN2009100968558A CN200910096855A CN101504775B CN 101504775 B CN101504775 B CN 101504775B CN 2009100968558 A CN2009100968558 A CN 2009100968558A CN 200910096855 A CN200910096855 A CN 200910096855A CN 101504775 B CN101504775 B CN 101504775B
- Authority
- CN
- China
- Prior art keywords
- image
- embedded
- difference
- images
- video
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 51
- 238000005457 optimization Methods 0.000 claims abstract description 7
- 230000008878 coupling Effects 0.000 claims description 10
- 238000010168 coupling process Methods 0.000 claims description 10
- 238000005859 coupling reaction Methods 0.000 claims description 10
- 230000003321 amplification Effects 0.000 claims description 5
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 5
- 239000004744 fabric Substances 0.000 claims description 4
- 238000012545 processing Methods 0.000 claims description 3
- 238000005516 engineering process Methods 0.000 claims description 2
- 238000012423 maintenance Methods 0.000 claims description 2
- 239000000203 mixture Substances 0.000 claims description 2
- 239000003086 colorant Substances 0.000 claims 1
- 230000015572 biosynthetic process Effects 0.000 abstract description 3
- 238000003786 synthesis reaction Methods 0.000 abstract description 3
- 238000013461 design Methods 0.000 abstract description 2
- 239000000463 material Substances 0.000 abstract description 2
- 230000000007 visual effect Effects 0.000 abstract description 2
- 238000004519 manufacturing process Methods 0.000 abstract 1
- 238000007619 statistical method Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000002194 synthesizing effect Effects 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000005465 channeling Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 238000012113 quantitative test Methods 0.000 description 1
- 238000002922 simulated annealing Methods 0.000 description 1
- 230000003595 spectral effect Effects 0.000 description 1
- 238000010183 spectrum analysis Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Abstract
The invention provides a method for automatically generating roaming videos based on image sets, which comprises the following steps: firstly, local matching; secondly, global optimization; thirdly, image synthesis; and fourthly, connection of all the frames into a video for output. The method takes a series of images as input, finds out an optimum browsing sequence, and forms a continuously amplified image browsing video with visual continuity. The invention simultaneously provides a more vivid image browsing mode, which makes users browse images in a more vivid mode just like roaming in image scenes; as for a series of images of a tourist attraction, the method can generate a video for well connecting different scenes of the tourist attraction in series and take the video as a propaganda film of the tourist attraction; and the production method has high speed, does not require scene design and image selection with superabundant manpower and material resources, and has high automation degree in the whole video generation process.
Description
Technical field
The present invention relates to a kind of based on the image browsing method of amplifying form.
Background technology
Be under the jurisdiction of feeling of unreality image/video process field based on the image browsing method of amplifying form, at present online popular a kind of feeling of unreality artistic expression that is become zoomquilt: artist's Freehandhand-drawing one is an image, form a flash video that amplifies playing image, the video of broadcast visually is continuous.But make such flash, need the well-designed and drafting scene of artist, need to consume a large amount of manpowers.
Summary of the invention
Technical matters to be solved by this invention provides a kind of roaming video automatic generation method based on image set, and it can generate the picture browsing video automatically based on Computer Processing.The present invention is by the following technical solutions for this reason: this method may further comprise the steps:
(1) local coupling: establishing in the image set arbitrarily, piece image is an image A, another width of cloth image is an image B arbitrarily, after image B dwindled, compare with the diverse location image in the image A, calculate the image difference between them, seek in the described image A and the position of image difference minimum between the described image B of dwindling, as the embedded location of described image B of dwindling in image A, this minimum image difference deposits image difference table in as two image final image difference with this position; Repeat above-mentioned steps, the image in the image set is mated in twos, deposit the minimum image difference that obtains in image difference table;
(2) global optimization: system finds out the original image collating sequence of image difference summation minimum according to the image difference table that this locality coupling obtains, and as play sequence, wherein, the image of embedding comes the back of the image that is embedded into;
(3) image is synthetic: according to the play sequence that obtains, by succession, establish that original image is i, i+1, i+2 in the sequence ... i+ (n-1), i+n dwindle i+1 and are embedded among the i according to the embedded location that obtains in the coupling of this locality, amplify the composograph of i and i+1, be full of form until i+1; Repeat said process, carry out i+1 and i+2 until the processing of i+ (n-1) with i+n; The composograph of each amplification stage is all as the interpolation image between the original image of front and back, the frame of all images output video;
At last, all frames are connected into video output.
Based on the image browsing method of amplifying form, the image difference of definition comprises following key element: color distinction, texture difference, image complexity.
Color is the notable attribute of image, compares with other features, and color characteristic calculates simple, stable in properties, and is all insensitive for rotation, translation, dimensional variation, shows very strong robustness.Color characteristic comprises color histogram, main color, mean flow rate etc.This method uses the mean pixel difference square (a sum squared of difference) on the hsv color space to weigh color distinction
Texture analysis is an important research direction of computer vision always, and its method can roughly be divided into statistical method and structural approach.Statistical method is that the space distribution information of the color intensity of image is added up, can be further divided into again traditional based on model statistical method and based on the method for spectrum analysis, as Markov random field model, Fourier spectral characteristic etc.Structural approach at first suppose texture pattern by texture cell according to certain regularly arranged composition, so texture analysis just becomes and determines these unit, their spatial disposition of quantitative test.This method uses the mean pixel difference of gradient image square to weigh the texture difference.
The image complexity problem is the research hot issue of computer vision in recent years, human eye is difficult to pick out an object from complex scene, but be easy to go out an object respectively from simple scene, so the zone that downscaled images is embedded into a relative complex is a better choice.
The global optimization problem also is a problem that is widely studied, and has many methods to be used to solve optimization problem, such as simulated annealing, and ant group algorithm, dynamic programming algorithm also is applicable to these a series of problems of solution.
Figure cuts algorithm and is widely used in computer vision field in the past few years, to solve label problem, energy minimization problem or the like, concrete application comprises image segmentation, image mosaic, video-splicing or the like, aspect the image mosaic, figure cuts algorithm extraordinary effect especially.
Because adopt technical scheme of the present invention, the present invention as input, finds a best to browse sequence with a series of images, form one section picture browsing video that constantly amplifies with visual continuity, the present invention also has following beneficial effect simultaneously:
1. a kind of more lively picture browsing mode is provided, has made the user with a kind of more lively mode browse graph picture, just as in the scene of image, roaming.
2. for a series of images of tourist attractions, this method can generate one section better video of the different scenes in this sight spot of series connection, as the propaganda film of these tourist attractions.
3. method for making speed is very fast, does not need the manpower and materials of overspending to go design scenario and select image, whole video generative process automaticity height.
Description of drawings
Fig. 1 is the process flow diagram of automatic generation method provided by the present invention.
Fig. 2 is the sub-process of local coupling step provided by the present invention.
Fig. 3 is the sub-process figure of image synthesis step provided by the present invention.
Fig. 4 is the process prescription figure of automatic generation method provided by the present invention, has demonstrated image set as input, exports best play sequence, carries out the composograph interpolation between the adjacent picture in the play sequence, forms the process of output video.
Fig. 5 is the simplified schematic diagram that automatic generation method provided by the present invention is handled example, it in the drawings, belong to a part of image in an image play sequence of first row for arranging by sequence, second row connects the wherein interpolation intermediate frame of two adjacent images for generating in the video.
Embodiment
At first be defined in the symbol that will use in the next explanation: image sequence is defined as I
1..., I
nWith I
iBe defined as target image, with I
I+1Down-scaled version note be P
I+1I
iAnd P
I+1Composograph note be S
I, i+1Mean pixel difference square (sum squared of difference) is abbreviated as SSD.
Fig. 1 is a basic flow sheet of the present invention, and the present invention as input, constantly is amplified in the video that in image switch as output with one section with a series of images.
Below each flow process of the present invention is elaborated:
1. local coupling
The local coupling with a series of images as input (flow process as shown in Figure 2), substantially flow process is as follows for it: establishing in the image arbitrarily, piece image is an image A, another width of cloth image is an image B arbitrarily, after image B dwindled, compare with the diverse location image in the image A, calculate the image difference between them, seek in the described image A and the position of image difference minimum between the described image B of dwindling, as the embedded location of described image B of dwindling in image A, this minimum image difference deposits image difference table in as two image final image difference with this position; Repeat above-mentioned steps, the image in the image set is mated in twos, deposit the minimum image difference that obtains in image difference table
The image difference comprises four aspects: color distinction, texture difference, the image complexity of embedded location, the embedded location position in image.
HSV spatial color difference: Δ c=α D (H)+β D (S)+γ D (V), wherein α: β: γ=4: 1: 1 is weighted value, D (H), D (S), D (V) represent 3 mean pixel SSD on the passage respectively
The texture difference: Δ T is defined as the mean pixel SSD through two width of cloth gradient images of medium filtering
Image complexity: use variance V to weigh the complexity of object region
Embedded location: to the position p definition penalty value P at edge
t(p), compare with the embedded location of image border near the embedded location of image central authorities, its image difference value is little, its penalty value P
t(p) little, for position intermediate, this is 0.Also the embedded location of promptly close image central authorities gives less punishment, gives bigger punishment in the position of image border.
Comparison procedure is carried out according to shown in Figure 2, certain embedded location is at first compared color distinction, if color distinction is greater than some setting thresholds, then skip this position, go to other positions to compare, otherwise continue relatively texture difference, image complexity, three factors of embedded location are up to the comprehensive evaluation that obtains this position.
Comprehensive above-mentioned several, the down-scaled version (size is r) that i is opened image embeds the consumption that j opens image and is defined as follows:
Wherein p is an embedded location, and G is the set of all positions, w
1=0.4, w
2=w
3=w
4=0.2, be weighted value, all the other several the factors that will consider for foregoing description image difference, last, consider the different big or small of image i, the consumption that image i is pasted into image j is defined as follows:
Wherein R gathers (generally choosing 16 * 16,32 * 32,64 * 64) for all resolution that we consider.
(i j), can obtain storing the table Best Map (image difference table) of difference between all images at last to calculate all images Cost between any two.
2. global optimization
The image difference table that global optimization obtains with this locality coupling is as input, problem will produce an optimum play sequence, make and do not duplicate image in the calling sequence, and the image difference sum minimum between the adjacent image, this problem solves by dynamic programming, set up a two-dimentional form BestSeq, the element Be stSeq in the form (i, j) .C
tExpression length is the minimum wastage in bulk or weight with j the image sequence that is ending of i, and form is gone back the index of previous image in the storage sequence simultaneously, and the dynamic programming recursion equation is expressed as follows:
Wherein Cost (j, k) be from Best Map, check in j is pasted consumption into k, Penalty is in order to prevent the appearance of multiimage in the sequence, if image j occurs in sequence before, Penalty is exactly a very big numeral, otherwise Penalty is exactly 0.
After dynamic programming finished, last column found an element of wastage in bulk or weight minimum in form, recalls by index information, can obtain an optimal sequence.
3. image is synthetic
After obtaining optimum play sequence, two images in front and back in the sequence are synthesized (flow process is as shown in Figure 3) at each amplification stage, image synthesis strategy of the present invention, employing figure cuts channeling conduct as a result, obtains embedded images P
I+1And I
iThe best of synthesizing is synthesized the border, expands by this border, generates to comprise P
I+1In the nuclear K of all pixel alpha values.
The energy function that figure cuts is defined as follows:
E wherein
DataBe the single order energy, E
SmoothIt is the second order energy.α is a relative importance, and L (p) expression is to the label of p point mark, and this point of 1 expression adopts P
I+1Pixel, I is adopted in 0 expression
iPixel.
The single order energy is defined as follows:
Wherein, n is P
I+1Size, establish k=0.6, dist is the function that calculates two pixel distances, center is P
I+1Geometric center.
The second order energy is defined as follows:
E
smooth(p,q,L(p),L(q))=||L(p).p-L(q).p||+||L(p).q-L(q).q||
Wherein, L (a) .b represents that L (a) goes up the color value of pixel b, and L (a)=0 expression is a target image, and L (a)=1 expression is the image that dwindles.
Use the energy function of above-mentioned definition, employing max-flow method can be in the hope of embedded images P
I+1And I
iThe best of synthesizing is synthesized the border, and the border is divided into A and two parts of E with image.By this border, generation comprises P
I+1In the nuclear K of all pixel alpha values, it is as follows to generate principle:
● for the pixel among the regional A, given alpha value
● for the pixel that belongs among the E, carry out 4 in abutting connection with expansion, given alpha value: n/w-δ
*Step (p)
Wherein, n is current P
I+1Size, w is the form size, δ is a given constant, the step number that step (p) remarked pixel p arrives through expansion.
Obtain composograph at last:
S=KP+(1-K)T
r
T wherein
rRepresent the object region corresponding with P.
In fact might exist figure to cut the unfavorable result of result,, then abandon employing figure and cut the result, then use Gaussian function to generate alpha nuclear K if figure cuts and obtains regional A less than 1/4 of whole region area:
(x wherein
0, y
0) be the coordinate of P geometric center, σ
x 2And σ
y 2Be variance, n is current P
I+1Size, w is the form size, k is decided to be 4.5 in order to regulate the smoothness of nuclear at this.
4. relevance between maintenance frame
In sequence, work as P
I+1Filled up form, composograph is by S
I, i+1Transferred S to
I+1, i+2, will produce vision sudden change at this, take place for preventing this sudden change, the composograph of front is carried out mixing of several steps with the composograph of back, the adding transition frames:
Tran=αS
i,i+1+(1-α)S
i+1,i+2
Wherein Tran represents transition frames, and the initial value of α is set at 0.9, and each one amplifies its value is reduced 0.025, is decreased to 0 until its value.
5. generation video
Through above-mentioned plurality of processes, at a frame of each amplification procedure generation video, all these frames have been formed the amplified video of output.Fig. 5 has showed the frame that wherein partly connects adjacent image.
Claims (10)
1. roaming video automatic generation method based on image set is characterized in that this method may further comprise the steps:
(1) local coupling: establishing in the image set arbitrarily, piece image is an image A, another width of cloth image is an image B arbitrarily, after image B dwindled, compare with the diverse location image in the image A, calculate the image difference between them, seek in the described image A and the position of image difference minimum between the image B of dwindling, as the embedded location of described image B of dwindling in image A, the minimum image difference deposits image difference table in as two image final image difference with this position; Repeat above-mentioned steps, the image in the image set is mated in twos, deposit the minimum image difference that obtains in image difference table;
(2) global optimization: according to the image difference table that this locality coupling obtains, find out the image collating sequence of image difference summation minimum, as play sequence, wherein, the image of embedding comes the back of the image that is embedded into;
(3) image is synthetic: according to the play sequence that obtains, by succession, establish that original image is i, i+1, i+2 in the sequence ... i+ (n-1), i+n dwindle i+1 and are embedded among the i according to the embedded location that obtains in the coupling of this locality, amplify the composograph of i and i+1, be full of form until i+1; Repeat said process, carry out i+1 and i+2 until the processing of i+ (n-1) with i+n; The composograph of each amplification stage is all as the interpolation image between the original image of front and back, the frame of all images output video;
At last, all frames are connected into video output.
2. a kind of roaming video automatic generation method based on image set according to claim 1 is characterized in that described image difference comprises 2 aspects: color distinction, texture difference; Wherein, color distinction uses the mean pixel difference on the hsv color space square to weigh; The texture difference uses the mean pixel difference of gradient image square to weigh.
3. a kind of roaming video automatic generation method based on image set according to claim 2 is characterized in that described image difference comprises 2 aspects: the image complexity of embedded location, the embedded location position in image; Image complexity uses variance to weigh, and the image complexity of embedded location is big more, and the image difference value is more little; Embedded location near image central authorities is compared with the embedded location of image border, and its image difference value is little.
4. according to claim 2 or 3 described a kind of roaming video automatic generation methods based on image set, it is characterized in that when calculating described image difference, at first calculate color distinction, if color distinction is then abandoned this image difference relatively greater than given threshold value, otherwise, continue to calculate remaining Several Factors and compare fully.
5. a kind of roaming video automatic generation method according to claim 1 based on image set, it is characterized in that in step (1), described image A is divided into and the described identical non-intersect subregion of image B size that dwindles, described image B of dwindling and described subregion are carried out the image difference relatively.
6. a kind of roaming video automatic generation method based on image set according to claim 1 is characterized in that in step (2), uses dynamic programming method to obtain described play sequence; If in two images of the adjacent front and back of play sequence, previous image is for being embedded into image, the downscaled images of the image in back is an embedded images, in output information, also comprise the best initial reduction dimension information of the embedded images in the play sequence and the positional information that is embedded into that is embedded into image.
7. a kind of roaming video automatic generation method according to claim 1 based on image set, it is characterized in that establishing in two images of adjacent front and back of play sequence, previous image is for being embedded into image, the downscaled images of the image in back is an embedded images, in step (3), to the embedded images in the composograph of each amplification stage be embedded into image and carry out image co-registration and handle.
8. a kind of roaming video automatic generation method according to claim 6 based on image set, it is characterized in that described image co-registration is handled adopts following disposal route: its employing figure cuts technology, obtains embedded images and is embedded into the minimum border of energy that merge the embedding target area of image; Under the condition that obtains the minimum border of energy, the pixel that the border is embedded in image gives initial alpha value, does 4 in abutting connection with expansion on the border, the pixel alpha value linear decrease that per step expansion obtains; Finally obtain the nuclear K of the alpha value of embedded images; Use K, the alpha blend of colors is carried out in embedded images and the embedding target area that is embedded into image, obtain composograph.
9. a kind of roaming video automatic generation method based on image set according to claim 1 is characterized in that composograph exports as frame of video, and described method is safeguarded its interframe continuity when synthetic video.
10. a kind of roaming video automatic generation method according to claim 9 based on image set, it is characterized in that between described maintenance frame the continuity step adopt keep before composograph carry out the mode that alpha mixes with current composograph.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2009100968558A CN101504775B (en) | 2009-03-19 | 2009-03-19 | Roaming video automatic generation method based on image set |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2009100968558A CN101504775B (en) | 2009-03-19 | 2009-03-19 | Roaming video automatic generation method based on image set |
Publications (2)
Publication Number | Publication Date |
---|---|
CN101504775A CN101504775A (en) | 2009-08-12 |
CN101504775B true CN101504775B (en) | 2011-08-31 |
Family
ID=40977010
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2009100968558A Expired - Fee Related CN101504775B (en) | 2009-03-19 | 2009-03-19 | Roaming video automatic generation method based on image set |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN101504775B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111241899B (en) * | 2019-03-26 | 2021-01-12 | 广西三笔科技有限公司 | Intelligent river channel control method |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040179262A1 (en) * | 2002-11-25 | 2004-09-16 | Dynamic Digital Depth Research Pty Ltd | Open GL |
CN1662933A (en) * | 2002-05-24 | 2005-08-31 | 德耐皮克斯情报图象公司 | Method and apparatus for comprehensive and multi-scale 3D image documentation and navigation |
WO2007016596A2 (en) * | 2005-07-29 | 2007-02-08 | Pamela Barber | Digital imaging method and apparatus |
US20070150188A1 (en) * | 2005-05-27 | 2007-06-28 | Outland Research, Llc | First-person video-based travel planning system |
-
2009
- 2009-03-19 CN CN2009100968558A patent/CN101504775B/en not_active Expired - Fee Related
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1662933A (en) * | 2002-05-24 | 2005-08-31 | 德耐皮克斯情报图象公司 | Method and apparatus for comprehensive and multi-scale 3D image documentation and navigation |
US20040179262A1 (en) * | 2002-11-25 | 2004-09-16 | Dynamic Digital Depth Research Pty Ltd | Open GL |
US20070150188A1 (en) * | 2005-05-27 | 2007-06-28 | Outland Research, Llc | First-person video-based travel planning system |
WO2007016596A2 (en) * | 2005-07-29 | 2007-02-08 | Pamela Barber | Digital imaging method and apparatus |
Also Published As
Publication number | Publication date |
---|---|
CN101504775A (en) | 2009-08-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102360490B (en) | Color conversion and editing propagation-based method for enhancing seasonal feature of image | |
CN100541524C (en) | Content-based method for filtering internet cartoon medium rubbish information | |
CN102567727B (en) | Method and device for replacing background target | |
CN107452010A (en) | A kind of automatically stingy nomography and device | |
CN108509978A (en) | The multi-class targets detection method and model of multi-stage characteristics fusion based on CNN | |
CN104732506A (en) | Character picture color style converting method based on face semantic analysis | |
CN103208115B (en) | Based on the saliency method for detecting area of geodesic line distance | |
CN102959951A (en) | Image processing device, image processing method, image processing program, and integrated circuit | |
CN106991529A (en) | City night lights economic index evaluation method based on cross-domain multidimensional big data | |
CN105205846B (en) | Ink animation production method | |
CN102903128A (en) | Video image content editing and spreading method based on local feature structure keeping | |
CN105957078A (en) | Multi-view video segmentation method based on graph cut | |
CN101436310A (en) | Method for automatically generating middle frame during two-dimension cartoon making process | |
CN101216948A (en) | Cartoon animation fabrication method based on video extracting and reusing | |
CN101098475A (en) | Interactive time-space accordant video matting method in digital video processing | |
CN105787948A (en) | Quick graph cutting method based on multiple deformation resolutions | |
CN105447900A (en) | Animation recording method and device | |
CN103051915A (en) | Manufacture method and manufacture device for interactive three-dimensional video key frame | |
CN106023276A (en) | Pencil drawing making method and pencil drawing making device based on image processing | |
CN101329768A (en) | Method for rapidly modeling of urban street based on image sequence | |
CN101504775B (en) | Roaming video automatic generation method based on image set | |
Fish et al. | Image morphing with perceptual constraints and stn alignment | |
CN102542528B (en) | Image conversion processing method and system | |
Ye et al. | Hybrid scheme of image’s regional colorization using mask r-cnn and Poisson editing | |
Chen et al. | iSlideShow: a content-aware slideshow system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20110831 |