CA2242166C - Display techniques for three dimensional virtual reality - Google Patents

Display techniques for three dimensional virtual reality Download PDF

Info

Publication number
CA2242166C
CA2242166C CA002242166A CA2242166A CA2242166C CA 2242166 C CA2242166 C CA 2242166C CA 002242166 A CA002242166 A CA 002242166A CA 2242166 A CA2242166 A CA 2242166A CA 2242166 C CA2242166 C CA 2242166C
Authority
CA
Canada
Prior art keywords
video
world
represented
computer graphics
representation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CA002242166A
Other languages
French (fr)
Other versions
CA2242166A1 (en
Inventor
Gianpaolo U. Carraro
James Robert Ensor
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nokia of America Corp
Original Assignee
Lucent Technologies Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lucent Technologies Inc filed Critical Lucent Technologies Inc
Publication of CA2242166A1 publication Critical patent/CA2242166A1/en
Application granted granted Critical
Publication of CA2242166C publication Critical patent/CA2242166C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/10Geometric effects
    • A63F13/10
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/45Controlling the progress of the video game
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/60Methods for processing data by generating or executing the game program
    • A63F2300/66Methods for processing data by generating or executing the game program for rendering three dimensional images
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/80Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
    • A63F2300/8082Virtual reality

Abstract

A limitation of a three-dimensional world in which objects in the distance may be represented in only two dimensions as a video on a screen occurs when a computer graphic object represented by computer graphics, e.g., in front of, to the side, above, or below the video screen, undergoes a trajectory that takes it to a location in the world that is not represented as computer graphics, but instead is within the field represented by the video, because such an object would disappear from view by the viewer. This limitation is overcome by having such an object be represented as video on the screen, rather than computer graphics. Thus, the computer graphics object "goes into the video" as video and remains visible to a viewer in front of the video screen, rather than becoming invisible because it is blocked from view by the video screen if it were to be generated at its proper location using computer graphic techniques.

Description

DISPLAY TECHNIQUES FOR THREE DIMENSIONAL VIRTUAL REALITY
Technical Field This invention relates to the integration of computer graphics and video to provide a realistic three dimensional virtual reality experience.
Background of the Invention The display of a three dimensional world to a viewer requires considerable computation power, and it is typically costly to develop the necessary highly detailed models required for doing so. In order to simplify the problem, a portion of the world that is in the distance may be represented in only two dimensions as a video displayed on a surface, e.g., a screen. By video it is meant the common usage of the term, such as the placing or projecting of predefined images on the surface, e.g., the electronic version of filmed moving pictures. Thus, such a world is essentially truncated in length to the screen on which the video is displayed. A great reduction in computation power and cost can be achieved by such an arrangement.
Summary of the Invention A limitation of such a world occurs when a computer graphic object represented by computer graphics, e.g., in front of, to the side of, above, or below the video screen, undergoes a trajectory that takes it to a location in the world that is not represented as computer graphics, but instead is within the field represented by the video.
We have recognized that, in accordance with the principles of the invention, when a computer graphic object undergoes a trajectory that takes it to a location in the world that is not represented as computer graphics, but instead is within the field represented by the video, such an object should be represented as video on the screen, rather than computer graphics. Thus, the computer graphics object "goes into the video" as video and remains visible to a viewer in front of the video screen, rather than becoming invisible because it is blocked from view by the video screen if it were to be generated at its proper location using computer graphic techniques.
In accordance with one aspect of the present invention there is provided a method for use in processing a view of a three dimensional world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video display on a video screen which exists in said first portion of said world, wherein said video is made up of a sequence of images, the method comprising the steps o~
determining that an object, other than said video screen, is being modeled as computer graphics and has moved to a location that is being represented by said video and at which s said object is at least partly occluded from view of a viewer of said world by said video screen although said object should not appear occluded to said viewer when at said location;
clhanging the representation of said object from one modeled as computer graphics to a video representation thereof; and displaying said video representation so that it appears to said viewer to be substantially integrated with said video on said video screen.
1o In accordance with another aspect of the present invention there is provided a system fir displaying an apparently three dimensional world to a user, in which a first part of said three dimensional world is displayed as computer graphics and a second part is displayed as video on a video screen, wherein said video is made up of a sequence of images, said system comprising: a processor for determining that an element displayed in said first part of said 15 world transited from said first part to said second part by having moved to a location that is being represented by said second part and at which said element is at least partly occluded from view of a viewer of said world by a said video screen although said object should not appear occluded to said viewer when at said location; and a rendering engine for displaying said world with said element displayed so that it appears to said viewer to be substantially 2o integrated into said second part rather than said first part.
In accordance with yet another aspect of the present invention there is provided a system for displaying an apparently three dimensional world to a user, in which a first part of said three dimensional world is displayed as computer graphics and a second part is displayed as video on a video screen, wherein said video is made up of a sequence of images, said 25 system comprising: a processor for determining that a portion of an element initially in said first part of said world has moved to a location within said second part by virtue of having moved to a location that is being represented by said second part and at which portion of said element is at least partly occluded from view of a viewer of said world by said video screen although said object should not appear occluded to said viewer when at said location; and a 3o rendering engine for displaying said world with said element displayed so that it appears to said viewer to be substantially integrated into said second part rather than said first part.
-la-In accordance with still yet another aspect of the present invention there is provided an apparatus for use in processing a view of a three dimensional world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video, wherein said video is made up of a sequence of images, the apparatus s comprising: means for determining that an object being modeled as computer graphics has moved to a location that is being represented by said video by having moved to a location that is being represented by said second portion and at which said object is at least partly occluded from view of a viewer of said world by said video although said object should not appear occluded to said viewer when at said location; and means for changing the 1o representation of said object from one modeled as computer graphics to a video representation thereof displayed so that it appears to said viewer to be substantially integrated into said second portion rather than said first portion.
The above-identified apparatus may include means for displaying said video representation of said object merged with said portion of said world that is represented by 15 said video.
In accordance with still yet another aspect of the present invention there is provided an apparatus for use in processing a view of a world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video on a video surface which exists in said first portion of said world, wherein said video is made 20 up of a sequence of images, said apparatus comprising: means for displaying and using a video representation of said object displayed on said video surface when at least a portion of said object is determined to be at a location in the world that is not represented as computer graphics but is instead within the field represented by said video on the video surface; and means for displaying said object as a computer graphics object in said first portion of said 25 world in addition to and independent from said video representation of said object on said video surface when said object is determined to be at a location in the world that is represented as computer graphics.
The above-identified apparatus may contain both of said means for displaying are comprised within a single system and or both of said means for displaying may be 30 implemented using a single processor running different software for each means.
In accordance with still yet another aspect of the present invention there is provided an apparatus for use in processing a view of a three dimensional world in which a first -lb-portion of said world is modeled as computer graphics and a second portion of said world is represented by a video on a video surface which exists in said first portion of said world, wherein said video is made up of a sequence of images, the apparatus comprising: means for merging the representation of an object that had been modeled as computer graphics with that of said video surface, so that said object is represented as part of said video displayed on said video surface as a function of a distance between said object and said video surface; and means for supplying said merged representation to a display for displaying said merged representation to a user.
Brief DescriQtion of the Drawings 1o FIGs. 1-4 show an example of that which a user sees according to the invention when a computer graphic object undergoes a trajectory that takes it to a location in the world that is not represented as computer graphics, but instead is within the field represented by the video, such an object should be represented as video on the screen, rather than computer graphics;
FIG. S shows an exemplary process for displaying the world to a user, in accordance with the principles of the invention;
FIG. 6 shows an exemplary process, in accordance with the principles of the invention, for composing a frame of video for display on a video screen;
FIG. 7 is a representation of a three dimensional world and an associated three dimensional display system according to an embodiment of the present invention; and 2o FIG. 8 is a representation of a three dimensional world and an associated view processing system according to an embodiment of the present invention.
Detailed Description To better understand the invention, FIGS. 1-4 show an example of that which a user sees according to the invention when a computer graphic object undergoes a trajectory that takes it to a location in the world that is not represented as computer graphics, but instead is within the field represented by the video, such an object should be represented as video on the screen, rather than computer graphics. In other words, FIGS. 1-4 show an example of a computer graphics object "going into the video" as video and remaining visible to a viewer in front of the video screen, rather than becoming invisible because it is blocked from view by -lc-the video screen if it were to be generated at its proper location using computer graphic techniques.
FIG. 1 shows world 101, which is a bicycle path in a park, e.g., Central Park in New York City. World 101 is divided into two portions, video screen 103, on which is shown the current frame of a video and the remainder of the world 105, which is represented using computer graphic techniques, and is thus referred to herein as computer graphics part (CG Part) 105. Within CG Part 105 there are various elements, such as bicyclist 107 representing the user, another bicyclist 109, and road 111.
Note that the viewpoint of the user is actually behind the representation of both to bicyclists 107 and 109. Also note that bicyclist 107, is moving slower than bicyclist 109. If this continues, eventually, bicyclist 109 will reach screen 103 while bicyclist 107 lags behind.
This occurs in FIG. 2, which shows world 101 of FIG. 1, but at a later time and from a side perspective. If the relative motion difference continues between bicyclists 107 and 109, part, and then all, of bicyclist 109 will intersect with, or go past, video screen 103. If this happens, part, and then all, of bicyclist 109 will become occluded by video screen 103 from a user in front of video screen 103, and so it will no longer be visible to such a user.
Therefore, in accordance with the principles of the invention, bicyclist 109 is represented as video 309 that is merged into the video being displayed on video screen 103 rather than being represented using computer graphic techniques as 109 within CG Part 105, as shown in FIG
3. A
2o corresponding side view is shown in FIG. 4.
The determination to represent an object, such as bicyclist 109, as video, such as video 309, that is merged into the video being displayed on the video screen rather than being represented using computer graphic techniques within CG Part 105, may be made when the object is within a predetermined distance from the video screen rather than requiring the change in representation to be at the exact point when the object intersects the video screen.. Furthermore, the distance employed may be determined dynamically, e.g:, as a function of the speed and trajectory of the bicyclist. The merging in may also take place only once it is determined that a predetermined portion of the object has actually intersected, or passed through, the video screen.
1o In other implementations of the invention, computer graphics representation may remain visible from other viewpoints, such as overhead. Thus, in accordance with an aspect of the invention, the determination as to whether or not to merge an object that is being represented using computer graphic techniques into the video being displayed on the video screen rather than continuing to be represented by using computer graphic techniques may be additionally a function of the viewpoint of the user.
Preferably, but not necessarily, no viewpoint should have both video 309 and computer graphics 109 visible at the same time.
In the event that bicyclist 107 speeds up or bicyclist X09 slows down, so that based on the relative motion between them bicyclist 109 would no longer be at or beyond video 2o screen 103, bicycle 109 may again be rendered using computer graphics techniques, as shown in FIG. 1.
FIG. 5 shows an exemplary process for displaying the world to a user, in accordance with the principles of the invention. The process is entered in step 501 whenever the world view of the user needs to be updated. It may be necessary to update the world view of the user a) whenever a new frame of video must be displayed;
b) when an object within computer graphics part of the world, e.g., CG Part 105, has moved; c) when the user's viewpoint is changed; or d) for any combination of the foregoing reasons.
Thereafter, steps 503 and 505 are performed, and these steps may be performed in parallel. In step 503, each object that is still within the view of the user and has not been 3o merged into the video is rendered within the computer graphics part of the world. In step 505 a frame of video is displayed on video screen. This may be either a new frame of video, if it is time to update the video frame being displayed, or it may be the same frame of video that was previously displayed, if the world view of the user is being updated because of a need to change the representation of the computer graphics part of the world.
Note that the frame of video that is displayed may be a "composed" frame of video in that includes a background video and video representing at least one object that was formerly represented the computer graphics portion of the world using computer graphics techniques. The process then exits in step 507.
FIG. 6 shows an exemplary process, in accordance with the principles of the invention, for composing a frame of video that is displayed in step 505 (FIG.
5). The process is entered in step 601 whenever it is determined that a new frame of video is required for display. Thus, execution of the process of FIG. 6 may be linked to step 505 or it may be independent thereof. Next, in step 603, the viewpoint of the user is determined. Thereafter, in step 605, the position of the video display surface, e.g., screen, is determined, and then, in step 607, the set of objects within the three dimensional world to for which merging is permitted is determined. Such objects may be identified by a set flag, such as a VIDEO PERMITTED flag, within their data structure.
Conditional branch point 609 tests to determine if there remain any objects in the set of objects for which merging is permitted that have not yet been checked to determine if they should be displayed as video or using computer graphic techniques. If the test result in step 609 is NO, indicating that all the objects for which merging is permitted have been checked to determine if they should be displayed as video or using computer graphic techniques, control passes to step 623 and the process is exited. If the test result in step 609 is YES, indicating that all the objects for which merging is permitted have not been checked to determine if they should be displayed as video or using computer graphic 2o techniques, control passes to step 611, in which the next object from the set that has not yet been checked is retrieved, and the retrieved object is made the current object.
Next, in step 613, a flag, e.g., a bit called VIDEO REPRESENTATION, for the current object is cleared. In step 615, the position of the current object is determined.
Optionally, the motion of the current object, e.g., its speed and trajectory, are also determined.
Thereafter, in accordance with the principles of the invention, conditional branch point 617 tests to determine if the current object should be represented by video. As noted above, if the object is at a point in its trajectory that takes it to a location in the world that is not represented as computer graphics, but instead is within the field 3o represented by the video, the object should be represented as video, rather than using computer graphics. More specifically, such a determination may be based, in accordance with an aspect of the invention, on the distance of the object from the video screen. This distance may be determined dynamically, e.g., as a function of the speed and trajectory of the bicyclist. The determination to represent an object as video may be made only if a predetermined portion of the object has actually intersected, or passed through, the video screen.
If the test result in step 617 is NO, indicating that the object may be displayed using computer graphics, control passes back to step 609 and the process continues as described above. If the test result in step 617 is YES, indicating that the computer graphics object must go into the video as video, control passes to step 619 in which the VIDEO REPRESENTATION bit for the current object is set. This indicates that the object is being represented as video and so should not be displayed in the computer graphics part of the world. Thereafter, in step 621, a video representation of the current object is retrieved and merged with the already composed video frame. This may be achieved by using a replacement technique to replace that part of the already composed video frame by retrieved 1 o video representation. Alternatively, a merging technique may be employed.
Control then passes back to step 609 and the process continues as described above.
The development of the merged video of step 621 may be achieved using an alternative method. Instead of developing a single frame of video for display, multiple, stacked video screens may be employed, with one screen displaying the frame that would be t 5 displayed if no objects were merged into the video and each of the other screens displaying the video representation of each object that is currently merged into the video. The viewer in front of all the screens would see a composite image containing all the video elements and appearing as a single screen. In such a case, step 505 (FIG. 5) would encompass displaying the videos of all of the screens.
20 Note that the video screen may be any surface and need not be simply a planar surface.
FIG. 7 shows a three dimensional world display system 50 for displaying an apparently three dimensional world 30 to a user on a display screen 33. The three dimensional world 30 represented on the display screen 33 comprises a first portion 31 25 displayed as computer graphics and a second portion 32 displayed as video.
The video is made up of a sequence of images.
With reference to FIG. 7, a processor 52 determines that an element (e.g. see bicyclist 109 in FIG. 1 - FIG. 2) displayed in said first portion 31 of the world 30 transited from the first portion 31 to the second portion 32. A rendering engine 51 displays the world 30 with the element displayed so that it appears to the viewer to be substantially integrated into the second portion 32 rather than the first portion 31 of the world 30.
The arrangement of the first and second portions 31, 32 is for pedagogic illustrative purposes only and is not representative. The first and second portions 31, 32 can also be arranged vertically or within each other (i.e. second portion 32 within first portion 31), for example.
to FIG. 8 shows a view processing system 40 for processing a view of a three dimensional world 30 to be displayed by a display screen 33 according to an embodiment of the present invention. The three dimensional world 30 represented on the display screen 33 comprises a first portion 31 modeled as computer graphics and a second portion represented by a video. The video is made up of a sequence of images.
With reference to FIG. 8, a video object locating mechanism 42 determines that an object (e.g. see police van 113 in FIG. 1 - FIG. 3) which is being modeled as computer graphics has moved to a location (second portion 32) that is being represented by the video by having moved to a location that is being represented by the second portion 32 and at which the object is at least partly occluded from view of a viewer of the world by the video 2o although the object should not appear occluded to the viewer when at the location. The video object locating mechanism 42 relays information about the determined object to a graphics-to-video / video-to-graphics conversion mechanism 41. The graphics-to-video /
video-to-graphics conversion mechanism 41 changes the representation of the object from one modeled as computer graphics to a video representation thereof. It thereby appears to the viewer that the object is substantially integrated into the second portion 32 rather than the first portion 31.
The foregoing merely illustrates the principles of the inventions. It will thus be appreciated that those skilled in the art wilt be able to devise various arrangements which, although not explicitly described or shown herein, embody the principles of the invention and 3o are included within its spirit and scope.

Claims (20)

Claims
1. A method for use in processing a view of a three dimensional world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video display on a video screen which exists in said first portion of said world, wherein said video is made up of a sequence of images, the method comprising the steps of:
determining that an object, other than said video screen, is being modeled as computer graphics and has moved to a location that is being represented by said video and at which said object is at least partly occluded from view of a viewer of said world by said video screen although said object should not appear occluded to said viewer when at said location;
changing the representation of said object from one modeled as computer graphics to a. video representation thereof; and displaying said video representation so that it appears to said viewer to be substantially integrated with said video on said video screen.
2. A method for use in processing a view of a three dimensional world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video displayed on a video surface which exists as a first object in said computer graphics portion of said world, comprising the step of merging the representation of a second object that had been modeled as computer graphics with that of said video surface, so that said second object is represented as part of said video displayed on said video surface as a function of a distance between said object and said video surface independent of a distance between a viewer and said object.
3. The method as defined in claim 2 wherein said second object is represented as part of said video displayed on said video surface as a further function of a speed of said object.
4. The method as defined in claim 2 wherein said second object is represented as part of said video displayed on said video surface as a further function of a trajectory of said object.
5. The method as defined in claim 2 wherein said second object is represented as part of said video displayed on said video surface as a further function of a viewpoint of a viewer of said three dimensional world.
6. A method for use in processing a view of a world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video on a video surface which exists in said computer graphics portion of said world, said video initially depicting at least a first object of said world, wherein said video is made up of a sequence of images, the method comprising the steps of:
changing the representation of a second object that is initially represented as computer graphics from a computer graphics representation to a video representation when said object undergoes a trajectory that takes at least a portion of said object to a location in the world that is not represented as computer graphics but instead is being represented by said video on said video surface and at which said second object is at least partly occluded from view of a viewer of said world by said video surface although said second object should not appear occluded to said viewer when at said location; and displaying said world so that it appears to said viewer that said video representation of said second object is substantially integrated with said video on said video surface.
7. The method as defined in claim 6 wherein said changing step is a function of a viewpoint of a viewer.
8. The method as defined in claim 6 wherein said changing step is a function of a speed of said object.
9. A method for use in processing a view of a world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video on a video surface which exists in said first portion of said world, wherein said video is made up of a sequence of images, the method comprising the steps of:
displaying an object using only a video representation of said object displayed on said video surface when at least a portion of said object is determined to be at a location in the world that is not represented as computer graphics but instead is within the field represented by said video on the video surface; and displaying said object as a computer graphics object in said first portion of said world in addition to and independent from the video representation of the object on said video surface when said second object is determined to be at a location in the world that is represented as computer graphics.
10. The method as defined in claim 9 wherein said video representation of said object is displayed on said video surface.
11. The method as defined in claim 9 further including the step of changing a representation of said object from computer graphics to video when said portion of said object transits from a location in the world that is represented as computer graphics to a location within the field represented on the video surface.
12. The method as defined in claim 9 further including the step of changing a representation of said object from video to computer graphics when said portion of said object transits from a location in the world represented on the video surface to a location within the field that is represented as computer graphics.
13. A system for displaying an apparently three dimensional world to a user, in which a first part of said three dimensional world is displayed as computer graphics and a second part is displayed as video on a video screen, wherein said video is made up of a sequence of images, said system comprising:
a processor for determining that an element displayed in said first part of said world transited from said first part to said second part by having moved to a location that is being represented by said second part and at which said element is at least partly occluded from view of a viewer of said world by said video screen although said element should not appear occluded to said viewer when at said location; and a rendering engine for displaying said world with said element displayed so that it appears to said viewer to be substantially integrated into said second part rather than said first part.
14. A system for displaying an apparently three dimensional world to a user, in which a first part of said three dimensional world is displayed as computer graphics and a second part is displayed as video on a video screen, wherein said video is made up of a sequence of images, said system comprising:
a processor for determining that a portion of an object initially in said first part of said world has moved to a location within said second part by virtue of having moved to a location that is being represented by said second part and at which portion of said object is at least partly occluded from view of a viewer of said world by said video screen although said object should not appear occluded to said viewer when at said location; and a rendering engine for displaying said world with said object displayed so that it appears to said viewer to be substantially integrated into said second part rather than said first part.
15. Apparatus for use in processing a view of a three dimensional world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video, wherein said video is made up of a sequence of images, the apparatus comprising:
means for determining that an object being modeled as computer graphics has moved to a location that is being represented by said video by having moved to a location that is being represented by said second portion and at which said object is at least partly occluded from view of a viewer of said world by said video although said object should not appear occluded to said viewer when at said location; and means for changing the representation of said object from one modeled as computer graphics to a video representation thereof displayed so that it appears to said viewer to be substantially integrated into said second portion rather than said first portion.
16. The apparatus as defined in claim 15 further comprising means for displaying said video representation of said object merged with said portion of said world that is represented by said video.
17. Apparatus for use in processing a view of a world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video on a video surface which exists in said first portion of said world, wherein said video is made up of a sequence of images, said apparatus comprising:
means for displaying an object using a video representation of said object displayed on said video surface when at least a portion of said object is determined to be at a location in the world that is not represented as computer graphics but is instead within the field represented by said video on the video surface; and means for displaying said object as a computer graphics object in said first portion of said world in addition to and independent from said video representation of said object on said video surface when said object is determined to be at a location in the world that is represented as computer graphics.
18. The apparatus as defined in claim 17 wherein both of said means for displaying are comprised within a single system.
19. The apparatus as defined in claim 17 wherein both of said means for displaying are implemented using a single processor running different software for each means.
20. Apparatus for use in processing a view of a three dimensional world in which a first portion of said world is modeled as computer graphics and a second portion of said world is represented by a video on a video surface which exists in said first portion of said world, wherein said video is made up of a sequence of images, the apparatus comprising:
means for merging the representation of an object that had been modeled as computer graphics with that of said video surface, so that said object is represented as part of said video displayed on said video surface as a function of a distance between said object and said video surface; and means for supplying said merged representation to a display for displaying said merged representation to a user.
CA002242166A 1997-08-04 1998-06-30 Display techniques for three dimensional virtual reality Expired - Fee Related CA2242166C (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/906,215 1997-08-04
US08/906,215 US6323861B1 (en) 1997-08-04 1997-08-04 Display techniques for three dimensional virtual reality

Publications (2)

Publication Number Publication Date
CA2242166A1 CA2242166A1 (en) 1999-02-04
CA2242166C true CA2242166C (en) 2005-08-16

Family

ID=25422101

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002242166A Expired - Fee Related CA2242166C (en) 1997-08-04 1998-06-30 Display techniques for three dimensional virtual reality

Country Status (4)

Country Link
US (1) US6323861B1 (en)
AU (1) AU7869598A (en)
CA (1) CA2242166C (en)
GB (1) GB2328101B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20010015674A (en) * 1998-07-30 2001-02-26 마츠시타 덴끼 산교 가부시키가이샤 Moving picture synthesizer
US9219945B1 (en) * 2011-06-16 2015-12-22 Amazon Technologies, Inc. Embedding content of personal media in a portion of a frame of streaming media indicated by a frame identifier
CN107797836B (en) * 2017-08-25 2020-03-27 深圳壹账通智能科技有限公司 Virtual reality page generation method and device, server and storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5261041A (en) * 1990-12-28 1993-11-09 Apple Computer, Inc. Computer controlled animation system based on definitional animated objects and methods of manipulating same

Also Published As

Publication number Publication date
GB2328101A (en) 1999-02-10
GB2328101B (en) 1999-06-30
CA2242166A1 (en) 1999-02-04
US6323861B1 (en) 2001-11-27
AU7869598A (en) 1999-02-11
GB9816367D0 (en) 1998-09-23

Similar Documents

Publication Publication Date Title
US6690393B2 (en) 3D environment labelling
EP0930584B1 (en) Method and apparatus for displaying panoramas with video data
Ware Designing with a 2½D attitude
AU6742700A (en) Display method for multiple layered screens
WO1999052079A1 (en) Object-based anti-aliasing
JP2003507828A (en) Controlling the depth movement of a visible display with a laminated screen
US7515156B2 (en) Method and apparatus for parallel speculative rendering of synthetic images
EP0899691B1 (en) Three dimensional virtual reality enhancement techniques
US6226009B1 (en) Display techniques for three dimensional virtual reality
CA2242166C (en) Display techniques for three dimensional virtual reality
Wloka et al. Interactive real-time motion blur
CA2270734C (en) Display techniques for three-dimensional virtual reality
Hirose CABIN-A multiscreen display for computer experiments
EP1111546A2 (en) Display techniques for three-dimensional virtual reality
CA2242150C (en) Display techniques for three dimensional virtual reality
US6466230B1 (en) Display techniques for object visibility in a three-dimensional virtual world
CN114935977B (en) Spatial anchor point processing method and device, electronic equipment and storage medium
JPH04220782A (en) Three-dimensional image display processing method
JP3739829B2 (en) Graphics equipment
JP2003108112A (en) Method and device for object display
Yildiz et al. User directed view synthesis on OMAP processors
Andiel et al. Position-and velocity-depending subpixel correction for spatially multiplexed autostereoscopic displays

Legal Events

Date Code Title Description
EEER Examination request
MKLA Lapsed
MKLA Lapsed

Effective date: 20090630