US20050090302A1 - System and method for utilizing vectors in a video game - Google Patents
System and method for utilizing vectors in a video game Download PDFInfo
- Publication number
- US20050090302A1 US20050090302A1 US10/737,143 US73714303A US2005090302A1 US 20050090302 A1 US20050090302 A1 US 20050090302A1 US 73714303 A US73714303 A US 73714303A US 2005090302 A1 US2005090302 A1 US 2005090302A1
- Authority
- US
- United States
- Prior art keywords
- vectors
- path
- vector
- color
- previous
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/50—Controlling the output signals based on the game progress
- A63F13/53—Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game
- A63F13/537—Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game using indicators, e.g. showing the condition of a game character on screen
- A63F13/5372—Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game using indicators, e.g. showing the condition of a game character on screen for tagging characters, objects or locations in the game scene, e.g. displaying a circle under the character controlled by the player
-
- A63F13/10—
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/45—Controlling the progress of the video game
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/80—Special adaptations for executing a specific game genre or game mode
- A63F13/803—Driving vehicles or craft, e.g. cars, airplanes, ships, robots or tanks
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/30—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device
- A63F2300/303—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device for displaying additional data, e.g. simulating a Head Up Display
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/30—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device
- A63F2300/303—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device for displaying additional data, e.g. simulating a Head Up Display
- A63F2300/306—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device for displaying additional data, e.g. simulating a Head Up Display for displaying a marker associated to an object or location in the game field
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/64—Methods for processing data by generating or executing the game program for computing dynamical parameters of game objects, e.g. motion determination or computation of frictional forces for a virtual car
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/80—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
- A63F2300/8017—Driving on land or water; Flying
Definitions
- This invention relates generally to computer-generated images and more particularly to a system and method for utilizing vectors in a video game.
- video game developers develop games that capitalize upon the competitive nature of many game players by recording aspects of a previous run, and then allowing subsequent game players access to previous run data. That is, subsequent players may compete against performances of previous players.
- the previous run may be, for example, based upon a fastest time, most points accumulated, a most efficient use of player character energy reserves, or any combinations of these or other attributes.
- a player character associated with a previous run may be simultaneously displayed with a current player character on a graphical display.
- a current player may observe a position of the player character associated with the previous run, and directly compete with this previous character.
- the current player may observe a single position of the previous character at any given time, the current player does not observe data associated with a path of the previous character.
- a current player character with access to path data associated with a previous run may be able to use the path data during game play.
- a system and method for utilizing vectors in a video game.
- the method comprises initiating a current video game session, computing vectors along one or more graphical paths of the video game, where each of the graphical paths are associated with a player character of the video game, and displaying the plurality of vectors along the graphical paths.
- the method comprises computing net resultant force vectors, where the vectors are computed in real time and are based upon phenomenological and/or physical laws of nature.
- the method comprises retrieving graphical path data associated with one or more previous runs, displaying the graphical path data as a string of vectors, and determining a color for each vector based upon an elapsed time of the current video game session and an elapsed time associated with the vector.
- the color of each vector may be based upon a player character state.
- the player character state represents a state of the player character, and includes states such as an “on the ground” state, an “airborne” state, and a “crashed” state.”
- a system of the present invention comprises a data cache configured to store graphical path data associated with a current video game session and one or more previous video game sessions, a processor configured to compute a plurality of force vectors associated with one or more graphical paths, and a display device configured to display the plurality of force vectors.
- the processor is configured to compute a color of a force vector from the plurality of force vectors, where the color of each force vector indicates a player character state.
- the graphical path data may be associated with previous game runs, including, but not limited to, previous “best time” runs, previous “worst time” runs, previous “average time” runs, and selectively chosen runs associated with a current game player or other game players.
- FIG. 1 is an exemplary block diagram of one embodiment of an electronic entertainment system, according to the present invention.
- FIG. 2 is an exemplary block diagram of one embodiment of the data cache of FIG. 1 ;
- FIG. 3A illustrates a bicycle on a roadway surface, according to an exemplary example of the present invention
- FIG. 3B illustrates forces acting in the y-z plane on a wheel of the bicycle illustrated in FIG. 3A ;
- FIG. 3C illustrates forces acting in the x-z plane on the wheel of the bicycle illustrated in FIG. 3A ;
- FIG. 3D shows a summation of the forces acting on the wheel illustrated in FIG. 3B ;
- FIG. 3E shows a summation of the forces acting on the wheel illustrated in FIG. 3C ;
- FIG. 3F shows a net resultant force acting on the wheel of the bicycle as a summation of the forces illustrated in FIG. 3D and FIG. 3E ;
- FIG. 4 is an exemplary screen-shot of net resultant force vectors as displayed at a given time during a game run from a vantage point of a player character A, according to one embodiment of the invention
- FIG. 5A is an illustration of path markers as displayed at a beginning of a time-trial run, according to one embodiment of the present invention.
- FIG. 5B is an illustration of path markers as displayed after an elapsed time t 1 from a start of the time-trial run as displayed in FIG. 5A , according to one embodiment of the present invention
- FIG. 5C is an illustration of path markers as displayed after an elapsed time t 2 from a start of the time-trial run as displayed in FIG. 5A , according to another embodiment of the present invention.
- FIG. 5D is an illustration of path markers as displayed after an elapsed time t 3 from a start of the time-trial run as displayed in FIG. 5A , according to yet another embodiment of the present invention.
- FIG. 5E is an exemplary screen-shot of path markers as displayed after an elapsed time t 4 from a beginning of a time-trial run.
- FIG. 6 is a flowchart of method steps for recording and displaying vectors along a graphical path, according to one embodiment of the present invention.
- a game player maneuvers a player character through a game environment from a starting point to an end point. For example, the player character traverses a racecourse in a time trial mode.
- Video game software records a path of the player character as it moves from the starting point to the end point. The path is recorded as a series of points represented by (x,y,z) Cartesian coordinates, for example.
- the game software also records a total time that it takes the player character to arrive at the end point and an elapsed time to each point in the path. For example, the elapsed time may be measured from an initial starting time associated with the starting point.
- the game software may also record other data at each point along the player character's path including, but not limited to, a magnitude and a direction of each force acting on the player character, an instantaneous energy consumption by the player character, a total energy consumption by the player character, and a player character travel state.
- the player character travel state may indicate that the character is “airborne,” “on the ground,” or “crashed.” Other player character states may be utilized in further embodiments.
- FIG. 1 is an exemplary block diagram of an electronic entertainment system 100 , according to one embodiment of the present invention.
- the system 100 comprises a main memory 110 , a central processing unit (CPU) 112 , a data cache 111 , a vector processing unit VPU 113 , a graphics processing unit (GPU) 114 , an input/output processor (IOP) 116 , an IOP memory 118 , a controller interface 120 , a memory card 122 , a Universal Serial Bus (USB) interface 124 , and an IEEE 1394 interface 126 .
- CPU central processing unit
- VPU vector processing unit
- GPU graphics processing unit
- IOP input/output processor
- IOP input/output processor
- the system 100 also comprises an operating system read-only memory (OS ROM) 128 , a sound processing unit (SPU) 132 , an optical disc control unit 134 and a hard disc drive (HDD) 136 , which are connected via a bus 146 to the IOP 116 .
- OS ROM operating system read-only memory
- SPU sound processing unit
- HDD hard disc drive
- the system 100 is preferably an electronic gaming console; however, the system 100 may also be implemented as any type of general-purpose computer, set-top box or hand-held gaming device. Alternative embodiments of the system 100 may comprise fewer, more, or other components.
- a user of the system 100 provides instructions via the controller interface 120 to the CPU 112 .
- the user may instruct the CPU 112 to store certain game information on the memory card 122 or may instruct a character in a game to perform some specified action.
- Other devices may be connected to the system 100 via the USB interface 124 and the IEEE 1394 interface 126 .
- the CPU 112 , the VPU 113 , the GPU 114 and the IOP 116 communicate via a system bus 144 .
- the CPU 112 communicates with the main memory 110 via a dedicated bus 142 , and with the data cache 111 via a dedicated bus 148 .
- the VPU 113 and the GPU 114 may also communicate via a dedicated bus 140 .
- the CPU 112 executes programs stored in the OS ROM 128 and the main memory 110 .
- the main memory 110 may contain pre-stored programs and may also contain programs transferred via the IOP 116 from a CD-ROM, a DVD-ROM or other optical disc (not shown) using the optical disc control unit 134 .
- the IOP 116 controls data exchanges between the CPU 112 , the VPU 113 , the GPU 114 and other devices of the system 100 , such as the controller interface 120 .
- the SPU 132 executes instructions to produce sound signals that are output on an audio device (not shown). Alternative embodiments may provide different communication paths between the various components.
- the CPU 112 stores and retrieves data associated with previous paths or runs (such as a “best time” run) in the data cache 111 .
- the data cache 111 is discussed further below in conjunction with FIG. 2 .
- the CPU 112 may retrieve data associated with a “best time” path generated by a developer of the game software and permanently stored on a game disc.
- the CPU 112 may instruct the optical disc control unit 134 to read the game disc.
- players may test themselves against the “best time” and learn an efficient path through the game environment.
- a player's “best time” path or any other paths may be stored on the memory card 122 , such as a removable flash card. The player then has an option, upon initiation of a new game session, of either racing against a “best time” path as stored on the memory card 122 or any other recorded path.
- the game software retrieves recorded data associated with a selected previous run, and then displays the data on a game screen as the character moves through the game environment, allowing the player character to race against the selected previous run.
- the GPU 114 executes drawing instructions from the CPU 112 and the VPU 113 to produce images for display on a display device (not shown).
- the VPU 113 transforms objects from three-dimensional coordinates to two-dimensional coordinates, and sends the two-dimensional coordinates to the GPU 114 .
- the game software displays the data associated with the previous run, such as previous run data, as a string of dynamically changing colored path markers situated in the game environment.
- the string of colored path markers displays a game developer's “best time” path through the game environment.
- the string of colored path markers displays a previous run path through the game environment by one game player from a group of game players.
- a path marker's color conveys information. For example, when the player selects a racecourse, the game software retrieves recorded data associated with a “best time” path before the current player character begins a time trial run, and displays a string of path markers. Each path marker of the string of path markers is initially a color co. In one embodiment of the present invention, the color co is gray. However, any color may represent the color co. In this embodiment, a path marker colored gray indicates that a player character associated with the “best time” run (hereinafter referred to as the “best time” player character) has not yet reached the path marker's position at a given elapsed time. Thus at any given elapsed time during game play, if the current player character is ahead of the “best time” player character, then the current player observes a string of path markers colored gray that represents the “best time” path.
- a path marker may change from the color c 0 to a color c 1 when the “best time” player character is positioned at the path marker and is traveling on the ground.
- a path marker may change from the color c 0 to a color c 2 when the “best time” player character is positioned at the path marker and is traveling in the air.
- a path marker may change from the color c 0 to a color c 3 when the “best time” player character is positioned at the path marker and is crashed.
- the color c 1 is green
- the color c 2 is blue
- the color c 3 is red.
- the scope of the present invention covers other colors for the above-described path markers.
- other colored path markers may represent alternate player character states, such as an energy level, an applied force magnitude, and an applied force direction, for example.
- FIG. 2 is an exemplary block diagram of one embodiment of the data cache 111 of FIG. 1 .
- the exemplary data cache 111 comprises a previous sub-cache 210 and a current sub-cache 220 .
- the previous sub-cache 210 stores data associated with n path markers of a previous run, where n is an integer.
- the previous sub-cache 210 is a “best time” sub cache which stores data associated with n path markers of a previous “best time” run.
- the current sub-cache 220 stores data generated during a current time-trial run, where the data is associated with m path markers, and m is an integer.
- the previous sub-cache 210 may store any of the following data: an x-position of path marker 1, a y-position of path marker 1, a z-position of path marker 1, an elapsed time t at path marker 1, a player character state at path marker 1, a player character speed and/or velocity at path marker 1, an x-component of a net resultant force (F x ) acting on the player character at path marker 1, a y-component of a net resultant force (F y ) acting on the player character at path marker 1, a z-component of a net resultant force (F z ) acting on the player character at path marker 1, and an instantaneous work performed by the net resultant forces (Fx, Fy, Fz) on the player character at path marker 1.
- the player character state at path marker 1 identifies a state of the player character.
- player character states include, but are not limited to, on the ground, in the air, and crashed.
- Alternate embodiments of the previous sub-cache 210 may comprise more, less, and/or other data associated with any path marker.
- the current sub-cache 220 comprises similar data categories as the previous sub-cache 210 , and thus the current sub-cache 220 will not be described in detail.
- the data cache 111 comprises a plurality of sub-caches (not shown) in which each sub-cache of the plurality of sub-caches may store data associated with previous game runs, which may include a “best time” run.
- the CPU 112 executes the video game software.
- the previous sub-cache 210 is a “best time” sub-cache 210 .
- the CPU 112 generates and records path markers based upon a distance a player character travels. For example, upon initiation of a time-trial run, when a “best time” path does not yet exist, the CPU 112 executes video game instructions that determine when a player character has traveled a predefined distance s from a starting point. The CPU 112 then generates a path marker 1 associated with the distance s, and may store the following data associated with the path marker 1 in the “best time” sub-cache 210 ( FIG.
- the player character state may indicate whether the player character is on the ground, in the air, or crashed at the path marker 1.
- Alternative embodiments may comprise additional or different states.
- each new path marker is located at a predefined player character travel distance s from a previous path marker position.
- the CPU 112 if the player character crashes before the player character travels the predefined distance s measured from the previous path marker, then the CPU 112 generates a path marker at or near the crash site, and stores the data associated with the crash site path marker in the “best time” sub-cache 210 .
- the CPU 112 generates and stores data associated with new path markers in the “best time” sub-cache 210 , where new path markers are separated by a predefined time interval.
- the CPU 112 retrieves the data from the “best time” sub-cache 210 , and displays a string of path markers on the display device (not shown).
- the string of path markers indicates the player character's path recorded during the “best time” run, for example.
- the “best time” path markers initially have a color co.
- the CPU 112 generates and records current path markers and data associated with the current path markers to the current sub-cache 220 .
- the CPU 112 compares a current elapsed time with an elapsed time t retrieved from the “best time” sub-cache 210 . For each “best time” path marker with an associated elapsed time t that is less than the current elapsed time, the CPU 112 changes the “best time” path marker color based upon the “best time” player character state associated with the “best time” path marker. For example, if the “best time” player character state indicates that the “best time” player character is on the ground at a given path marker, then the CPU 112 instructs the GPU 114 to change the color of the given path marker from c 0 to c 1 .
- the CPU 112 instructs the GPU 114 to change the color of the given path marker from c 0 to c 2 .
- the CPU 112 instructs the GPU 114 to change the color of the given path marker from c 0 to c 3 .
- the CPU 112 compares an elapsed time t associated with the marker m from the current sub-cache 220 (also referred to as a total elapsed time) with an elapsed time t associated with the marker n from the “best time” sub-cache 210 (also referred to as a “best time” total elapsed time). If the total elapsed time from the repeat time-trial run is less than the “best time” total elapsed time, then the CPU 112 overwrites data of the “best time” sub-cache 210 with data of the current sub-cache 220 .
- data from the repeat time-trial run is now stored in the “best time” sub-cache 210 .
- data generated from subsequent time-trial runs and stored in the current sub-cache 220 are compared to data stored in the “best time” sub-cache 210 .
- data associated with a “best time” run may already be stored or predetermined in the “best time” sub-cache 210 prior to initiation of the time-trial run.
- the CPU 112 retrieves the data from the “best time” sub cache 210 and displays a string of path markers on the display device (not shown) that indicate a path of a “best time” player character.
- the CPU 112 computes a net resultant force acting on the player character at each of the path markers generated during the time-trial run.
- the net resultant force comprises Cartesian components (Fx, Fy, Fz).
- the scope of the invention covers net resultant force components defined with respect to any coordinate system. These forces may be based upon the governing laws of physics and/or any artificially contrived laws.
- the CPU 112 may compute the net resultant forces based upon phenomenological laws (i.e., contrived physical laws) that dictate motion of a player character as the player character collides with extraneous objects, for example.
- Extraneous objects comprise any objects within a game environment, but external to the player character.
- extraneous objects may include, but are not limited to, a roadway surface, rocks, ledges, holes, bumps, walls, other game characters, and other player characters.
- the game software comprises phenomenological laws which the CPU 112 processes to direct the motion of the player character when the player character interacts with an extraneous object. For example, if the player character hits a rock, the CPU 112 retrieves from main memory 110 ( FIG.
- a phenomenological law associated with hitting a rock uses the phenomenological law and other variables and/or constants retrieved from the main memory 110 or the data cache 111 (e.g., speed and mass of the player character) to compute a net resultant force acting on the player character from the impact with the rock.
- the CPU 112 may compute net resultant forces based upon real laws of nature, such as the law of gravity, or laws based upon water and wind resistance.
- resistive frictional forces such as water and wind resistance forces, may depend upon a shape and a size of a player character or of apparel worn by the player character.
- the CPU 112 may compute net resultant forces based upon a combination of phenomenological laws and real laws of nature. For example, if a bicycle hits a wall, the CPU 112 may compute a gravitational force, a frictional wind force, and a phenomenological impact force due to a collision with the wall. The CPU 112 then computes a net resultant force acting on the bicycle as a vector sum of the gravitational force, wind force, and impact force. The CPU 112 may then instruct the GPU 114 ( FIG. 1 ) to display the net resultant force on the display device (not shown) and/or may store the net resultant force in the data cache 111 ( FIG. 1 ) for retrieval and display during subsequent game play.
- FIG. 3A illustrates a bicycle 302 on a roadway surface 304
- FIG. 3B illustrates four forces acting on a wheel 305 of the bicycle 302 in the y-z plane.
- some of the forces are illustrated to act on a point of contact between the roadway surface 304 and the wheel 305
- other forces are shown to act on a center of mass of the wheel 305 (i.e., geometric center of the wheel in this embodiment).
- the forces comprise a force of gravity F g , a frictional force F u , a pedaling force F p (i.e., the force transferred from the bicycle pedals (not shown) via a bicycle chain (not shown) to the wheels 305 and applied to the roadway surface 304 ), and an impulsive force F i .
- the impulsive force F i is typically a force of high magnitude and short duration that may be incurred as the wheel 305 strikes irregularities in the roadway surface 304 , for example.
- FIG. 3C illustrates additional forces that may act on the wheel 305 as viewed along line AB ( FIG. 3A ), and acting on the wheel 305 in the x-z plane.
- These forces may comprise a lateral centripetal force F c applied to the wheel 305 by the roadway surface 304 when the bicycle 302 is turning (e.g., a centripetal force is generated when the driver is leaning into a turn), and a normal force F n applied perpendicular to the roadway surface 304 .
- FIG. 3D shows a summation of all the forces acting on the wheel 305 in the y-z plane as illustrated in FIG. 3B as the y-z resultant force F yz .
- FIG. 3E shows a summation of all the forces acting on the wheel in the x-z plane as illustrated in FIG. 3C as the x-z resultant force F xz .
- FIG. 3F shows a net resultant force F r acting on the wheel 305 of the bicycle 302 as a summation of the y-z resultant force F yz ( FIG. 3D ) and the x-z resultant force F xz ( FIG. 3E ).
- a net resultant force may also be computed as a summation of resultant forces acting on other components of the bicycle 302 and on a rider (not shown) of the bicycle 302 .
- Other forces may be computed as well, including frictional forces due to air and water (e.g., rain) resistance, and based upon bicycle speed and bicycle structural design and/or rider clothing, for example.
- the net resultant forces may be recorded during a time trial, and then graphically displayed during subsequent runs to inform game players of previous actions and assist in game decisions that may improve game play performance. Additionally, the net resultant forces may be computed and graphically displayed in real time during one or more game players' current runs to provide additional information regarding path smoothness and/or locations of impediments to travel, such as obstacle or barriers. Other applications may be considered as well.
- FIG. 4 is a screen-shot of net resultant force vectors as displayed at a given time during a game run from a vantage point of player character A, according to one embodiment of the present invention.
- the CPU 112 FIG. 1
- the CPU 112 computes and displays net resultant force vectors along paths associated with each game player. For example, net resultant force vectors, such as net resultant force vectors 405 , 410 , and 415 are displayed along a path traveled by player character D.
- a net resultant force vector 420 is displayed along a path traveled by player character B
- a net resultant force vector 425 is displayed along a path traveled by player character C
- a net resultant force vector 430 is displayed instantaneously acting on player character A (i.e., current player).
- other net resultant force vectors are computed and displayed along the paths of player characters B, C, and D, they are not readily visible in the FIG. 4A embodiment of the present invention. Therefore, only the most visible net resultant force vectors 405 - 430 are illustrated in this exemplary example utilizing the present invention.
- Each net resultant force vector of the net resultant force vectors 405 - 430 have a magnitude (exemplified by a vector length) and a direction.
- the net resultant force vectors 405 - 430 are colored to represent a player character's travel state. The significance of colored net resultant force vectors will be discussed further below in conjunction with path markers and FIGS. 5A-5E .
- Player characters B, C, and D may be other current player characters competing against player character A, previous player characters representing previous game paths, or a combination of other current player characters and previous player characters.
- a player selects player character A ( FIG. 4 ), for example, to compete in a time-trial run.
- Player character A then competes against a previous run as displayed via a string of net resultant force vectors.
- the previous run may be a previous “best time” run.
- the previous run force vectors not only illustrate the previous path of a player character, but also magnitude and direction of net resultant forces encountered by the player character.
- Player character A may use the information provided by the string of net resultant force vectors to avoid obstacles and select paths in order to improve performance.
- a game developer may utilize the computed and displayed net resultant force vectors to develop and debug new versions of video games, or fine tune game dynamics of existing video games. Other uses of net resultant force vectors are contemplated and covered by the present invention.
- FIG. 5A is an exemplary illustration of path markers as displayed at a beginning of a time-trial run, according to one embodiment of the present invention.
- a player character 510 is located at a starting position A, and path markers 1-13 represent a path of a “best time” run.
- the path markers 1-13 are represented by circles, however the scope of the invention covers markers of any shape or size.
- FIG. 5A is an aerial view of the path markers 1-13 and the player character 510 .
- a time-trial run's game environment is typically viewed from a perspective of the player character 510 .
- the player character 510 may directly observe path markers 1-13.
- path markers visible to the player character 510 are a color c 0 .
- FIG. 5B is an exemplary illustration of the path markers of FIG. 5A as displayed after an elapsed time t 1 from a beginning of a time-trial run.
- a player character 510 travels along a path designated by positions A-D, and at time t 1 , the player character 510 is located at the position D.
- a “best time” player character (not shown) is located at a path marker 3 at time t 1 .
- the CPU 112 instructs the GPU 114 ( FIG. 1 ) to change the color of each path marker 1-13 from c 0 to another color, dependent upon a “best time” player character state associated with each path marker, if the elapsed time t associated with each respective path marker is less than or equal to the elapsed time t 1 .
- path markers 1-3 have elapsed times t that are less than or equal to the elapsed time t 1
- the “best time” player character states associated with path markers 1-3 indicate that the “best time” player character is traveling on the ground.
- the GPU 114 changes color of the path markers 1-3 from a color c 0 to a color c 1 . Consequently, the path markers 4-13 with elapsed times t-greater than the elapsed time t 1 remain unchanged.
- the path markers 1-3 disappear from the display device (not shown), if the player character 510 is past the path markers 1-3.
- FIG. 5C is an exemplary illustration of the path markers of FIG. 5A as displayed after an elapsed time t 2 from a beginning of a time-trial run.
- the player character 510 travels along a path designated by positions A-D, and at time t 2 , the player character 510 is located at a position D.
- a “best time” player character (not shown) associated with a “best time” path (represented by circular path markers 1-13) is located either between path markers 8 and 9, or at the path marker 8 at the elapsed time t 2 . Since the path markers 9-13 have elapsed times that are greater than the elapsed time t 2 , the path markers 9-13 remain unchanged.
- the path markers 1-8 will change color. For example, since the path markers 1-5 and 8 have elapsed times that are less than or equal to the elapsed time t 2 and since the “best time” player character states associated with the path markers 1-5 and 8 indicate that the “best time” player character is traveling on the ground, the GPU 114 changes color of the path markers 1-5 and 8 from a color c 0 to a color c 1 .
- the GPU 114 changes color of the path markers 6-7 from a color c 0 to a color c 2 .
- the color c 1 is green and the color c 2 is blue, although any color may be assigned to the path markers.
- FIG. 5D is an exemplary illustration of the path markers of FIG. 5A as displayed after an elapsed time t 3 from a beginning of a time-trial run.
- the player character 510 travels along a path designated by positions A-D, and at time t 3 , the player character 510 is located at a position D.
- the “best time” player character (not shown) associated with a “best time” path (represented by circular path markers 1-13) is located either between path markers 12 and 13, or at the path marker 12 in this example. Since the path marker 13 has an elapsed time that is greater than the elapsed time t 3 , the path marker 13 remains unchanged.
- the path markers 1-12 have elapsed times less than or equal to t 3 , the path markers 1-12 will change color.
- the GPU 114 changes color of the path markers 1-5, 8-10 and 12 from a color c 0 to a color c 1 .
- the GPU 114 changes color of the path markers 6-7 from a color c 0 to a color c 2 .
- the path marker 11 has an elapsed time that is less than or equal to the elapsed time t 3 and the “best time” player character state associated with the path marker 11 indicates that the “best time” player character crashed at a location represented by the path marker 11, the GPU 114 changes color of the path marker 11 from a color c 0 to a color c 3 .
- the colors c 1 , c 2 , and c 3 are green, blue and red, respectively, although any color may be assigned to the path markers. It should be noted that in the examples of FIGS. 5A-5D , not all player character states may be utilized during a trial run. For example, the “crashed” state associated with the color c 3 may not be displayed if the “best time” character did not crash during the “best time” run.
- FIG. 5E is a screen-shot of path markers as displayed at an elapsed time t 4 from a beginning of a time-trial run, according to one embodiment of the invention.
- FIG. 5E includes color-coded path markers 515 - 545 that represent a “best time” player character's “best time” path.
- the “best time” player character (not shown) is located either between path markers 540 and 545 or at the path marker 540 .
- the path markers 515 - 540 have a color c 1 since, in this example, the “best time” player character is traveling on the ground at each of the path markers 515 - 540 .
- the color c 1 associated with the path marker 540 is brighter than the color c 1 associated with the path markers 515 - 535 , indicating that the “best time” player character is located at or near the path marker 540 at the elapsed time t 4 .
- the path markers 515 - 535 are colored green, and the path marker 540 is colored a bright green.
- the path marker 545 has a color c 0 since an elapsed time t associated with the path marker 545 is greater than the elapsed time t 4 (i.e., the “best time” player character has not yet reached the path marker 545 at the elapsed time t 4 ).
- one embodiment of the present invention allows a player to view a color-coded previous game path via path markers or net resultant force vectors that dynamically change color dependent upon (1) a current elapsed time t associated with a current time-trial run as compared to elapsed times recorded in a previous game run and (2) previous player character states.
- the player can quickly gauge game progress as measured against the previous game path, and subsequently optimize variations on player character input controls as determined by viewing the previous player character state via color-coded path markers or color-coded net resultant force vectors.
- the previous game path is a “best time” path
- the previous player character states are the “best time” player character states.
- the present invention dynamically computes net resultant force vectors in real time, and displays the net resultant force vectors along paths of one or more player characters.
- the net resultant force vectors may or may nor be color-coded.
- a non-color coded net resultant force vector displayed at a given point on a given path associated with a given player character represents the net resultant force acting on the given player character at the given point.
- a color-coded net resultant force vector acting on a player character offers additional game information, such as the player character state.
- FIG. 6 is an exemplary flowchart of method steps for recording and displaying force vectors along a graphical path, according to one embodiment of the present invention.
- a player of a video game loads the video game into the entertainment system 100 ( FIG. 1 ), and selects a game event via the controller interface 120 ( FIG. 1 ).
- the CPU 112 ( FIG. 1 ) then initiates parameters associated with the selected game event.
- the parameters may include, though are not limited to, environmental parameters (e.g., scenery and weather), racecourse parameters (e.g., track layout and obstacles), and player character, clothing, and mode of transportation parameters (e.g., type of player and bicycle).
- step 610 the player decides whether to race against a previous run via inputs to the controller interface 120 . If, in step 610 , the player decides to race against the previous run, then the CPU 112 retrieves previous run path data from the memory card 122 ( FIG. 1 ) or the previous sub-cache 210 ( FIG. 2 ). In one embodiment of the present invention, the previous run is a “best time” run. Then, in step 615 , the CPU 112 instructs the GPU 114 ( FIG. 1 ) to display the previous run path data, including force vectors representing the previous run path, on a display device (not shown). In one embodiment of the present invention, the force vectors initially have a color co.
- step 620 the player initiates game play, and the CPU 112 instructs the GPU 114 to dynamically change appearance of the force vectors, such as force vector color, as a player character traverses a course associated with the selected game event.
- the CPU 112 compares an elapsed time of the selected game event with elapsed times associated with the previous run force vectors and path markers, and changes colors of those previous run force vectors with elapsed times less than or equal to the elapsed time of the present game event.
- the color of a given force vector changes based upon the previous character state.
- the color of the force vector changes from the color c 0 to a color c 1 if a player character associated with the previous run event (i.e., the previous run player character) is in contact with the ground at the given path marker.
- the color of a given force vector changes from the color c 0 to a color c 2 if the previous run player character is airborne at the given path marker.
- the color of a given force vector changes from the color c 0 to a color c 3 if the previous run player character crashes at the given path marker.
- the CPU 112 computes and stores path data associated with the selected game event in the current sub-cache 220 ( FIG. 2 ), including new force vectors associated with the current player character's path.
- the path data may also comprise (x,y,z) position of path markers, elapsed time at each path marker, player character state at each path marker, and other data important to graphical path characterization.
- Path data may comprise other or different attributes depending on the nature of the game.
- the CPU 112 compares a total elapsed time of the completed game event with a total elapsed time of the previous run event.
- the CPU 112 records data associated with the completed event as previous run path data in the previous sub-cache 210 .
- the previous run event is a “best time” event
- the CPU 112 determines that the total elapsed time of the completed game event is less than the total elapsed time of the “best time” event, then the CPU 112 transfers the path data of the completed game event from the current sub-cache 220 to the previous sub-cache 210 , overwriting any “best time” data stored in the previous sub-cache 210 .
- the current run data may be stored as a previous run data in any of a plurality of previous “non-best time” sub-caches (not shown), or a user may chose whether to save data associated with the completed run to the previous sub-cache 210 or any of the previous “non-best time” sub-caches.
- the player decides whether to play again in step 635 . If the player decides to play again, then the method continues at step 610 . However, if the player decides to not race again, the method is complete.
- the player may instruct the CPU 112 to store the path data of the completed game event on the memory card 122 ( FIG. 1 ) to be retrieved at a later time when the player reloads the video game for another session.
- the GPU 114 does not dynamically change color of the force vectors as a player character traverses a course associated with the selected game event.
- the CPU 112 may store and retrieve data from “non-best time” sub-caches associated with “non-best time” game paths, including, but not limited to, “worst time” paths associated with a current player or any player, “average time” paths associated with a current player or any player, a previous path of a current player selected from one or more previous paths, or a previous path of another player selected from one or more of the other player's previous paths.
- the game software displays the data stored in the “non-best time” sub-caches as a string of dynamically changing colored path markers in which each path marker's color conveys information.
- the game software displays the data stored in the “non-best time” sub-caches as a string of dynamically changing colored force vectors in which each force vector's color, direction, and magnitude conveys information.
- data stored in the “worst time” sub-cache is overwritten with data from the current player's game run if the total elapsed game time associated with the current player's game run is greater than the total elapsed time stored in the “worst time” sub-cache.
- a current player selectively chooses whether to save data associated with a completed current game run in any of the sub-caches. For example, a current player may selectively choose not to allow data associated with a completed current game run to be compiled with the data stored in an “average time” sub-cache.
- the present invention may readily be implemented using configurations other than those described in the preferred embodiments above.
- the present invention may be utilized in any type of racing game such as, but not limited to, car racing, horse jumping, air racing, etc. Therefore, these and other variations upon the preferred embodiments are intended to be covered by the present invention.
Abstract
Description
- This application is a continuation-in-part application of U.S. patent application Ser. No. 10/691,929 filed on Oct. 22, 2003 entitled “System and Method for Recording and Displaying a Graphical Path in a Video Game,” which is hereby incorporated by reference.
- 1. Field of the Invention
- This invention relates generally to computer-generated images and more particularly to a system and method for utilizing vectors in a video game.
- 2. Description of the Background Art
- One appealing aspect of video games is competition. Presently video game developers develop games that capitalize upon the competitive nature of many game players by recording aspects of a previous run, and then allowing subsequent game players access to previous run data. That is, subsequent players may compete against performances of previous players. The previous run may be, for example, based upon a fastest time, most points accumulated, a most efficient use of player character energy reserves, or any combinations of these or other attributes.
- In such video games, a player character associated with a previous run may be simultaneously displayed with a current player character on a graphical display. Thus, a current player may observe a position of the player character associated with the previous run, and directly compete with this previous character. Although the current player may observe a single position of the previous character at any given time, the current player does not observe data associated with a path of the previous character. A current player character with access to path data associated with a previous run may be able to use the path data during game play.
- It would be advantageous to implement a system and method for recording and dynamically displaying data along a graphical path associated with a previous game run, such as a previous “best game” run, thereby allowing a subsequent player of the video game to base competitive game decisions upon a previous game graphical path, and other data associated with the previous game graphical path.
- In accordance with embodiments of the present invention, a system and method is disclosed for utilizing vectors in a video game. The method comprises initiating a current video game session, computing vectors along one or more graphical paths of the video game, where each of the graphical paths are associated with a player character of the video game, and displaying the plurality of vectors along the graphical paths. In a further embodiment of the present invention, the method comprises computing net resultant force vectors, where the vectors are computed in real time and are based upon phenomenological and/or physical laws of nature.
- In another embodiment of the present invention, the method comprises retrieving graphical path data associated with one or more previous runs, displaying the graphical path data as a string of vectors, and determining a color for each vector based upon an elapsed time of the current video game session and an elapsed time associated with the vector. In addition, the color of each vector may be based upon a player character state. The player character state represents a state of the player character, and includes states such as an “on the ground” state, an “airborne” state, and a “crashed” state.”
- A system of the present invention comprises a data cache configured to store graphical path data associated with a current video game session and one or more previous video game sessions, a processor configured to compute a plurality of force vectors associated with one or more graphical paths, and a display device configured to display the plurality of force vectors. In a further embodiment, the processor is configured to compute a color of a force vector from the plurality of force vectors, where the color of each force vector indicates a player character state.
- In further embodiments of the present invention, the graphical path data may be associated with previous game runs, including, but not limited to, previous “best time” runs, previous “worst time” runs, previous “average time” runs, and selectively chosen runs associated with a current game player or other game players.
-
FIG. 1 is an exemplary block diagram of one embodiment of an electronic entertainment system, according to the present invention; -
FIG. 2 is an exemplary block diagram of one embodiment of the data cache ofFIG. 1 ; -
FIG. 3A illustrates a bicycle on a roadway surface, according to an exemplary example of the present invention; -
FIG. 3B illustrates forces acting in the y-z plane on a wheel of the bicycle illustrated inFIG. 3A ; -
FIG. 3C illustrates forces acting in the x-z plane on the wheel of the bicycle illustrated inFIG. 3A ; -
FIG. 3D shows a summation of the forces acting on the wheel illustrated inFIG. 3B ; -
FIG. 3E shows a summation of the forces acting on the wheel illustrated inFIG. 3C ; -
FIG. 3F shows a net resultant force acting on the wheel of the bicycle as a summation of the forces illustrated inFIG. 3D andFIG. 3E ; -
FIG. 4 is an exemplary screen-shot of net resultant force vectors as displayed at a given time during a game run from a vantage point of a player character A, according to one embodiment of the invention; -
FIG. 5A is an illustration of path markers as displayed at a beginning of a time-trial run, according to one embodiment of the present invention; -
FIG. 5B is an illustration of path markers as displayed after an elapsed time t1 from a start of the time-trial run as displayed inFIG. 5A , according to one embodiment of the present invention; -
FIG. 5C is an illustration of path markers as displayed after an elapsed time t2 from a start of the time-trial run as displayed inFIG. 5A , according to another embodiment of the present invention; -
FIG. 5D is an illustration of path markers as displayed after an elapsed time t3 from a start of the time-trial run as displayed inFIG. 5A , according to yet another embodiment of the present invention; -
FIG. 5E is an exemplary screen-shot of path markers as displayed after an elapsed time t4 from a beginning of a time-trial run; and -
FIG. 6 is a flowchart of method steps for recording and displaying vectors along a graphical path, according to one embodiment of the present invention. - In a video game, a game player maneuvers a player character through a game environment from a starting point to an end point. For example, the player character traverses a racecourse in a time trial mode. Video game software records a path of the player character as it moves from the starting point to the end point. The path is recorded as a series of points represented by (x,y,z) Cartesian coordinates, for example. The game software also records a total time that it takes the player character to arrive at the end point and an elapsed time to each point in the path. For example, the elapsed time may be measured from an initial starting time associated with the starting point. The game software may also record other data at each point along the player character's path including, but not limited to, a magnitude and a direction of each force acting on the player character, an instantaneous energy consumption by the player character, a total energy consumption by the player character, and a player character travel state. In one embodiment of the present invention, the player character travel state may indicate that the character is “airborne,” “on the ground,” or “crashed.” Other player character states may be utilized in further embodiments.
-
FIG. 1 is an exemplary block diagram of anelectronic entertainment system 100, according to one embodiment of the present invention. Thesystem 100 comprises amain memory 110, a central processing unit (CPU) 112, adata cache 111, a vectorprocessing unit VPU 113, a graphics processing unit (GPU) 114, an input/output processor (IOP) 116, anIOP memory 118, acontroller interface 120, amemory card 122, a Universal Serial Bus (USB)interface 124, and anIEEE 1394interface 126. Thesystem 100 also comprises an operating system read-only memory (OS ROM) 128, a sound processing unit (SPU) 132, an opticaldisc control unit 134 and a hard disc drive (HDD) 136, which are connected via abus 146 to theIOP 116. Thesystem 100 is preferably an electronic gaming console; however, thesystem 100 may also be implemented as any type of general-purpose computer, set-top box or hand-held gaming device. Alternative embodiments of thesystem 100 may comprise fewer, more, or other components. - A user of the
system 100 provides instructions via thecontroller interface 120 to theCPU 112. For example, the user may instruct theCPU 112 to store certain game information on thememory card 122 or may instruct a character in a game to perform some specified action. Other devices may be connected to thesystem 100 via theUSB interface 124 and theIEEE 1394interface 126. - The
CPU 112, theVPU 113, theGPU 114 and theIOP 116 communicate via asystem bus 144. TheCPU 112 communicates with themain memory 110 via adedicated bus 142, and with thedata cache 111 via adedicated bus 148. TheVPU 113 and theGPU 114 may also communicate via adedicated bus 140. TheCPU 112 executes programs stored in theOS ROM 128 and themain memory 110. Themain memory 110 may contain pre-stored programs and may also contain programs transferred via theIOP 116 from a CD-ROM, a DVD-ROM or other optical disc (not shown) using the opticaldisc control unit 134. TheIOP 116 controls data exchanges between theCPU 112, theVPU 113, theGPU 114 and other devices of thesystem 100, such as thecontroller interface 120. TheSPU 132 executes instructions to produce sound signals that are output on an audio device (not shown). Alternative embodiments may provide different communication paths between the various components. - In one embodiment of the present invention, the
CPU 112 stores and retrieves data associated with previous paths or runs (such as a “best time” run) in thedata cache 111. Thedata cache 111 is discussed further below in conjunction withFIG. 2 . For example, theCPU 112 may retrieve data associated with a “best time” path generated by a developer of the game software and permanently stored on a game disc. TheCPU 112 may instruct the opticaldisc control unit 134 to read the game disc. Thus, players may test themselves against the “best time” and learn an efficient path through the game environment. In addition, a player's “best time” path or any other paths may be stored on thememory card 122, such as a removable flash card. The player then has an option, upon initiation of a new game session, of either racing against a “best time” path as stored on thememory card 122 or any other recorded path. - When the player begins a new run of a racecourse, the game software retrieves recorded data associated with a selected previous run, and then displays the data on a game screen as the character moves through the game environment, allowing the player character to race against the selected previous run. The
GPU 114 executes drawing instructions from theCPU 112 and theVPU 113 to produce images for display on a display device (not shown). Typically, theVPU 113 transforms objects from three-dimensional coordinates to two-dimensional coordinates, and sends the two-dimensional coordinates to theGPU 114. - In one embodiment of the present invention, the game software displays the data associated with the previous run, such as previous run data, as a string of dynamically changing colored path markers situated in the game environment. In another embodiment of the present invention, the string of colored path markers displays a game developer's “best time” path through the game environment. In a further embodiment, the string of colored path markers displays a previous run path through the game environment by one game player from a group of game players.
- In a further embodiment of the present invention, a path marker's color conveys information. For example, when the player selects a racecourse, the game software retrieves recorded data associated with a “best time” path before the current player character begins a time trial run, and displays a string of path markers. Each path marker of the string of path markers is initially a color co. In one embodiment of the present invention, the color co is gray. However, any color may represent the color co. In this embodiment, a path marker colored gray indicates that a player character associated with the “best time” run (hereinafter referred to as the “best time” player character) has not yet reached the path marker's position at a given elapsed time. Thus at any given elapsed time during game play, if the current player character is ahead of the “best time” player character, then the current player observes a string of path markers colored gray that represents the “best time” path.
- However, if the current player character is behind the “best time” player character at any given elapsed time, then the current player observes path markers of a different color. For example, a path marker may change from the color c0 to a color c1 when the “best time” player character is positioned at the path marker and is traveling on the ground. In addition, a path marker may change from the color c0 to a color c2 when the “best time” player character is positioned at the path marker and is traveling in the air. Finally, a path marker may change from the color c0 to a color c3 when the “best time” player character is positioned at the path marker and is crashed. In exemplary embodiments of the present invention, the color c1 is green, the color c2 is blue, and the color c3 is red. The scope of the present invention covers other colors for the above-described path markers. Further, other colored path markers may represent alternate player character states, such as an energy level, an applied force magnitude, and an applied force direction, for example.
-
FIG. 2 is an exemplary block diagram of one embodiment of thedata cache 111 ofFIG. 1 . Theexemplary data cache 111 comprises a previous sub-cache 210 and acurrent sub-cache 220. The previous sub-cache 210 stores data associated with n path markers of a previous run, where n is an integer. In one embodiment of the present invention, the previous sub-cache 210 is a “best time” sub cache which stores data associated with n path markers of a previous “best time” run. In contrast, the current sub-cache 220 stores data generated during a current time-trial run, where the data is associated with m path markers, and m is an integer. Using path marker 1 (i.e., Mrk 1) of the previous sub-cache 210 as an illustrative example of data associated with any path marker, the previous sub-cache 210 may store any of the following data: an x-position ofpath marker 1, a y-position ofpath marker 1, a z-position ofpath marker 1, an elapsed time t atpath marker 1, a player character state atpath marker 1, a player character speed and/or velocity atpath marker 1, an x-component of a net resultant force (Fx) acting on the player character atpath marker 1, a y-component of a net resultant force (Fy) acting on the player character atpath marker 1, a z-component of a net resultant force (Fz) acting on the player character atpath marker 1, and an instantaneous work performed by the net resultant forces (Fx, Fy, Fz) on the player character atpath marker 1. The player character state atpath marker 1 identifies a state of the player character. For example, player character states include, but are not limited to, on the ground, in the air, and crashed. Alternate embodiments of the previous sub-cache 210 may comprise more, less, and/or other data associated with any path marker. Thecurrent sub-cache 220 comprises similar data categories as the previous sub-cache 210, and thus thecurrent sub-cache 220 will not be described in detail. In a further embodiment of the present invention, thedata cache 111 comprises a plurality of sub-caches (not shown) in which each sub-cache of the plurality of sub-caches may store data associated with previous game runs, which may include a “best time” run. - In an exemplary “best time” embodiment of the present invention, the CPU 112 (
FIG. 1 ) executes the video game software. In this embodiment of the present invention, the previous sub-cache 210 is a “best time” sub-cache 210. TheCPU 112 generates and records path markers based upon a distance a player character travels. For example, upon initiation of a time-trial run, when a “best time” path does not yet exist, theCPU 112 executes video game instructions that determine when a player character has traveled a predefined distance s from a starting point. TheCPU 112 then generates apath marker 1 associated with the distance s, and may store the following data associated with thepath marker 1 in the “best time” sub-cache 210 (FIG. 2 ): (x,y,z) coordinates of thepath marker 1, an elapsed time t at thepath marker 1, a player character state at thepath marker 1, a player character speed at thepath marker 1, net resultant forces (Fx,Fy,Fz) acting on the player character at thepath marker 1, and an instantaneous work performed by the player character at thepath marker 1. According to exemplary embodiments of the present invention, the player character state may indicate whether the player character is on the ground, in the air, or crashed at thepath marker 1. Alternative embodiments may comprise additional or different states. - As the player character proceeds along the time-trial run for the above example, the
CPU 112 repeats the above-described process of generating and storing data associated with new path markers in the “best time” sub-cache 210. Preferably, each new path marker is located at a predefined player character travel distance s from a previous path marker position. - Continuing with the above example, in one embodiment of the present invention, if the player character crashes before the player character travels the predefined distance s measured from the previous path marker, then the
CPU 112 generates a path marker at or near the crash site, and stores the data associated with the crash site path marker in the “best time” sub-cache 210. - In an alternate embodiment of the present invention, the
CPU 112 generates and stores data associated with new path markers in the “best time” sub-cache 210, where new path markers are separated by a predefined time interval. - In one embodiment, if the player decides to repeat the time-trial run, the
CPU 112 retrieves the data from the “best time” sub-cache 210, and displays a string of path markers on the display device (not shown). The string of path markers indicates the player character's path recorded during the “best time” run, for example. In one embodiment of the present invention, the “best time” path markers initially have a color co. As the current player character initiates the repeat time-trial run, theCPU 112 generates and records current path markers and data associated with the current path markers to thecurrent sub-cache 220. - Further, during the repeat time-trial run the
CPU 112 compares a current elapsed time with an elapsed time t retrieved from the “best time” sub-cache 210. For each “best time” path marker with an associated elapsed time t that is less than the current elapsed time, theCPU 112 changes the “best time” path marker color based upon the “best time” player character state associated with the “best time” path marker. For example, if the “best time” player character state indicates that the “best time” player character is on the ground at a given path marker, then theCPU 112 instructs theGPU 114 to change the color of the given path marker from c0 to c1. Similarly, if the “best time” player character state indicates that the “best time” player character is in the air at the given path marker, then theCPU 112 instructs theGPU 114 to change the color of the given path marker from c0 to c2. Finally, if the “best time” player character state indicates that the “best time” player character crashed at the given path marker, then theCPU 112 instructs theGPU 114 to change the color of the given path marker from c0 to c3. - At the end of the repeat time-trial run, the
CPU 112 compares an elapsed time t associated with the marker m from the current sub-cache 220 (also referred to as a total elapsed time) with an elapsed time t associated with the marker n from the “best time” sub-cache 210 (also referred to as a “best time” total elapsed time). If the total elapsed time from the repeat time-trial run is less than the “best time” total elapsed time, then theCPU 112 overwrites data of the “best time” sub-cache 210 with data of thecurrent sub-cache 220. Thus, data from the repeat time-trial run is now stored in the “best time” sub-cache 210. Subsequently, data generated from subsequent time-trial runs and stored in thecurrent sub-cache 220 are compared to data stored in the “best time” sub-cache 210. - In an alternate embodiment, data associated with a “best time” run may already be stored or predetermined in the “best time” sub-cache 210 prior to initiation of the time-trial run. In this embodiment, upon initiation of the time trail run, the
CPU 112 retrieves the data from the “best time” sub cache 210 and displays a string of path markers on the display device (not shown) that indicate a path of a “best time” player character. - In exemplary embodiments of the present invention, the
CPU 112 computes a net resultant force acting on the player character at each of the path markers generated during the time-trial run. In one embodiment of the present invention, the net resultant force comprises Cartesian components (Fx, Fy, Fz). However, the scope of the invention covers net resultant force components defined with respect to any coordinate system. These forces may be based upon the governing laws of physics and/or any artificially contrived laws. For example, theCPU 112 may compute the net resultant forces based upon phenomenological laws (i.e., contrived physical laws) that dictate motion of a player character as the player character collides with extraneous objects, for example. Extraneous objects comprise any objects within a game environment, but external to the player character. For example, given that the player character is a rider on a bicycle, extraneous objects may include, but are not limited to, a roadway surface, rocks, ledges, holes, bumps, walls, other game characters, and other player characters. In this embodiment of the present invention, the game software comprises phenomenological laws which theCPU 112 processes to direct the motion of the player character when the player character interacts with an extraneous object. For example, if the player character hits a rock, theCPU 112 retrieves from main memory 110 (FIG. 1 ) a phenomenological law associated with hitting a rock, and uses the phenomenological law and other variables and/or constants retrieved from themain memory 110 or the data cache 111 (e.g., speed and mass of the player character) to compute a net resultant force acting on the player character from the impact with the rock. - In addition, the
CPU 112 may compute net resultant forces based upon real laws of nature, such as the law of gravity, or laws based upon water and wind resistance. For example, resistive frictional forces, such as water and wind resistance forces, may depend upon a shape and a size of a player character or of apparel worn by the player character. - In a further embodiment of the present invention, the
CPU 112 may compute net resultant forces based upon a combination of phenomenological laws and real laws of nature. For example, if a bicycle hits a wall, theCPU 112 may compute a gravitational force, a frictional wind force, and a phenomenological impact force due to a collision with the wall. TheCPU 112 then computes a net resultant force acting on the bicycle as a vector sum of the gravitational force, wind force, and impact force. TheCPU 112 may then instruct the GPU 114 (FIG. 1 ) to display the net resultant force on the display device (not shown) and/or may store the net resultant force in the data cache 111 (FIG. 1 ) for retrieval and display during subsequent game play. - During play of a video game, a multitude of force vectors acting on a player character may be calculated. For example,
FIG. 3A illustrates abicycle 302 on aroadway surface 304, andFIG. 3B illustrates four forces acting on awheel 305 of thebicycle 302 in the y-z plane. For ease of illustration, some of the forces are illustrated to act on a point of contact between theroadway surface 304 and thewheel 305, and other forces are shown to act on a center of mass of the wheel 305 (i.e., geometric center of the wheel in this embodiment). The forces comprise a force of gravity Fg, a frictional force Fu, a pedaling force Fp (i.e., the force transferred from the bicycle pedals (not shown) via a bicycle chain (not shown) to thewheels 305 and applied to the roadway surface 304), and an impulsive force Fi. The impulsive force Fi is typically a force of high magnitude and short duration that may be incurred as thewheel 305 strikes irregularities in theroadway surface 304, for example. -
FIG. 3C illustrates additional forces that may act on thewheel 305 as viewed along line AB (FIG. 3A ), and acting on thewheel 305 in the x-z plane. These forces may comprise a lateral centripetal force Fc applied to thewheel 305 by theroadway surface 304 when thebicycle 302 is turning (e.g., a centripetal force is generated when the driver is leaning into a turn), and a normal force Fn applied perpendicular to theroadway surface 304. -
FIG. 3D shows a summation of all the forces acting on thewheel 305 in the y-z plane as illustrated inFIG. 3B as the y-z resultant force Fyz.FIG. 3E shows a summation of all the forces acting on the wheel in the x-z plane as illustrated inFIG. 3C as the x-z resultant force Fxz. Finally,FIG. 3F shows a net resultant force Fr acting on thewheel 305 of thebicycle 302 as a summation of the y-z resultant force Fyz (FIG. 3D ) and the x-z resultant force Fxz (FIG. 3E ). A net resultant force may also be computed as a summation of resultant forces acting on other components of thebicycle 302 and on a rider (not shown) of thebicycle 302. Other forces may be computed as well, including frictional forces due to air and water (e.g., rain) resistance, and based upon bicycle speed and bicycle structural design and/or rider clothing, for example. - In game play, the net resultant forces may be recorded during a time trial, and then graphically displayed during subsequent runs to inform game players of previous actions and assist in game decisions that may improve game play performance. Additionally, the net resultant forces may be computed and graphically displayed in real time during one or more game players' current runs to provide additional information regarding path smoothness and/or locations of impediments to travel, such as obstacle or barriers. Other applications may be considered as well.
-
FIG. 4 is a screen-shot of net resultant force vectors as displayed at a given time during a game run from a vantage point of player character A, according to one embodiment of the present invention. In theFIG. 4 embodiment of the present invention, the CPU 112 (FIG. 1 ) computes and displays net resultant force vectors along paths associated with each game player. For example, net resultant force vectors, such as netresultant force vectors resultant force vector 420 is displayed along a path traveled by player character B, a netresultant force vector 425 is displayed along a path traveled by player character C, and a netresultant force vector 430 is displayed instantaneously acting on player character A (i.e., current player). Although other net resultant force vectors are computed and displayed along the paths of player characters B, C, and D, they are not readily visible in theFIG. 4A embodiment of the present invention. Therefore, only the most visible net resultant force vectors 405-430 are illustrated in this exemplary example utilizing the present invention. Each net resultant force vector of the net resultant force vectors 405-430 have a magnitude (exemplified by a vector length) and a direction. In one embodiment of the present invention, the net resultant force vectors 405-430 are colored to represent a player character's travel state. The significance of colored net resultant force vectors will be discussed further below in conjunction with path markers andFIGS. 5A-5E . Player characters B, C, and D may be other current player characters competing against player character A, previous player characters representing previous game paths, or a combination of other current player characters and previous player characters. - In another embodiment of the present invention, a player selects player character A (
FIG. 4 ), for example, to compete in a time-trial run. Player character A then competes against a previous run as displayed via a string of net resultant force vectors. The previous run may be a previous “best time” run. The previous run force vectors not only illustrate the previous path of a player character, but also magnitude and direction of net resultant forces encountered by the player character. Player character A may use the information provided by the string of net resultant force vectors to avoid obstacles and select paths in order to improve performance. - In yet a further embodiment of the present invention, a game developer may utilize the computed and displayed net resultant force vectors to develop and debug new versions of video games, or fine tune game dynamics of existing video games. Other uses of net resultant force vectors are contemplated and covered by the present invention.
-
FIG. 5A is an exemplary illustration of path markers as displayed at a beginning of a time-trial run, according to one embodiment of the present invention. At the beginning of the time-trial run, aplayer character 510 is located at a starting position A, and path markers 1-13 represent a path of a “best time” run. In this embodiment of the present invention, the path markers 1-13 are represented by circles, however the scope of the invention covers markers of any shape or size. For descriptive purposes,FIG. 5A is an aerial view of the path markers 1-13 and theplayer character 510. However, during game play, a time-trial run's game environment is typically viewed from a perspective of theplayer character 510. For example, theplayer character 510 may directly observe path markers 1-13. In theFIG. 5A embodiment, since a game player has not yet initiated the time-trial run, path markers visible to the player character 510 (i.e., the path markers 1-13) are a color c0. -
FIG. 5B is an exemplary illustration of the path markers ofFIG. 5A as displayed after an elapsed time t1 from a beginning of a time-trial run. Aplayer character 510 travels along a path designated by positions A-D, and at time t1, theplayer character 510 is located at the position D. However in this example, a “best time” player character (not shown) is located at apath marker 3 at time t1. - In one embodiment of the present invention, the CPU 112 (
FIG. 1 ) instructs the GPU 114 (FIG. 1 ) to change the color of each path marker 1-13 from c0 to another color, dependent upon a “best time” player character state associated with each path marker, if the elapsed time t associated with each respective path marker is less than or equal to the elapsed time t1. For example, in theFIG. 5B embodiment, path markers 1-3 have elapsed times t that are less than or equal to the elapsed time t1, and the “best time” player character states associated with path markers 1-3 indicate that the “best time” player character is traveling on the ground. Therefore, theGPU 114 changes color of the path markers 1-3 from a color c0 to a color c1. Consequently, the path markers 4-13 with elapsed times t-greater than the elapsed time t1 remain unchanged. In an alternate embodiment, the path markers 1-3 disappear from the display device (not shown), if theplayer character 510 is past the path markers 1-3. -
FIG. 5C is an exemplary illustration of the path markers ofFIG. 5A as displayed after an elapsed time t2 from a beginning of a time-trial run. Theplayer character 510 travels along a path designated by positions A-D, and at time t2, theplayer character 510 is located at a position D. However in this example, a “best time” player character (not shown) associated with a “best time” path (represented by circular path markers 1-13) is located either betweenpath markers path marker 8 at the elapsed time t2. Since the path markers 9-13 have elapsed times that are greater than the elapsed time t2, the path markers 9-13 remain unchanged. However, because the path markers 1-8 have elapsed times less than or equal to t2, the path markers 1-8 will change color. For example, since the path markers 1-5 and 8 have elapsed times that are less than or equal to the elapsed time t2 and since the “best time” player character states associated with the path markers 1-5 and 8 indicate that the “best time” player character is traveling on the ground, theGPU 114 changes color of the path markers 1-5 and 8 from a color c0 to a color c1. Furthermore, since the path markers 6-7 have elapsed times that are less than or equal to the elapsed time t2 and the “best time” player character states associated with the path markers 6-7 in this example indicate that the player character is airborne, theGPU 114 changes color of the path markers 6-7 from a color c0 to a color c2. In one embodiment of the present invention the color c1 is green and the color c2 is blue, although any color may be assigned to the path markers. -
FIG. 5D is an exemplary illustration of the path markers ofFIG. 5A as displayed after an elapsed time t3 from a beginning of a time-trial run. Theplayer character 510 travels along a path designated by positions A-D, and at time t3, theplayer character 510 is located at a position D. However, the “best time” player character (not shown) associated with a “best time” path (represented by circular path markers 1-13) is located either betweenpath markers path marker 12 in this example. Since thepath marker 13 has an elapsed time that is greater than the elapsed time t3, thepath marker 13 remains unchanged. However, because the path markers 1-12 have elapsed times less than or equal to t3, the path markers 1-12 will change color. Thus in this example, since the path markers 1-5, 8-10 and 12 have elapsed times that are less than or equal to the elapsed time t3 and “best time” player character states associated with the path markers 1-5, 8-10 and 12 indicate that the “best time” player character is traveling on the ground, theGPU 114 changes color of the path markers 1-5, 8-10 and 12 from a color c0 to a color c1. Furthermore, since the path markers 6-7 have elapsed times that are less than or equal to the elapsed time t3 and the “best time” player character states associated with the path markers 6-7 indicate that the “best time” player character is airborne, theGPU 114 changes color of the path markers 6-7 from a color c0 to a color c2. Finally, since thepath marker 11 has an elapsed time that is less than or equal to the elapsed time t3 and the “best time” player character state associated with thepath marker 11 indicates that the “best time” player character crashed at a location represented by thepath marker 11, theGPU 114 changes color of thepath marker 11 from a color c0 to a color c3. In one embodiment of the present invention the colors c1, c2, and c3 are green, blue and red, respectively, although any color may be assigned to the path markers. It should be noted that in the examples ofFIGS. 5A-5D , not all player character states may be utilized during a trial run. For example, the “crashed” state associated with the color c3 may not be displayed if the “best time” character did not crash during the “best time” run. -
FIG. 5E is a screen-shot of path markers as displayed at an elapsed time t4 from a beginning of a time-trial run, according to one embodiment of the invention.FIG. 5E includes color-coded path markers 515-545 that represent a “best time” player character's “best time” path. At the elapsed time t4, the “best time” player character (not shown) is located either betweenpath markers path marker 540. Thus, the path markers 515-540 have a color c1 since, in this example, the “best time” player character is traveling on the ground at each of the path markers 515-540. In this exemplary embodiment, the color c1 associated with thepath marker 540 is brighter than the color c1 associated with the path markers 515-535, indicating that the “best time” player character is located at or near thepath marker 540 at the elapsed time t4. For example, the path markers 515-535 are colored green, and thepath marker 540 is colored a bright green. Further, thepath marker 545 has a color c0 since an elapsed time t associated with thepath marker 545 is greater than the elapsed time t4 (i.e., the “best time” player character has not yet reached thepath marker 545 at the elapsed time t4). - In operation, one embodiment of the present invention allows a player to view a color-coded previous game path via path markers or net resultant force vectors that dynamically change color dependent upon (1) a current elapsed time t associated with a current time-trial run as compared to elapsed times recorded in a previous game run and (2) previous player character states. Thus, the player can quickly gauge game progress as measured against the previous game path, and subsequently optimize variations on player character input controls as determined by viewing the previous player character state via color-coded path markers or color-coded net resultant force vectors. In one embodiment of the present invention, the previous game path is a “best time” path, and the previous player character states are the “best time” player character states.
- In a further embodiment, the present invention dynamically computes net resultant force vectors in real time, and displays the net resultant force vectors along paths of one or more player characters. The net resultant force vectors may or may nor be color-coded. A non-color coded net resultant force vector displayed at a given point on a given path associated with a given player character represents the net resultant force acting on the given player character at the given point. A color-coded net resultant force vector acting on a player character offers additional game information, such as the player character state.
-
FIG. 6 is an exemplary flowchart of method steps for recording and displaying force vectors along a graphical path, according to one embodiment of the present invention. Instep 605, a player of a video game loads the video game into the entertainment system 100 (FIG. 1 ), and selects a game event via the controller interface 120 (FIG. 1 ). The CPU 112 (FIG. 1 ) then initiates parameters associated with the selected game event. The parameters may include, though are not limited to, environmental parameters (e.g., scenery and weather), racecourse parameters (e.g., track layout and obstacles), and player character, clothing, and mode of transportation parameters (e.g., type of player and bicycle). - In
step 610, the player decides whether to race against a previous run via inputs to thecontroller interface 120. If, instep 610, the player decides to race against the previous run, then theCPU 112 retrieves previous run path data from the memory card 122 (FIG. 1 ) or the previous sub-cache 210 (FIG. 2 ). In one embodiment of the present invention, the previous run is a “best time” run. Then, instep 615, theCPU 112 instructs the GPU 114 (FIG. 1 ) to display the previous run path data, including force vectors representing the previous run path, on a display device (not shown). In one embodiment of the present invention, the force vectors initially have a color co. - In
step 620, the player initiates game play, and theCPU 112 instructs theGPU 114 to dynamically change appearance of the force vectors, such as force vector color, as a player character traverses a course associated with the selected game event. For example, theCPU 112 compares an elapsed time of the selected game event with elapsed times associated with the previous run force vectors and path markers, and changes colors of those previous run force vectors with elapsed times less than or equal to the elapsed time of the present game event. According to exemplary embodiments, the color of a given force vector changes based upon the previous character state. For example, the color of the force vector changes from the color c0 to a color c1 if a player character associated with the previous run event (i.e., the previous run player character) is in contact with the ground at the given path marker. In another embodiment, the color of a given force vector changes from the color c0 to a color c2 if the previous run player character is airborne at the given path marker. In yet a further embodiment, the color of a given force vector changes from the color c0 to a color c3 if the previous run player character crashes at the given path marker. - In addition, the
CPU 112 computes and stores path data associated with the selected game event in the current sub-cache 220 (FIG. 2 ), including new force vectors associated with the current player character's path. The path data may also comprise (x,y,z) position of path markers, elapsed time at each path marker, player character state at each path marker, and other data important to graphical path characterization. Path data may comprise other or different attributes depending on the nature of the game. - In
step 625, theCPU 112 compares a total elapsed time of the completed game event with a total elapsed time of the previous run event. Instep 630, theCPU 112 records data associated with the completed event as previous run path data in the previous sub-cache 210. In one embodiment in which the previous run event is a “best time” event, if theCPU 112 determines that the total elapsed time of the completed game event is less than the total elapsed time of the “best time” event, then theCPU 112 transfers the path data of the completed game event from thecurrent sub-cache 220 to the previous sub-cache 210, overwriting any “best time” data stored in the previous sub-cache 210. In alternate embodiments, the current run data may be stored as a previous run data in any of a plurality of previous “non-best time” sub-caches (not shown), or a user may chose whether to save data associated with the completed run to the previous sub-cache 210 or any of the previous “non-best time” sub-caches. Finally, the player decides whether to play again instep 635. If the player decides to play again, then the method continues atstep 610. However, if the player decides to not race again, the method is complete. - Referring back to step 630, in another embodiment of the present invention the player may instruct the
CPU 112 to store the path data of the completed game event on the memory card 122 (FIG. 1 ) to be retrieved at a later time when the player reloads the video game for another session. Referring back to step 620, in yet another embodiment of the present invention, theGPU 114 does not dynamically change color of the force vectors as a player character traverses a course associated with the selected game event. Finally, referring back to step 610, if the player decides not to race against a previous run event, then instep 640 the player initiates game play, and theCPU 112 computes and dynamically displays force vectors associated with the initiating player character and any other player characters' paths through the game environment. The method then continues atstep 635. - The present invention has been described above with reference to exemplary embodiments. Other embodiments will be apparent to those skilled in the art in light of this disclosure. For example, the
CPU 112 may store and retrieve data from “non-best time” sub-caches associated with “non-best time” game paths, including, but not limited to, “worst time” paths associated with a current player or any player, “average time” paths associated with a current player or any player, a previous path of a current player selected from one or more previous paths, or a previous path of another player selected from one or more of the other player's previous paths. The game software then displays the data stored in the “non-best time” sub-caches as a string of dynamically changing colored path markers in which each path marker's color conveys information. Alternatively, the game software displays the data stored in the “non-best time” sub-caches as a string of dynamically changing colored force vectors in which each force vector's color, direction, and magnitude conveys information. In one embodiment of the present invention, if a current player chooses to race against a “worst time” path, data stored in the “worst time” sub-cache is overwritten with data from the current player's game run if the total elapsed game time associated with the current player's game run is greater than the total elapsed time stored in the “worst time” sub-cache. In other embodiments of the present invention, a current player selectively chooses whether to save data associated with a completed current game run in any of the sub-caches. For example, a current player may selectively choose not to allow data associated with a completed current game run to be compiled with the data stored in an “average time” sub-cache. - The present invention may readily be implemented using configurations other than those described in the preferred embodiments above. In a further example, the present invention may be utilized in any type of racing game such as, but not limited to, car racing, horse jumping, air racing, etc. Therefore, these and other variations upon the preferred embodiments are intended to be covered by the present invention.
Claims (35)
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/737,143 US20050090302A1 (en) | 2003-10-22 | 2003-12-15 | System and method for utilizing vectors in a video game |
EP04256342A EP1541208A1 (en) | 2003-10-22 | 2004-10-14 | System and method for utilizing vectors in a video game |
JP2004306046A JP2005125098A (en) | 2003-10-22 | 2004-10-20 | System and method of using vector in video game |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/691,929 US8133115B2 (en) | 2003-10-22 | 2003-10-22 | System and method for recording and displaying a graphical path in a video game |
US10/737,143 US20050090302A1 (en) | 2003-10-22 | 2003-12-15 | System and method for utilizing vectors in a video game |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/691,929 Continuation-In-Part US8133115B2 (en) | 2003-10-22 | 2003-10-22 | System and method for recording and displaying a graphical path in a video game |
Publications (1)
Publication Number | Publication Date |
---|---|
US20050090302A1 true US20050090302A1 (en) | 2005-04-28 |
Family
ID=34521982
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/691,929 Expired - Fee Related US8133115B2 (en) | 2003-10-22 | 2003-10-22 | System and method for recording and displaying a graphical path in a video game |
US10/737,143 Abandoned US20050090302A1 (en) | 2003-10-22 | 2003-12-15 | System and method for utilizing vectors in a video game |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/691,929 Expired - Fee Related US8133115B2 (en) | 2003-10-22 | 2003-10-22 | System and method for recording and displaying a graphical path in a video game |
Country Status (3)
Country | Link |
---|---|
US (2) | US8133115B2 (en) |
EP (1) | EP1541207A1 (en) |
JP (1) | JP3821822B2 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080094390A1 (en) * | 2004-09-03 | 2008-04-24 | Konami Digital Entertainment Co., Ltd. | Video Generation Device, Load Display Method, Recording Medium, and Program |
US20090118003A1 (en) * | 2004-09-09 | 2009-05-07 | Konami Digital Entertainment Co., Ltd. | Image Creating Device, Load Display Method, Recording Medium, and Program |
US20100035689A1 (en) * | 2008-08-11 | 2010-02-11 | Yaniv Altshuler | System and method for a gaming community |
US9342817B2 (en) | 2011-07-07 | 2016-05-17 | Sony Interactive Entertainment LLC | Auto-creating groups for sharing photos |
US10786736B2 (en) | 2010-05-11 | 2020-09-29 | Sony Interactive Entertainment LLC | Placement of user information in a game space |
US10991140B2 (en) | 2014-01-03 | 2021-04-27 | Creative Technology Ltd | System suitable for one or both of audio processing and graphics processing and a method of processing in association therewith |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060071933A1 (en) | 2004-10-06 | 2006-04-06 | Sony Computer Entertainment Inc. | Application binary interface for multi-pass shaders |
US7636126B2 (en) | 2005-06-22 | 2009-12-22 | Sony Computer Entertainment Inc. | Delay matching in audio/video systems |
CN100410955C (en) * | 2005-09-30 | 2008-08-13 | 腾讯科技(深圳)有限公司 | Method and device for tracking in three-dimensional game scene |
US7965859B2 (en) | 2006-05-04 | 2011-06-21 | Sony Computer Entertainment Inc. | Lighting control of a user environment via a display device |
US7880746B2 (en) | 2006-05-04 | 2011-02-01 | Sony Computer Entertainment Inc. | Bandwidth management through lighting control of a user environment via a display device |
JP4404932B2 (en) * | 2008-02-13 | 2010-01-27 | 株式会社コナミデジタルエンタテインメント | Game system and game terminal |
US10046241B1 (en) * | 2010-11-01 | 2018-08-14 | Ronald Charles Krosky | Output production |
JP2012155551A (en) * | 2011-01-26 | 2012-08-16 | Square Enix Co Ltd | Character input device and character input program |
US9233309B2 (en) * | 2012-12-27 | 2016-01-12 | Sony Computer Entertainment America Llc | Systems and methods for enabling shadow play for video games based on prior user plays |
US10269260B2 (en) | 2015-08-07 | 2019-04-23 | Honda Motor Co., Ltd. | System and method for coaching a driver |
JP6767319B2 (en) * | 2017-07-31 | 2020-10-14 | 株式会社ソニー・インタラクティブエンタテインメント | Information processing device and file copy method |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4068847A (en) * | 1976-06-23 | 1978-01-17 | The Magnavox Company | Chroma and luminance signal generator for video games |
US4116444A (en) * | 1976-07-16 | 1978-09-26 | Atari, Inc. | Method for generating a plurality of moving objects on a video display screen |
US4905147A (en) * | 1986-10-15 | 1990-02-27 | Logg George E | Collision detection system for video system |
US5265888A (en) * | 1990-06-22 | 1993-11-30 | Nintendo Co., Ltd. | Game apparatus and memory cartridge used therefor |
US5269687A (en) * | 1990-08-01 | 1993-12-14 | Atari Games Corporation | System and method for recursive driver training |
US5354202A (en) * | 1990-08-01 | 1994-10-11 | Atari Games Corporation | System and method for driver training with multiple driver competition |
US6141041A (en) * | 1998-06-22 | 2000-10-31 | Lucent Technologies Inc. | Method and apparatus for determination and visualization of player field coverage in a sporting event |
US6220962B1 (en) * | 1997-11-20 | 2001-04-24 | Nintendo Co., Ltd | Video game apparatus having a direction pointing mark and player object displayed on a game screen |
US6233291B1 (en) * | 1995-05-01 | 2001-05-15 | Metalink Ltd. | Method and apparatus for performing data sorting in a decoder |
US6273814B1 (en) * | 1999-10-25 | 2001-08-14 | Square Co., Ltd. | Game apparatus and method for controlling timing for executive action by game character |
Family Cites Families (363)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB958588A (en) | 1960-12-08 | 1964-05-21 | Gen Electric Co Ltd | Improvements in or relating to colour television receivers |
US3147341A (en) | 1961-04-24 | 1964-09-01 | Gen Electric | Automatic brightness-contrast control using photoresistive element to control brightness and agc voltages in response to ambinent light |
US3943277A (en) | 1969-02-20 | 1976-03-09 | The United States Of America As Represented By The Secretary Of The Navy | Digital memory area correlation tracker |
US3717345A (en) * | 1971-03-15 | 1973-02-20 | Prayfel Inc | Computerized race game |
NL7509871A (en) | 1975-08-20 | 1977-02-22 | Philips Nv | COLOR TV CHROMA KEY SIGNAL GENERATOR. |
US4090216A (en) | 1976-05-26 | 1978-05-16 | Gte Sylvania Incorporated | Ambient light contrast and color control circuit |
US4133004A (en) | 1977-11-02 | 1979-01-02 | Hughes Aircraft Company | Video correlation tracker |
US4448200A (en) | 1978-03-27 | 1984-05-15 | University Of Southern California | System and method for dynamic background subtraction |
JPS5513582A (en) | 1978-07-13 | 1980-01-30 | Sanyo Electric Co Ltd | Color television receiver |
US4241341A (en) | 1979-03-05 | 1980-12-23 | Thorson Mark R | Apparatus for scan conversion |
US4321635A (en) | 1979-04-20 | 1982-03-23 | Teac Corporation | Apparatus for selective retrieval of information streams or items |
US4355334A (en) | 1981-05-29 | 1982-10-19 | Zenith Radio Corporation | Dimmer and dimmer override control for a display device |
JPS5846783A (en) | 1981-09-12 | 1983-03-18 | Sony Corp | Chromakey device |
JPS58195957A (en) | 1982-05-11 | 1983-11-15 | Casio Comput Co Ltd | Program starting system by voice |
US4514727A (en) | 1982-06-28 | 1985-04-30 | Trw Inc. | Automatic brightness control apparatus |
US4757525A (en) | 1982-09-29 | 1988-07-12 | Vmx, Inc. | Electronic audio communications system with voice command features |
FI68131C (en) | 1983-06-30 | 1985-07-10 | Valtion Teknillinen | REFERENCE FOR A WINDOW MACHINE WITH A GLASS LED WITH A LASER INDICATOR |
IL69327A (en) | 1983-07-26 | 1986-11-30 | Elscint Ltd | Automatic misregistration correction |
US4675562A (en) | 1983-08-01 | 1987-06-23 | Fairchild Semiconductor Corporation | Method and apparatus for dynamically controlling the timing of signals in automatic test systems |
IL72685A (en) | 1983-08-30 | 1988-08-31 | Gen Electric | Advanced video object generator |
US4646075A (en) | 1983-11-03 | 1987-02-24 | Robert Bosch Corporation | System and method for a data processing pipeline |
US4649504A (en) | 1984-05-22 | 1987-03-10 | Cae Electronics, Ltd. | Optical position and orientation measurement techniques |
US5555532A (en) | 1984-05-23 | 1996-09-10 | The United States Of America As Represented By The Secretary Of The Navy | Method and apparatus for target imaging with sidelooking sonar |
US4658247A (en) | 1984-07-30 | 1987-04-14 | Cornell Research Foundation, Inc. | Pipelined, line buffered real-time color graphics display system |
JPH0746391B2 (en) | 1984-09-14 | 1995-05-17 | 株式会社日立製作所 | Graphic seeding device |
US4672564A (en) | 1984-11-15 | 1987-06-09 | Honeywell Inc. | Method and apparatus for determining location and orientation of objects |
US4683466A (en) | 1984-12-14 | 1987-07-28 | Honeywell Information Systems Inc. | Multiple color generation on a display |
US4737921A (en) | 1985-06-03 | 1988-04-12 | Dynamic Digital Displays, Inc. | Three dimensional medical image display system |
DE3650494T2 (en) | 1985-07-05 | 1996-10-02 | Dainippon Printing Co Ltd | DESIGNING METHOD AND DEVICE OF THREE-DIMENSIONAL CONTAINERS |
JPH0814854B2 (en) | 1985-10-11 | 1996-02-14 | 株式会社日立製作所 | 3D graphic display device |
IL77610A (en) | 1986-01-15 | 1994-01-25 | Technion Res & Dev Foundation | Single camera three-dimensional head position sensing system |
US4843568A (en) | 1986-04-11 | 1989-06-27 | Krueger Myron W | Real time perception of and response to the actions of an unencumbered participant/user |
JP2603947B2 (en) | 1986-09-26 | 1997-04-23 | オリンパス光学工業株式会社 | Apparatus for detecting corresponding areas between primary color images |
US4807158A (en) | 1986-09-30 | 1989-02-21 | Daleco/Ivex Partners, Ltd. | Method and apparatus for sampling images to simulate movement within a multidimensional space |
US4905168A (en) | 1986-10-15 | 1990-02-27 | Atari Games Corporation | Object processing for video system using slips and linked list |
JPS63177193A (en) | 1987-01-19 | 1988-07-21 | 株式会社日立製作所 | Display device |
US4864515A (en) | 1987-03-30 | 1989-09-05 | Honeywell Inc. | Electronic sensing screen for measuring projectile parameters |
FR2613572B1 (en) | 1987-04-03 | 1993-01-22 | Thomson Csf | LIGHT DATA VISUALIZATION SYSTEM WITH IMPROVED READABILITY |
US5014327A (en) | 1987-06-15 | 1991-05-07 | Digital Equipment Corporation | Parallel associative memory having improved selection and decision mechanisms for recognizing and sorting relevant patterns |
US4980823A (en) | 1987-06-22 | 1990-12-25 | International Business Machines Corporation | Sequential prefetching with deconfirmation |
US4860197A (en) | 1987-07-31 | 1989-08-22 | Prime Computer, Inc. | Branch cache system with instruction boundary determination independent of parcel boundary |
US5162781A (en) | 1987-10-02 | 1992-11-10 | Automated Decisions, Inc. | Orientational mouse computer input system |
US5363120A (en) | 1987-10-14 | 1994-11-08 | Wang Laboratories, Inc. | Computer input device using orientation sensor |
US4866637A (en) | 1987-10-30 | 1989-09-12 | International Business Machines Corporation | Pipelined lighting model processing system for a graphics workstation's shading function |
US4901064A (en) | 1987-11-04 | 1990-02-13 | Schlumberger Technologies, Inc. | Normal vector shading for 3-D graphics display system |
US4942538A (en) | 1988-01-05 | 1990-07-17 | Spar Aerospace Limited | Telerobotic tracker |
US5369737A (en) | 1988-03-21 | 1994-11-29 | Digital Equipment Corporation | Normalization of vectors associated with a display pixels of computer generated images |
GB8808608D0 (en) | 1988-04-12 | 1988-05-11 | Boc Group Plc | Dry pump with booster |
US5448687A (en) | 1988-09-13 | 1995-09-05 | Computer Design, Inc. | Computer-assisted design system for flattening a three-dimensional surface and for wrapping a flat shape to a three-dimensional surface |
US4933864A (en) | 1988-10-04 | 1990-06-12 | Transitions Research Corporation | Mobile robot navigation employing ceiling light fixtures |
US5045843B1 (en) | 1988-12-06 | 1996-07-16 | Selectech Ltd | Optical pointing device |
US5222203A (en) | 1989-01-20 | 1993-06-22 | Daikin Industries, Ltd. | Method and apparatus for displaying translucent surface |
US5034986A (en) | 1989-03-01 | 1991-07-23 | Siemens Aktiengesellschaft | Method for detecting and tracking moving objects in a digital image sequence having a stationary background |
US4969036A (en) | 1989-03-31 | 1990-11-06 | Bir Bhanu | System for computing the self-motion of moving images devices |
WO1990016037A1 (en) | 1989-06-20 | 1990-12-27 | Fujitsu Limited | Method for measuring position and posture of object |
US5367615A (en) | 1989-07-10 | 1994-11-22 | General Electric Company | Spatial augmentation of vertices and continuous level of detail transition for smoothly varying terrain polygon density |
FR2652972B1 (en) | 1989-10-06 | 1996-11-29 | Thomson Video Equip | METHOD AND DEVICE FOR INTEGRATING SELF-ADAPTIVE COLOR VIDEO IMAGES. |
GB9001468D0 (en) | 1990-01-23 | 1990-03-21 | Sarnoff David Res Center | Computing multiple motions within an image region |
US5668646A (en) | 1990-02-06 | 1997-09-16 | Canon Kabushiki Kaisha | Apparatus and method for decoding differently encoded multi-level and binary image data, the later corresponding to a color in the original image |
ATE137377T1 (en) | 1990-02-06 | 1996-05-15 | Canon Kk | IMAGE PROCESSING DEVICE |
JPH03274869A (en) | 1990-03-23 | 1991-12-05 | Victor Co Of Japan Ltd | Moving vector detecting circuit |
US5064291A (en) | 1990-04-03 | 1991-11-12 | Hughes Aircraft Company | Method and apparatus for inspection of solder joints utilizing shape determination from shading |
US5128671A (en) | 1990-04-12 | 1992-07-07 | Ltv Aerospace And Defense Company | Control device having multiple degrees of freedom |
DE69130549T2 (en) | 1990-06-11 | 1999-05-27 | Hitachi Ltd | Device for generating an object movement path |
US5253339A (en) | 1990-07-26 | 1993-10-12 | Sun Microsystems, Inc. | Method and apparatus for adaptive Phong shading |
US5208763A (en) | 1990-09-14 | 1993-05-04 | New York University | Method and apparatus for determining position and orientation of mechanical objects |
US5274560A (en) | 1990-12-03 | 1993-12-28 | Audio Navigation Systems, Inc. | Sensor free vehicle navigation system utilizing a voice input/output interface for routing a driver from his source point to his destination point |
US5268996A (en) | 1990-12-20 | 1993-12-07 | General Electric Company | Computer image generation method for determination of total pixel illumination due to plural light sources |
US5128794A (en) | 1990-12-31 | 1992-07-07 | Honeywell Inc. | Scanning laser helmet mounted sight |
US5534917A (en) | 1991-05-09 | 1996-07-09 | Very Vivid, Inc. | Video image based control system |
US5548667A (en) | 1991-05-24 | 1996-08-20 | Sony Corporation | Image processing system and method thereof in which three dimensional shape is reproduced from two dimensional image data |
US5227985A (en) | 1991-08-19 | 1993-07-13 | University Of Maryland | Computer vision system for position monitoring in three dimensions using non-coplanar light sources attached to a monitored object |
US5305389A (en) | 1991-08-30 | 1994-04-19 | Digital Equipment Corporation | Predictive cache system |
US5212888A (en) | 1991-09-16 | 1993-05-25 | Calcomp Inc. | Dual function sensor for a pen plotter |
US5537638A (en) | 1991-10-25 | 1996-07-16 | Hitachi, Ltd. | Method and system for image mapping |
US5335557A (en) | 1991-11-26 | 1994-08-09 | Taizo Yasutake | Touch sensitive input control device |
US5631697A (en) | 1991-11-27 | 1997-05-20 | Hitachi, Ltd. | Video camera capable of automatic target tracking |
US5734384A (en) | 1991-11-29 | 1998-03-31 | Picker International, Inc. | Cross-referenced sectioning and reprojection of diagnostic image volumes |
US5230623A (en) | 1991-12-10 | 1993-07-27 | Radionics, Inc. | Operating pointer with interactive computergraphics |
US5680487A (en) | 1991-12-23 | 1997-10-21 | Texas Instruments Incorporated | System and method for determining optical flow |
US5590248A (en) | 1992-01-02 | 1996-12-31 | General Electric Company | Method for reducing the complexity of a polygonal mesh |
US5953485A (en) | 1992-02-07 | 1999-09-14 | Abecassis; Max | Method and system for maintaining audio during video control |
EP0627103B1 (en) | 1992-02-18 | 1996-01-10 | EVANS & SUTHERLAND COMPUTER CORPORATION | Image texturing system having theme cells |
US5577179A (en) | 1992-02-25 | 1996-11-19 | Imageware Software, Inc. | Image editing system |
US5307137A (en) | 1992-03-16 | 1994-04-26 | Mark F. Jones | Terrain imaging apparatus and method |
JP3107452B2 (en) | 1992-04-28 | 2000-11-06 | 株式会社日立製作所 | Texture mapping method and apparatus |
US5450504A (en) | 1992-05-19 | 1995-09-12 | Calia; James | Method for finding a most likely matching of a target facial image in a data base of facial images |
US5366376A (en) * | 1992-05-22 | 1994-11-22 | Atari Games Corporation | Driver training system and method with performance data feedback |
JP3391405B2 (en) | 1992-05-29 | 2003-03-31 | 株式会社エフ・エフ・シー | Object identification method in camera image |
US5473736A (en) | 1992-06-08 | 1995-12-05 | Chroma Graphics | Method and apparatus for ordering and remapping colors in images of real two- and three-dimensional objects |
IL102289A (en) | 1992-06-24 | 1997-08-14 | R Technologies Ltd B V | Method and system for processing moving images |
US5446714A (en) | 1992-07-21 | 1995-08-29 | Pioneer Electronic Corporation | Disc changer and player that reads and stores program data of all discs prior to reproduction and method of reproducing music on the same |
DE69331031T2 (en) | 1992-07-27 | 2002-07-04 | Matsushita Electric Ind Co Ltd | Device for parallel imaging |
US5361385A (en) | 1992-08-26 | 1994-11-01 | Reuven Bakalash | Parallel computing system for volumetric modeling, data processing and visualization |
JP3244798B2 (en) | 1992-09-08 | 2002-01-07 | 株式会社東芝 | Moving image processing device |
US5982352A (en) | 1992-09-18 | 1999-11-09 | Pryor; Timothy R. | Method for providing human input to a computer |
JP2682559B2 (en) | 1992-09-30 | 1997-11-26 | インターナショナル・ビジネス・マシーンズ・コーポレイション | Apparatus and method for displaying image of object on display device and computer graphics display system |
GB2271259A (en) | 1992-10-02 | 1994-04-06 | Canon Res Ct Europe Ltd | Processing image data |
US5469193A (en) | 1992-10-05 | 1995-11-21 | Prelude Technology Corp. | Cordless pointing apparatus |
EP0598295B1 (en) | 1992-11-17 | 1998-10-14 | Matsushita Electric Industrial Co., Ltd. | Video and audio signal multiplexing apparatus and separating apparatus |
US5387943A (en) | 1992-12-21 | 1995-02-07 | Tektronix, Inc. | Semiautomatic lip sync recovery system |
US5890122A (en) | 1993-02-08 | 1999-03-30 | Microsoft Corporation | Voice-controlled computer simulateously displaying application menu and list of available commands |
US5660547A (en) * | 1993-02-17 | 1997-08-26 | Atari Games Corporation | Scenario development system for vehicle simulators |
US5508744A (en) | 1993-03-12 | 1996-04-16 | Thomson Consumer Electronics, Inc. | Video signal compression with removal of non-correlated motion vectors |
JP3679426B2 (en) | 1993-03-15 | 2005-08-03 | マサチューセッツ・インスティチュート・オブ・テクノロジー | A system that encodes image data into multiple layers, each representing a coherent region of motion, and motion parameters associated with the layers. |
EP0622747B1 (en) | 1993-04-01 | 2000-05-31 | Sun Microsystems, Inc. | Method and apparatus for an adaptive texture mapping controller |
GB9308952D0 (en) | 1993-04-30 | 1993-06-16 | Philips Electronics Uk Ltd | Tracking objects in video sequences |
US5297061A (en) | 1993-05-19 | 1994-03-22 | University Of Maryland | Three dimensional pointing device monitored by computer vision |
EP0633546B1 (en) | 1993-07-02 | 2003-08-27 | Siemens Corporate Research, Inc. | Background recovery in monocular vision |
JPH0757117A (en) | 1993-07-09 | 1995-03-03 | Silicon Graphics Inc | Forming method of index to texture map and computer control display system |
US5550960A (en) | 1993-08-02 | 1996-08-27 | Sun Microsystems, Inc. | Method and apparatus for performing dynamic texture mapping for complex surfaces |
US5598514A (en) | 1993-08-09 | 1997-01-28 | C-Cube Microsystems | Structure and method for a multistandard video encoder/decoder |
JP2916076B2 (en) | 1993-08-26 | 1999-07-05 | シャープ株式会社 | Image display device |
DE69419291T2 (en) | 1993-09-03 | 1999-12-30 | Canon Kk | Form measuring apparatus |
JP2552427B2 (en) | 1993-12-28 | 1996-11-13 | コナミ株式会社 | Tv play system |
FR2714502A1 (en) | 1993-12-29 | 1995-06-30 | Philips Laboratoire Electroniq | An image processing method and apparatus for constructing from a source image a target image with perspective change. |
US5559950A (en) | 1994-02-02 | 1996-09-24 | Video Lottery Technologies, Inc. | Graphics processor enhancement unit |
US5699497A (en) | 1994-02-17 | 1997-12-16 | Evans & Sutherland Computer Corporation | Rendering global macro texture, for producing a dynamic image, as on computer generated terrain, seen from a moving viewpoint |
US5611000A (en) | 1994-02-22 | 1997-03-11 | Digital Equipment Corporation | Spline-based image registration |
DE69515565T2 (en) * | 1994-07-15 | 2000-11-02 | Dade Chemistry Systems Inc | Analyzer |
AU699764B2 (en) | 1994-09-06 | 1998-12-17 | Research Foundation Of The State University Of New York, The | Apparatus and method for real-time volume visualization |
US5920842A (en) | 1994-10-12 | 1999-07-06 | Pixel Instruments | Signal synchronization |
JP2642070B2 (en) | 1994-11-07 | 1997-08-20 | インターナショナル・ビジネス・マシーンズ・コーポレイション | Method and system for generating quadrilateral mesh |
US5649032A (en) | 1994-11-14 | 1997-07-15 | David Sarnoff Research Center, Inc. | System for automatically aligning images to form a mosaic image |
EP0715453B1 (en) | 1994-11-28 | 2014-03-26 | Canon Kabushiki Kaisha | Camera controller |
JP3578498B2 (en) | 1994-12-02 | 2004-10-20 | 株式会社ソニー・コンピュータエンタテインメント | Image information processing device |
GB2295936B (en) | 1994-12-05 | 1997-02-05 | Microsoft Corp | Progressive image transmission using discrete wavelet transforms |
GB9501832D0 (en) | 1995-01-31 | 1995-03-22 | Videologic Ltd | Texturing and shading of 3-d images |
US5818553A (en) | 1995-04-10 | 1998-10-06 | Norand Corporation | Contrast control for a backlit LCD |
US5757360A (en) | 1995-05-03 | 1998-05-26 | Mitsubishi Electric Information Technology Center America, Inc. | Hand held computer control device |
US5672820A (en) | 1995-05-16 | 1997-09-30 | Boeing North American, Inc. | Object location identification system for providing location data of an object being pointed at by a pointing device |
US5913727A (en) | 1995-06-02 | 1999-06-22 | Ahdoot; Ned | Interactive movement and contact simulation game |
US5572261A (en) | 1995-06-07 | 1996-11-05 | Cooper; J. Carl | Automatic audio to video timing measurement device and method |
US5617407A (en) | 1995-06-21 | 1997-04-01 | Bareis; Monica M. | Optical disk having speech recognition templates for information access |
US5805745A (en) | 1995-06-26 | 1998-09-08 | Lucent Technologies Inc. | Method for locating a subject's lips in a facial image |
US6002738A (en) | 1995-07-07 | 1999-12-14 | Silicon Graphics, Inc. | System and method of performing tomographic reconstruction and volume rendering using texture mapping |
US5704024A (en) | 1995-07-20 | 1997-12-30 | Silicon Graphics, Inc. | Method and an apparatus for generating reflection vectors which can be unnormalized and for using these reflection vectors to index locations on an environment map |
US5870097A (en) | 1995-08-04 | 1999-02-09 | Microsoft Corporation | Method and system for improving shadowing in a graphics rendering system |
US5852443A (en) | 1995-08-04 | 1998-12-22 | Microsoft Corporation | Method and system for memory decomposition in a graphics rendering system |
US6016150A (en) | 1995-08-04 | 2000-01-18 | Microsoft Corporation | Sprite compositor and method for performing lighting and shading operations using a compositor to combine factored image layers |
US5977977A (en) | 1995-08-04 | 1999-11-02 | Microsoft Corporation | Method and system for multi-pass rendering |
US5808617A (en) | 1995-08-04 | 1998-09-15 | Microsoft Corporation | Method and system for depth complexity reduction in a graphics rendering system |
US5864342A (en) | 1995-08-04 | 1999-01-26 | Microsoft Corporation | Method and system for rendering graphical objects to image chunks |
JP3203160B2 (en) | 1995-08-09 | 2001-08-27 | 三菱電機株式会社 | Volume rendering apparatus and method |
US5856844A (en) | 1995-09-21 | 1999-01-05 | Omniplanar, Inc. | Method and apparatus for determining position and orientation |
US5959673A (en) | 1995-10-05 | 1999-09-28 | Microsoft Corporation | Transform coding of dense motion vector fields for frame and object based video coding applications |
US5818424A (en) | 1995-10-19 | 1998-10-06 | International Business Machines Corporation | Rod shaped device and data acquisition apparatus for determining the position and orientation of an object in space |
KR100261076B1 (en) | 1995-11-09 | 2000-07-01 | 윤종용 | Rendering method and apparatus of performing bump mapping and phong shading at the same time |
US5825308A (en) | 1996-11-26 | 1998-10-20 | Immersion Human Interface Corporation | Force feedback interface having isotonic and isometric functionality |
FR2741769B1 (en) | 1995-11-23 | 1997-12-19 | Thomson Broadcast Systems | PROCESS FOR PROCESSING THE SIGNAL CONSTITUTED BY A SUBJECT EVOLVING BEFORE A COLORED BACKGROUND AND DEVICE IMPLEMENTING THIS METHOD |
FR2741770B1 (en) | 1995-11-23 | 1998-01-02 | Thomson Broadcast Systems | METHOD FOR CALCULATING A CUTTING KEY OF A SUBJECT EVOLVING IN FRONT OF A COLORED BACKGROUND AND DEVICE IMPLEMENTING THIS METHOD |
US6028593A (en) | 1995-12-01 | 2000-02-22 | Immersion Corporation | Method and apparatus for providing simulated physical interactions within computer generated environments |
US5963209A (en) | 1996-01-11 | 1999-10-05 | Microsoft Corporation | Encoding and progressive transmission of progressive meshes |
US5574836A (en) | 1996-01-22 | 1996-11-12 | Broemmelsiek; Raymond M. | Interactive display apparatus and method with viewer position compensation |
JP3668547B2 (en) | 1996-01-29 | 2005-07-06 | ヤマハ株式会社 | Karaoke equipment |
US6049619A (en) | 1996-02-12 | 2000-04-11 | Sarnoff Corporation | Method and apparatus for detecting moving objects in two- and three-dimensional scenes |
AU2123297A (en) | 1996-02-12 | 1997-08-28 | Golf Age Technologies | Golf driving range distancing apparatus and methods |
US6009188A (en) | 1996-02-16 | 1999-12-28 | Microsoft Corporation | Method and system for digital plenoptic imaging |
US5982390A (en) | 1996-03-25 | 1999-11-09 | Stan Stoneking | Controlling personality manifestations by objects in a computer-assisted animation environment |
US5764803A (en) | 1996-04-03 | 1998-06-09 | Lucent Technologies Inc. | Motion-adaptive modelling of scene content for very low bit rate model-assisted coding of video sequences |
US5889505A (en) | 1996-04-04 | 1999-03-30 | Yale University | Vision-based six-degree-of-freedom computer input device |
US6348921B1 (en) | 1996-04-12 | 2002-02-19 | Ze Hong Zhao | System and method for displaying different portions of an object in different levels of detail |
US6018347A (en) | 1996-04-12 | 2000-01-25 | Multigen Paradigm, Inc. | Methods and apparatus for rendering three-dimensional images |
US5894308A (en) | 1996-04-30 | 1999-04-13 | Silicon Graphics, Inc. | Interactively reducing polygon count in three-dimensional graphic objects |
GB9609197D0 (en) | 1996-05-02 | 1996-07-03 | Philips Electronics Nv | Process control with evaluation of stored referential expressions |
US5805170A (en) | 1996-05-07 | 1998-09-08 | Microsoft Corporation | Systems and methods for wrapping a closed polygon around an object |
US5769718A (en) | 1996-05-15 | 1998-06-23 | Rieder; William R. | Video game apparatus and medium readable by a computer stored with video game program |
US6034693A (en) | 1996-05-28 | 2000-03-07 | Namco Ltd. | Image synthesizing apparatus, image synthesizing method and information storage medium |
JP3664336B2 (en) | 1996-06-25 | 2005-06-22 | 株式会社日立メディコ | Method and apparatus for setting viewpoint position and gaze direction in 3D image construction method |
WO1998002223A1 (en) | 1996-07-11 | 1998-01-22 | Sega Enterprises, Ltd. | Voice recognizer, voice recognizing method and game machine using them |
GB9616184D0 (en) | 1996-08-01 | 1996-09-11 | Philips Electronics Nv | Virtual environment navigation |
US5933150A (en) | 1996-08-06 | 1999-08-03 | Interval Research Corporation | System for image manipulation and animation using embedded constraint graphics |
US5781194A (en) | 1996-08-29 | 1998-07-14 | Animatek International, Inc. | Real-time projection of voxel-based object |
JP3358169B2 (en) | 1996-08-30 | 2002-12-16 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Mirror surface rendering method and apparatus |
JP3387750B2 (en) | 1996-09-02 | 2003-03-17 | 株式会社リコー | Shading processing equipment |
US5786801A (en) | 1996-09-06 | 1998-07-28 | Sony Corporation | Back light control apparatus and method for a flat display system |
US5854632A (en) | 1996-10-15 | 1998-12-29 | Real 3D | Apparatus and method for simulating specular reflection in a computer graphics/imaging system |
US5886702A (en) | 1996-10-16 | 1999-03-23 | Real-Time Geometry Corporation | System and method for computer modeling of 3D objects or surfaces by mesh constructions having optimal quality characteristics and dynamic resolution capabilities |
US5812136A (en) | 1996-10-30 | 1998-09-22 | Microsoft Corporation | System and method for fast rendering of a three dimensional graphical object |
US5935198A (en) | 1996-11-22 | 1999-08-10 | S3 Incorporated | Multiplier with selectable booth encoders for performing 3D graphics interpolations with two multiplies in a single pass through the multiplier |
US6157386A (en) | 1997-10-10 | 2000-12-05 | Cirrus Logic, Inc | MIP map blending in a graphics processor |
US5899810A (en) | 1997-01-24 | 1999-05-04 | Kaon Interactive Corporation | Distributed game architecture to overcome system latency |
JPH10211358A (en) | 1997-01-28 | 1998-08-11 | Sega Enterp Ltd | Game apparatus |
US6121953A (en) | 1997-02-06 | 2000-09-19 | Modern Cartoons, Ltd. | Virtual reality system for sensing facial movements |
JP3870422B2 (en) | 1997-02-20 | 2007-01-17 | ソニー株式会社 | Video signal processing apparatus and method, image composition apparatus, and editing apparatus |
US5870098A (en) | 1997-02-26 | 1999-02-09 | Evans & Sutherland Computer Corporation | Method for rendering shadows on a graphical display |
US5880736A (en) | 1997-02-28 | 1999-03-09 | Silicon Graphics, Inc. | Method system and computer program product for shading |
US5949424A (en) | 1997-02-28 | 1999-09-07 | Silicon Graphics, Inc. | Method, system, and computer program product for bump mapping in tangent space |
US6803964B1 (en) | 1997-03-21 | 2004-10-12 | International Business Machines Corporation | Method and apparatus for processing digital data |
US6137492A (en) | 1997-04-03 | 2000-10-24 | Microsoft Corporation | Method and system for adaptive refinement of progressive meshes |
JP4244391B2 (en) | 1997-04-04 | 2009-03-25 | ソニー株式会社 | Image conversion apparatus and image conversion method |
US6058397A (en) | 1997-04-08 | 2000-05-02 | Mitsubishi Electric Information Technology Center America, Inc. | 3D virtual environment creation management and delivery system |
US5864742A (en) | 1997-04-11 | 1999-01-26 | Eastman Kodak Company | Copy restrictive system using microdots to restrict copying of color-reversal documents |
US5917937A (en) | 1997-04-15 | 1999-06-29 | Microsoft Corporation | Method for performing stereo matching to recover depths, colors and opacities of surface elements |
US6130673A (en) | 1997-04-18 | 2000-10-10 | Silicon Graphics, Inc. | Editing a surface |
US6175367B1 (en) | 1997-04-23 | 2001-01-16 | Siligon Graphics, Inc. | Method and system for real time illumination of computer generated images |
US5912830A (en) | 1997-04-30 | 1999-06-15 | Hewlett-Packard Co. | System and method for conditionally calculating exponential values in a geometry accelerator |
CA2250021C (en) | 1997-05-19 | 2007-02-06 | Matsushita Electric Industrial Co., Ltd. | Graphic display apparatus, synchronous reproduction method, and av synchronous reproduction apparatus |
US5817494A (en) | 1997-05-21 | 1998-10-06 | Incyte Pharmaceuticals, Inc. | Ubiquitin conjugation proteins |
GB2326781B (en) | 1997-05-30 | 2001-10-10 | British Broadcasting Corp | Video and audio signal processing |
NL1006177C2 (en) | 1997-05-30 | 1998-12-07 | Delaval Stork V O F | Rotor shaft for a rotating machine and rotating machine provided with such a rotor shaft. |
US6222555B1 (en) | 1997-06-18 | 2001-04-24 | Christofferson Enterprises, Llc | Method for automatically smoothing object level of detail transitions for regular objects in a computer graphics display system |
US6072504A (en) | 1997-06-20 | 2000-06-06 | Lucent Technologies Inc. | Method and apparatus for tracking, storing, and synthesizing an animated version of object motion |
US6208347B1 (en) | 1997-06-23 | 2001-03-27 | Real-Time Geometry Corporation | System and method for computer modeling of 3D objects and 2D images by mesh constructions that incorporate non-spatial data such as color or texture |
US6226006B1 (en) | 1997-06-27 | 2001-05-01 | C-Light Partners, Inc. | Method and apparatus for providing shading in a graphic display system |
US5990901A (en) | 1997-06-27 | 1999-11-23 | Microsoft Corporation | Model based image editing and correction |
US6049636A (en) | 1997-06-27 | 2000-04-11 | Microsoft Corporation | Determining a rectangular box encompassing a digital picture within a digital image |
US5914724A (en) | 1997-06-30 | 1999-06-22 | Sun Microsystems, Inc | Lighting unit for a three-dimensional graphics accelerator with improved handling of incoming color values |
JP3372832B2 (en) | 1997-07-25 | 2003-02-04 | コナミ株式会社 | GAME DEVICE, GAME IMAGE PROCESSING METHOD, AND COMPUTER-READABLE RECORDING MEDIUM CONTAINING GAME IMAGE PROCESSING PROGRAM |
US6018349A (en) | 1997-08-01 | 2000-01-25 | Microsoft Corporation | Patch-based alignment method and apparatus for construction of image mosaics |
US6044181A (en) | 1997-08-01 | 2000-03-28 | Microsoft Corporation | Focal length estimation method and apparatus for construction of panoramic mosaic images |
US6009190A (en) | 1997-08-01 | 1999-12-28 | Microsoft Corporation | Texture map construction method and apparatus for displaying panoramic image mosaics |
US5987164A (en) | 1997-08-01 | 1999-11-16 | Microsoft Corporation | Block adjustment method and apparatus for construction of image mosaics |
US5986668A (en) | 1997-08-01 | 1999-11-16 | Microsoft Corporation | Deghosting method and apparatus for construction of image mosaics |
US6720949B1 (en) | 1997-08-22 | 2004-04-13 | Timothy R. Pryor | Man machine interfaces and applications |
AUPO894497A0 (en) | 1997-09-02 | 1997-09-25 | Xenotech Research Pty Ltd | Image processing method and apparatus |
US6496189B1 (en) | 1997-09-29 | 2002-12-17 | Skyline Software Systems Ltd. | Remote landscape display and pilot training |
US6101289A (en) | 1997-10-15 | 2000-08-08 | Electric Planet, Inc. | Method and apparatus for unencumbered capture of an object |
US6072494A (en) | 1997-10-15 | 2000-06-06 | Electric Planet, Inc. | Method and apparatus for real-time gesture recognition |
US6031934A (en) | 1997-10-15 | 2000-02-29 | Electric Planet, Inc. | Computer vision system for subject characterization |
US6037947A (en) | 1997-10-16 | 2000-03-14 | Sun Microsystems, Inc. | Graphics accelerator with shift count generation for handling potential fixed-point numeric overflows |
US5905894A (en) | 1997-10-29 | 1999-05-18 | Microsoft Corporation | Meta-programming methods and apparatus |
JP3890781B2 (en) * | 1997-10-30 | 2007-03-07 | 株式会社セガ | Computer-readable storage medium, game device, and game image display method |
US6320580B1 (en) | 1997-11-07 | 2001-11-20 | Sega Enterprises, Ltd. | Image processing apparatus |
JP3119608B2 (en) * | 1997-11-19 | 2000-12-25 | コナミ株式会社 | Competitive video game device, character movement control method in competitive video game, and recording medium storing character movement control program |
US6162123A (en) | 1997-11-25 | 2000-12-19 | Woolston; Thomas G. | Interactive electronic sword game |
US6010403A (en) | 1997-12-05 | 2000-01-04 | Lbe Technologies, Inc. | System and method for displaying an interactive event |
DK0928659T3 (en) | 1997-12-19 | 2002-08-12 | Esab Ab | Welding apparatus |
US6356288B1 (en) | 1997-12-22 | 2002-03-12 | U.S. Philips Corporation | Diversion agent uses cinematographic techniques to mask latency |
JPH11203501A (en) | 1998-01-14 | 1999-07-30 | Sega Enterp Ltd | Picture processor and picture processing method |
US6172354B1 (en) | 1998-01-28 | 2001-01-09 | Microsoft Corporation | Operator input device |
US6014144A (en) | 1998-02-03 | 2000-01-11 | Sun Microsystems, Inc. | Rapid computation of local eye vectors in a fixed point lighting unit |
US6850236B2 (en) | 1998-02-17 | 2005-02-01 | Sun Microsystems, Inc. | Dynamically adjusting a sample-to-pixel filter in response to user input and/or sensor input |
US6577312B2 (en) | 1998-02-17 | 2003-06-10 | Sun Microsystems, Inc. | Graphics system configured to filter samples using a variable support filter |
US6275187B1 (en) | 1998-03-03 | 2001-08-14 | General Electric Company | System and method for directing an adaptive antenna array |
US6181988B1 (en) | 1998-04-07 | 2001-01-30 | Raytheon Company | Guidance system having a body fixed seeker with an adjustable look angle |
US6100898A (en) | 1998-04-08 | 2000-08-08 | Webtv Networks, Inc. | System and method of selecting level of detail in texture mapping |
US6578197B1 (en) | 1998-04-08 | 2003-06-10 | Silicon Graphics, Inc. | System and method for high-speed execution of graphics application programs including shading language instructions |
US6313841B1 (en) | 1998-04-13 | 2001-11-06 | Terarecon, Inc. | Parallel volume rendering system with a resampling module for parallel and perspective projections |
JP3829014B2 (en) | 1998-06-03 | 2006-10-04 | 株式会社コナミデジタルエンタテインメント | Video game equipment |
WO1999064944A2 (en) | 1998-06-08 | 1999-12-16 | Microsoft Corporation | Compression of time-dependent geometry |
JP2000090289A (en) | 1998-07-13 | 2000-03-31 | Sony Corp | Device and method for processing image and medium |
US6563499B1 (en) | 1998-07-20 | 2003-05-13 | Geometrix, Inc. | Method and apparatus for generating a 3D region from a surrounding imagery |
US6646639B1 (en) | 1998-07-22 | 2003-11-11 | Nvidia Corporation | Modified method and apparatus for improved occlusion culling in graphics systems |
GB9817834D0 (en) | 1998-08-14 | 1998-10-14 | British Telecomm | Predicting avatar movement in a distributed virtual environment |
AU5688199A (en) | 1998-08-20 | 2000-03-14 | Raycer, Inc. | System, apparatus and method for spatially sorting image data in a three-dimensional graphics pipeline |
US6771264B1 (en) | 1998-08-20 | 2004-08-03 | Apple Computer, Inc. | Method and apparatus for performing tangent space lighting and bump mapping in a deferred shading graphics processor |
US7251315B1 (en) | 1998-09-21 | 2007-07-31 | Microsoft Corporation | Speech processing for telephony API |
TW495710B (en) | 1998-10-15 | 2002-07-21 | Primax Electronics Ltd | Voice control module for control of game controller |
JP2000137828A (en) | 1998-10-30 | 2000-05-16 | Fujitsu Ltd | Shading arithmetic unit and image processor using the arithmetic unit |
GB2343598B (en) | 1998-11-06 | 2003-03-19 | Videologic Ltd | Image processing apparatus |
US6342885B1 (en) | 1998-11-12 | 2002-01-29 | Tera Recon Inc. | Method and apparatus for illuminating volume data in a rendering pipeline |
US6396490B1 (en) | 1998-12-04 | 2002-05-28 | Intel Corporation | Efficient representation of connectivity information in progressive mesh update record |
JP3748172B2 (en) | 1998-12-09 | 2006-02-22 | 富士通株式会社 | Image processing device |
JP2000181676A (en) | 1998-12-11 | 2000-06-30 | Nintendo Co Ltd | Image processor |
US6738059B1 (en) | 1998-12-18 | 2004-05-18 | Kabushiki Kaisha Sega Enterprises | Apparatus and methods for image processing using mixed display objects |
US6414960B1 (en) | 1998-12-29 | 2002-07-02 | International Business Machines Corp. | Apparatus and method of in-service audio/video synchronization testing |
US6356263B2 (en) | 1999-01-27 | 2002-03-12 | Viewpoint Corporation | Adaptive subdivision of mesh models |
JP2000218036A (en) | 1999-01-28 | 2000-08-08 | Square Co Ltd | Game device, game control method and recording medium therefor |
JP2000229172A (en) | 1999-02-10 | 2000-08-22 | Konami Co Ltd | Game system and computer readable storage medium on which game program is recorded |
JP3972230B2 (en) | 1999-02-15 | 2007-09-05 | 株式会社セガ | GAME DEVICE, GAME DEVICE CONTROL METHOD, AND RECORDING MEDIUM |
JP2000237453A (en) | 1999-02-16 | 2000-09-05 | Taito Corp | Problem solution type vehicle game device |
JP2000311251A (en) | 1999-02-26 | 2000-11-07 | Toshiba Corp | Device and method for generating animation and storage medium |
US6313842B1 (en) | 1999-03-03 | 2001-11-06 | Discreet Logic Inc. | Generating image data |
DE19917660A1 (en) | 1999-04-19 | 2000-11-02 | Deutsch Zentr Luft & Raumfahrt | Method and input device for controlling the position of an object to be graphically represented in a virtual reality |
US6226007B1 (en) | 1999-05-21 | 2001-05-01 | Sun Microsystems, Inc. | Method and apparatus for modeling specular reflection |
US6917692B1 (en) | 1999-05-25 | 2005-07-12 | Thomson Licensing S.A. | Kalman tracking of color objects |
JP3431535B2 (en) | 1999-05-26 | 2003-07-28 | 株式会社ナムコ | GAME SYSTEM AND INFORMATION STORAGE MEDIUM |
US6489955B1 (en) | 1999-06-07 | 2002-12-03 | Intel Corporation | Ray intersection reduction using directionally classified target lists |
US6717579B1 (en) | 1999-06-10 | 2004-04-06 | Dassault Systemes | Reflection line control |
US6504538B1 (en) | 1999-07-01 | 2003-01-07 | Microsoft Corporation | Method and system for generating light values for a set of vertices |
US6421057B1 (en) | 1999-07-15 | 2002-07-16 | Terarecon, Inc. | Configurable volume rendering pipeline |
DE60025725T2 (en) * | 1999-07-15 | 2006-11-09 | Midway Games West Inc., Milpitas | AUTHORIZED SYSTEM AND METHOD WITH IMPROVED SIMULATION OF A VIRTUAL COMPETITOR |
JP2001029649A (en) | 1999-07-21 | 2001-02-06 | Taito Corp | Game machine executing speech visual display by speech recognition |
US6417836B1 (en) | 1999-08-02 | 2002-07-09 | Lucent Technologies Inc. | Computer input device having six degrees of freedom for controlling movement of a three-dimensional object |
JP2001086486A (en) | 1999-09-14 | 2001-03-30 | Matsushita Electric Ind Co Ltd | Monitor camera system and display method for monitor camera |
JP2001079263A (en) | 1999-09-16 | 2001-03-27 | Sega Corp | Game device, data processing method, and information recording medium |
US6554707B1 (en) | 1999-09-24 | 2003-04-29 | Nokia Corporation | Interactive voice, wireless game system using predictive command input |
JP3270928B2 (en) | 1999-09-30 | 2002-04-02 | コナミ株式会社 | Field map generation method, video game system, and recording medium |
WO2001029768A2 (en) | 1999-10-18 | 2001-04-26 | S3 Incorporated | Multi-stage fixed cycle pipe-lined lighting equation evaluator |
US6798411B1 (en) | 1999-10-29 | 2004-09-28 | Intel Corporation | Image processing |
US6571208B1 (en) | 1999-11-29 | 2003-05-27 | Matsushita Electric Industrial Co., Ltd. | Context-dependent acoustic models for medium and large vocabulary speech recognition with eigenvoice training |
US6686924B1 (en) | 2000-02-02 | 2004-02-03 | Ati International, Srl | Method and apparatus for parallel processing of geometric aspects of video graphics data |
US6664955B1 (en) | 2000-03-15 | 2003-12-16 | Sun Microsystems, Inc. | Graphics system configured to interpolate pixel values |
US6426755B1 (en) | 2000-05-16 | 2002-07-30 | Sun Microsystems, Inc. | Graphics system using sample tags for blur |
US6594388B1 (en) | 2000-05-25 | 2003-07-15 | Eastman Kodak Company | Color image reproduction of scenes with preferential color mapping and scene-dependent tone scaling |
US6864895B1 (en) | 2000-05-30 | 2005-03-08 | Hewlett-Packard Development Company, L.P. | Pseudo-linear frame buffer mapping system and method |
US6690372B2 (en) | 2000-05-31 | 2004-02-10 | Nvidia Corporation | System, method and article of manufacture for shadow mapping |
US6717599B1 (en) | 2000-06-29 | 2004-04-06 | Microsoft Corporation | Method, system, and computer program product for implementing derivative operators with graphics hardware |
US6795068B1 (en) | 2000-07-21 | 2004-09-21 | Sony Computer Entertainment Inc. | Prop input device and method for mapping an object from a two-dimensional camera image to a three-dimensional space for controlling action in a game program |
US6825851B1 (en) | 2000-08-23 | 2004-11-30 | Nintendo Co., Ltd. | Method and apparatus for environment-mapped bump-mapping in a graphics system |
US6316974B1 (en) | 2000-08-26 | 2001-11-13 | Rgb Systems, Inc. | Method and apparatus for vertically locking input and output signals |
US6744442B1 (en) | 2000-08-29 | 2004-06-01 | Harris Corporation | Texture mapping system used for creating three-dimensional urban models |
KR20000072753A (en) | 2000-09-25 | 2000-12-05 | 오재덕 | An intelligent toy with Wireless Local Communications. |
GB2367471B (en) | 2000-09-29 | 2002-08-14 | Pixelfusion Ltd | Graphics system |
US6853382B1 (en) | 2000-10-13 | 2005-02-08 | Nvidia Corporation | Controller for a memory system having multiple partitions |
US6765573B2 (en) | 2000-10-26 | 2004-07-20 | Square Enix Co., Ltd. | Surface shading using stored texture map based on bidirectional reflectance distribution function |
JP2002222436A (en) | 2000-11-22 | 2002-08-09 | Sony Computer Entertainment Inc | Object control method, object control processing program executed by computer, computer readable-recording medium with recorded object control processing program executed by computer, and program executing device executing object control processing program |
JP2002159749A (en) | 2000-11-28 | 2002-06-04 | Namco Ltd | Game system and program |
US6850243B1 (en) | 2000-12-07 | 2005-02-01 | Nvidia Corporation | System, method and computer program product for texture address operations based on computations involving other textures |
US6778181B1 (en) | 2000-12-07 | 2004-08-17 | Nvidia Corporation | Graphics processing system having a virtual texturing array |
US6928433B2 (en) | 2001-01-05 | 2005-08-09 | Creative Technology Ltd | Automatic hierarchical categorization of music by metadata |
US6646640B2 (en) | 2001-02-06 | 2003-11-11 | Sony Computer Entertainment Inc. | System and method for creating real-time shadows of complex transparent objects |
US6995788B2 (en) | 2001-10-10 | 2006-02-07 | Sony Computer Entertainment America Inc. | System and method for camera navigation |
US7162314B2 (en) | 2001-03-05 | 2007-01-09 | Microsoft Corporation | Scripting solution for interactive audio generation |
JP3581835B2 (en) | 2001-03-14 | 2004-10-27 | 株式会社イマジカ | Color conversion method and apparatus in chroma key processing |
US6493858B2 (en) | 2001-03-23 | 2002-12-10 | The Board Of Trustees Of The Leland Stanford Jr. University | Method and system for displaying VLSI layout data |
US6741259B2 (en) | 2001-03-30 | 2004-05-25 | Webtv Networks, Inc. | Applying multiple texture maps to objects in three-dimensional imaging processes |
US6961055B2 (en) | 2001-05-09 | 2005-11-01 | Free Radical Design Limited | Methods and apparatus for constructing virtual environments |
US7085722B2 (en) | 2001-05-14 | 2006-08-01 | Sony Computer Entertainment America Inc. | System and method for menu-driven voice control of characters in a game environment |
US6639594B2 (en) | 2001-06-03 | 2003-10-28 | Microsoft Corporation | View-dependent image synthesis |
US7006101B1 (en) | 2001-06-08 | 2006-02-28 | Nvidia Corporation | Graphics API with branching capabilities |
US7162716B2 (en) | 2001-06-08 | 2007-01-09 | Nvidia Corporation | Software emulator for optimizing application-programmable vertex processing |
US6781594B2 (en) | 2001-08-21 | 2004-08-24 | Sony Computer Entertainment America Inc. | Method for computing the intensity of specularly reflected light |
JP2003338972A (en) | 2001-09-12 | 2003-11-28 | Fuji Photo Film Co Ltd | Image processing system, imaging unit, apparatus and method for processing image and program |
US7081893B2 (en) | 2001-10-10 | 2006-07-25 | Sony Computer Entertainment America Inc. | System and method for point pushing to render polygons in environments with changing levels of detail |
AU2002342040A1 (en) | 2001-10-10 | 2003-04-22 | Sony Computer Entertainment America Inc. | Dynamically loaded game software for smooth play. |
US7046245B2 (en) | 2001-10-10 | 2006-05-16 | Sony Computer Entertainment America Inc. | System and method for environment mapping |
JP3493189B2 (en) * | 2001-10-11 | 2004-02-03 | コナミ株式会社 | Game progress control program, game progress control method, and video game apparatus |
US7443401B2 (en) | 2001-10-18 | 2008-10-28 | Microsoft Corporation | Multiple-level graphics processing with animation interval generation |
CN1801949A (en) | 2001-10-29 | 2006-07-12 | 松下电器产业株式会社 | Video/audio synchronizing apparatus |
US7034784B2 (en) | 2001-11-22 | 2006-04-25 | Sharp Kabushiki Kaisha | Optical shifter and optical display system |
JP3732168B2 (en) * | 2001-12-18 | 2006-01-05 | 株式会社ソニー・コンピュータエンタテインメント | Display device, display system and display method for objects in virtual world, and method for setting land price and advertising fee in virtual world where they can be used |
US6657624B2 (en) | 2001-12-21 | 2003-12-02 | Silicon Graphics, Inc. | System, method, and computer program product for real-time shading of computer generated images |
US7212247B2 (en) | 2002-01-31 | 2007-05-01 | Thomson Licensing | Audio/video system providing variable delay |
US7159212B2 (en) | 2002-03-08 | 2007-01-02 | Electronic Arts Inc. | Systems and methods for implementing shader-driven compilation of rendering assets |
US7009605B2 (en) | 2002-03-20 | 2006-03-07 | Nvidia Corporation | System, method and computer program product for generating a shader program |
US6912010B2 (en) | 2002-04-15 | 2005-06-28 | Tektronix, Inc. | Automated lip sync error correction |
US6956871B2 (en) | 2002-04-19 | 2005-10-18 | Thomson Licensing | Apparatus and method for synchronization of audio and video streams |
US6903738B2 (en) | 2002-06-17 | 2005-06-07 | Mitsubishi Electric Research Laboratories, Inc. | Image-based 3D modeling rendering system |
US6831641B2 (en) | 2002-06-17 | 2004-12-14 | Mitsubishi Electric Research Labs, Inc. | Modeling and rendering of surface reflectance fields of 3D objects |
GB0220138D0 (en) | 2002-08-30 | 2002-10-09 | Kaydara Inc | Matte extraction using fragment processors |
US7212248B2 (en) | 2002-09-09 | 2007-05-01 | The Directv Group, Inc. | Method and apparatus for lipsync measurement and correction |
US7339589B2 (en) | 2002-10-24 | 2008-03-04 | Sony Computer Entertainment America Inc. | System and method for video choreography |
US7180529B2 (en) | 2002-12-19 | 2007-02-20 | Eastman Kodak Company | Immersive image viewing system and method |
US7072792B2 (en) * | 2002-12-24 | 2006-07-04 | Daniel Freifeld | Racecourse lap counter and racecourse for radio controlled vehicles |
US7214133B2 (en) * | 2003-05-09 | 2007-05-08 | Microsoft Corporation | Method and apparatus for retrieving recorded races for use in a game |
JP4712712B2 (en) | 2003-05-29 | 2011-06-29 | イーエイティー.ティーブイ、インコーポレイテッド | System for presenting multimedia content |
US7428000B2 (en) | 2003-06-26 | 2008-09-23 | Microsoft Corp. | System and method for distributed meetings |
US7382369B2 (en) | 2003-10-10 | 2008-06-03 | Microsoft Corporation | Systems and methods for robust sampling for real-time relighting of objects in natural lighting environments |
KR100810514B1 (en) | 2003-10-28 | 2008-03-07 | 삼성전자주식회사 | Display apparatus and control method thereof |
GB2411065A (en) | 2004-02-11 | 2005-08-17 | Vodafone Plc | Mobile telephone terminal with camera and using display to illuminate scene. |
US7570267B2 (en) | 2004-05-03 | 2009-08-04 | Microsoft Corporation | Systems and methods for providing an enhanced graphics pipeline |
US7333150B2 (en) | 2004-05-14 | 2008-02-19 | Pixel Instruments Corporation | Method, system, and program product for eliminating error contribution from production switchers with internal DVEs |
KR100703334B1 (en) | 2004-08-20 | 2007-04-03 | 삼성전자주식회사 | Apparatus and method for displaying image in mobile terminal |
US20060046834A1 (en) | 2004-08-30 | 2006-03-02 | Aruze Corp. | Gaming machine, control method for gaming machine, game system, server, mobile device and game program |
WO2006033360A1 (en) | 2004-09-21 | 2006-03-30 | Nikon Corporation | Mobile information device |
US20060071933A1 (en) | 2004-10-06 | 2006-04-06 | Sony Computer Entertainment Inc. | Application binary interface for multi-pass shaders |
KR100631603B1 (en) | 2004-10-25 | 2006-10-09 | 엘지전자 주식회사 | How to improve image quality of mobile communication terminal |
US20060209210A1 (en) | 2005-03-18 | 2006-09-21 | Ati Technologies Inc. | Automatic audio and video synchronization |
US7636126B2 (en) | 2005-06-22 | 2009-12-22 | Sony Computer Entertainment Inc. | Delay matching in audio/video systems |
US7589723B2 (en) | 2005-07-25 | 2009-09-15 | Microsoft Corporation | Real-time rendering of partially translucent objects |
US9697230B2 (en) | 2005-11-09 | 2017-07-04 | Cxense Asa | Methods and apparatus for dynamic presentation of advertising, factual, and informational content using enhanced metadata in search-driven media applications |
JP2007271999A (en) | 2006-03-31 | 2007-10-18 | Renesas Technology Corp | Photomask, method for manufacturing the same, and method for manufacturing electronic device |
US7965859B2 (en) | 2006-05-04 | 2011-06-21 | Sony Computer Entertainment Inc. | Lighting control of a user environment via a display device |
US7880746B2 (en) | 2006-05-04 | 2011-02-01 | Sony Computer Entertainment Inc. | Bandwidth management through lighting control of a user environment via a display device |
JP2007334664A (en) | 2006-06-15 | 2007-12-27 | Sharp Corp | Process controller, process control method, and computer program |
US7907117B2 (en) | 2006-08-08 | 2011-03-15 | Microsoft Corporation | Virtual controller for visual displays |
US7896733B2 (en) | 2006-09-14 | 2011-03-01 | Nintendo Co., Ltd. | Method and apparatus for providing interesting and exciting video game play using a stability/energy meter |
US20080111789A1 (en) | 2006-11-09 | 2008-05-15 | Intelligence Frontier Media Laboratory Ltd | Control device with hybrid sensing system comprised of video-based pattern recognition and electronic signal transmission |
KR100849693B1 (en) | 2006-12-28 | 2008-07-31 | 엔에이치엔(주) | Method and system for generating static-shadow |
JP5285234B2 (en) | 2007-04-24 | 2013-09-11 | 任天堂株式会社 | Game system, information processing system |
US8566158B2 (en) | 2008-01-28 | 2013-10-22 | At&T Intellectual Property I, Lp | System and method for harvesting advertising data for dynamic placement into end user data streams |
-
2003
- 2003-10-22 US US10/691,929 patent/US8133115B2/en not_active Expired - Fee Related
- 2003-12-15 US US10/737,143 patent/US20050090302A1/en not_active Abandoned
-
2004
- 2004-10-14 EP EP04256331A patent/EP1541207A1/en not_active Withdrawn
- 2004-10-19 JP JP2004304868A patent/JP3821822B2/en not_active Expired - Fee Related
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4068847A (en) * | 1976-06-23 | 1978-01-17 | The Magnavox Company | Chroma and luminance signal generator for video games |
US4116444A (en) * | 1976-07-16 | 1978-09-26 | Atari, Inc. | Method for generating a plurality of moving objects on a video display screen |
US4905147A (en) * | 1986-10-15 | 1990-02-27 | Logg George E | Collision detection system for video system |
US5265888A (en) * | 1990-06-22 | 1993-11-30 | Nintendo Co., Ltd. | Game apparatus and memory cartridge used therefor |
US5269687A (en) * | 1990-08-01 | 1993-12-14 | Atari Games Corporation | System and method for recursive driver training |
US5354202A (en) * | 1990-08-01 | 1994-10-11 | Atari Games Corporation | System and method for driver training with multiple driver competition |
US5577913A (en) * | 1990-08-01 | 1996-11-26 | Atari Games Corporation | System and method for driver training with multiple driver competition |
US6233291B1 (en) * | 1995-05-01 | 2001-05-15 | Metalink Ltd. | Method and apparatus for performing data sorting in a decoder |
US6220962B1 (en) * | 1997-11-20 | 2001-04-24 | Nintendo Co., Ltd | Video game apparatus having a direction pointing mark and player object displayed on a game screen |
US6141041A (en) * | 1998-06-22 | 2000-10-31 | Lucent Technologies Inc. | Method and apparatus for determination and visualization of player field coverage in a sporting event |
US6273814B1 (en) * | 1999-10-25 | 2001-08-14 | Square Co., Ltd. | Game apparatus and method for controlling timing for executive action by game character |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080094390A1 (en) * | 2004-09-03 | 2008-04-24 | Konami Digital Entertainment Co., Ltd. | Video Generation Device, Load Display Method, Recording Medium, and Program |
US7843453B2 (en) | 2004-09-03 | 2010-11-30 | Konami Digital Entertainment Co., Ltd. | Video generation device, load display method, recording medium, and program |
US20090118003A1 (en) * | 2004-09-09 | 2009-05-07 | Konami Digital Entertainment Co., Ltd. | Image Creating Device, Load Display Method, Recording Medium, and Program |
US20100035689A1 (en) * | 2008-08-11 | 2010-02-11 | Yaniv Altshuler | System and method for a gaming community |
US10786736B2 (en) | 2010-05-11 | 2020-09-29 | Sony Interactive Entertainment LLC | Placement of user information in a game space |
US11478706B2 (en) | 2010-05-11 | 2022-10-25 | Sony Interactive Entertainment LLC | Placement of user information in a game space |
US9342817B2 (en) | 2011-07-07 | 2016-05-17 | Sony Interactive Entertainment LLC | Auto-creating groups for sharing photos |
US10991140B2 (en) | 2014-01-03 | 2021-04-27 | Creative Technology Ltd | System suitable for one or both of audio processing and graphics processing and a method of processing in association therewith |
US11410358B2 (en) | 2014-01-03 | 2022-08-09 | Creative Technology Ltd | System suitable for efficient communication of media stream and a method in association therewith |
Also Published As
Publication number | Publication date |
---|---|
JP3821822B2 (en) | 2006-09-13 |
EP1541207A1 (en) | 2005-06-15 |
US8133115B2 (en) | 2012-03-13 |
JP2005125095A (en) | 2005-05-19 |
US20050090312A1 (en) | 2005-04-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20050090302A1 (en) | System and method for utilizing vectors in a video game | |
JP3972230B2 (en) | GAME DEVICE, GAME DEVICE CONTROL METHOD, AND RECORDING MEDIUM | |
US6171186B1 (en) | Game processing method, game device, image processing device, image processing method, and recording medium | |
US6409597B1 (en) | Video game machine, screen display method for video game, and recording medium containing screen display program | |
US8523646B2 (en) | Program for racing game device, storage medium storing the program, and racing game device | |
US6494784B1 (en) | Driving game machine and a storage medium for storing a driving game program | |
US6733383B2 (en) | Systems and methods for simulating game state changes responsive to an interrupt condition | |
JP4035867B2 (en) | Image processing apparatus, image processing method, and medium | |
US20100087248A1 (en) | Method, storage medium, and apparatus for generating image of virtual space | |
JP3463065B2 (en) | Driver training system | |
JPWO2009028690A1 (en) | GAME DEVICE, GAME PROGRAM, AND GAME DEVICE CONTROL METHOD | |
EP1541208A1 (en) | System and method for utilizing vectors in a video game | |
JP3335445B2 (en) | 3D simulator | |
JPH0970483A (en) | Game device | |
JP4579889B2 (en) | Game machine, game system, game program | |
JP4318834B2 (en) | GAME DEVICE AND INFORMATION STORAGE MEDIUM | |
JP3165768B2 (en) | Video game equipment | |
JP3995817B2 (en) | Game machine and information storage medium | |
JP3827697B2 (en) | GAME DEVICE, ITS CONTROL METHOD, AND PROGRAM | |
JPH08332280A (en) | Game device | |
Erdelyi | Using Computer Vision Techniques to Play an Existing Video Game | |
JP3827696B2 (en) | GAME DEVICE, ITS CONTROL METHOD, AND PROGRAM | |
Anderson | Innovative navigation artificial intelligence for motor racing games | |
JPH08318048A (en) | Three-dimensional game device and picture composition method | |
JP2005237822A (en) | Game device and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY COMPUTER ENTERTAINMENT AMERICA INC., CALIFORN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CAMPBELL, SCOTT;REEL/FRAME:014809/0940 Effective date: 20031208 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: SONY INTERACTIVE ENTERTAINMENT AMERICA LLC, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:SONY COMPUTER ENTERTAINMENT AMERICA LLC;REEL/FRAME:038626/0637 Effective date: 20160331 Owner name: SONY INTERACTIVE ENTERTAINMENT AMERICA LLC, CALIFO Free format text: CHANGE OF NAME;ASSIGNOR:SONY COMPUTER ENTERTAINMENT AMERICA LLC;REEL/FRAME:038626/0637 Effective date: 20160331 |