WO1997005926A1 - Interactive computer controlled exercise system - Google Patents
Interactive computer controlled exercise system Download PDFInfo
- Publication number
- WO1997005926A1 WO1997005926A1 PCT/US1996/012369 US9612369W WO9705926A1 WO 1997005926 A1 WO1997005926 A1 WO 1997005926A1 US 9612369 W US9612369 W US 9612369W WO 9705926 A1 WO9705926 A1 WO 9705926A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- ofthe
- computer
- exercise
- virtual
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/016—Input arrangements with force or tactile feedback as computer generated output to the user
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B21/00—Exercising apparatus for developing or strengthening the muscles or joints of the body by working against a counterforce, with or without measuring devices
- A63B21/15—Arrangements for force transmissions
- A63B21/151—Using flexible elements for reciprocating movements, e.g. ropes or chains
- A63B21/154—Using flexible elements for reciprocating movements, e.g. ropes or chains using special pulley-assemblies
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B24/00—Electric or electronic controls for exercising apparatus of preceding groups; Controlling or monitoring of exercises, sportive games, training or athletic performances
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B71/0619—Displays, user interfaces and indicating devices, specially adapted for sport equipment, e.g. display mounted on treadmills
- A63B71/0622—Visual, audio or audio-visual systems for entertaining, instructing or motivating the user
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B71/0619—Displays, user interfaces and indicating devices, specially adapted for sport equipment, e.g. display mounted on treadmills
- A63B71/0622—Visual, audio or audio-visual systems for entertaining, instructing or motivating the user
- A63B2071/0638—Displaying moving images of recorded environment, e.g. virtual environment
- A63B2071/0641—Displaying moving images of recorded environment, e.g. virtual environment with a marker advancing in function of the exercise
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B71/0619—Displays, user interfaces and indicating devices, specially adapted for sport equipment, e.g. display mounted on treadmills
- A63B71/0622—Visual, audio or audio-visual systems for entertaining, instructing or motivating the user
- A63B2071/0638—Displaying moving images of recorded environment, e.g. virtual environment
- A63B2071/0644—Displaying moving images of recorded environment, e.g. virtual environment with display speed of moving landscape controlled by the user's performance
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B71/00—Games or sports accessories not covered in groups A63B1/00 - A63B69/00
- A63B71/06—Indicating or scoring devices for games or players, or for other sports activities
- A63B71/0619—Displays, user interfaces and indicating devices, specially adapted for sport equipment, e.g. display mounted on treadmills
- A63B2071/0658—Position or arrangement of display
- A63B2071/0661—Position or arrangement of display arranged on the user
- A63B2071/0666—Position or arrangement of display arranged on the user worn on the head or face, e.g. combined with goggles or glasses
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B2220/00—Measuring of physical parameters relating to sporting activity
- A63B2220/30—Speed
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S482/00—Exercise devices
- Y10S482/901—Exercise devices having computer circuitry
- Y10S482/902—Employing specific graphic or video display
Definitions
- This invention relates to computer controlled exercise machines and provides the user with a wide variety of interactive exercise options controlled by software.
- a virtual reality environment or a hybrid of virtual and real environments is provided which permits the user to perform significant physical exertion by applying forces to the machine while viewing images in such a fashion that the user perceives that he is exerting forces against the objects the images represent.
- the system may incorporate advanced image presentation methods, such as computer graphics, holography, or 3-D motion picture projection, and may include interactive entertainment, competitive, or motivational software, together with sound, vibration, and other special effects to enhance the sense of realism experienced by the user.
- controls incorporating force sensing and position detection may be used.
- Virtual reality systems applied to simulations such as preparation for space missions or strictly for entertainment also use force and position detectors.
- these inventions differ fundamentally from exercise machines in that they are not specifically adapted to apply the type of load to the user which results in the most beneficial conditioning effect.
- exertion to the point of fatigue, as well as to the limits of strength are important to the achievement of health benefits by the user.
- Exercise machines are designed to maximize the safe strain on the user.
- interactive robotics and machine controls are designed to minimize strain on the operator and to reduce fatigue thereby increasing worker output and accuracy.
- Computer controlled exercise machines which utilize a virtual reality or hybrid environment must be robust, must be capable of applying loads up to the usual limits of human strength, and must be suited to many repetitive cycles
- Delicate sensors susceptible to flex fatigue failure are undesirable Components must be adapted to withstand exposure to sweat without corrosion or electrical failure Skin abrasion and blister formation should be avoided
- Typical virtual reality components such as sensor-equipped gloves and body suits, and heavy visual display helmets are not well adapted to exercise applications
- U S Patent No 5, 184,319 entitled FORCE FEEDBACK AND TEXTURES SIMULATING INTERFACE DEVICE discloses a force-actuating device which transmits force to a pressure-sensing body part and also includes sensors to measure the actual force applied to the body part as well as additional sensors to measure joint angles These components may be incorporated into a "feedback glove" permitting texture information and other simulation of grasping in a virtual reality system
- the force transmitters and sensors are intended for relatively low loads, such as at finger joints, and no applications for exercise are anticipated
- the invention is principally concerned with transfer of information at the man-machine interface rather than the transfer of significant conditioning forces at the interface as are required for effective exercise
- U S Patent No 5,227,985 entitled COMPUTER VISION SYSTEM FOR POSITION MONITORING IN THREE DIMENSIONS USING NON-COPLANER LIGHT SOURCES ATTACHED TO A MONITORED OBJECT relates to " monitoring the position and orientation of an object while it is displaced by an operator, who is enabled to interactively control various devices such as teleoperated mechanisms or graphics displays presented to his eyes while his motions are sensed and computed "
- the invention finds application in “communication methods with the virtual world", and although it could be incorporated as a component of a computer controlled exercise machine the patent makes no mention of and does not anticipate this use
- Exercise machines ofthe present invention range from relatively simple means of creating the illusion in the mind of the user that he is exercising in a computer-presented virtual reality environment to more complex highly interactive systems in which the user can control the "place” in the virtual world that he chooses to "go” and the activities that he chooses to do in that "place".
- the former includes a virtual reality exercise bicycle in which the user controls the speed and direction of his trip on a virtual road via the handle bars, pedals and brakes, or a treadmill on which the user runs in a virtual race.
- virtual reality as used in the present technological vernacular is not succinctly defined. It generally refers to computer systems which create the illusion that the user is actually present in a tangible physical environment and can interact with objects which appear to be present when in actuality the images viewed are of objects which are not physically present and are displayed using information present in the computer. As used in this patent document, the term “virtual reality” is used to refer to any illusionary environment which visually creates the false impression that the individual is actually present in a three-dimensional space and can interact with objects appearing to be present.
- the image presented may be derived from computer graphics, photography, or art work, so long as the false illusion is created that the individual is interactively present within the space represented by the image.
- hybrid virtual reality refers to the inclusion of real objects in addition to images representing objects that are not physically present, both within view ofthe individual experiencing the "hybrid virtual reality".
- a system where the user both sees his real hand and sees the image of an object not actually present that appears to be held in his hand is an example of "hybrid virtual reality”.
- Interactive "virtual reality” or “hybrid virtual reality” exercise machines of the present invention provide a visual image of an illusory environment with which the user interacts by applying force to the machine structure.
- the position of "virtual" objects presented in the visual field by a computer via a display such as a head-mounted video screen, the orientation ofthe user with respect to the "virtual” objects, and the forces applied between the user and the machine structure at various locations, are all monitored by the computer.
- the parts ofthe machine structure are physically moved, creating a changing arrangement of machine elements during use.
- the forces required of the user to cause these changes are controlled by the computer in such a way that the user perceives that he is interacting with the
- the computer may utilize a CD-ROM device or other high capacity memory device with advanced graphics, and may also incorporate 3-D images.
- One preferred embodiment ofthe invention is intended for installation in health clubs and gymnasiums. It includes a booth-like enclosure which contains an exercise mechanism and provides an externally darkened setting permitting images projected on a partially transparent visor to be clearly viewed.
- the user wears a lightweight exercise suit which has a fluorescent surface illuminated by ultraviolet light.
- the inside ofthe booth is a black-light-absorbent surface and, other than his own body, only the "handle grip" portions ofthe machine structure are visible to the user through the visor.
- the software, sensors, and computational algorithms are such that the computer matches the overlay of the projected image with the real positions of the exercise machine "handles" such that the handles appear to be part ofthe picture.
- the device is a hybrid between actual reality and virtual reality.
- the machine provides a particular force of resistance and path of motion ofthe handle, which is kept in correspondence to a changing computer-generated image of the scene with the door swinging open on its hinges
- the user perceives that he has opened a virtual door, when actually he has only pulled on a portion ofthe machine's mechanism
- the exercise machine may have many interchangeable "handles" of different size, shape, hardness, and texture, that are held in a storage area and may be removably affixed to the machine utilizing a mechanism similar to a tool changer such as used with a computer controlled milling machine This is accomplished without the attention ofthe user, and permits an appropriate surface to be contacted by the user to exert the necessary forces for the exercise while reasonably well matching the surface represented in the image
- a tool changer such as used with a computer controlled milling machine
- an exercise regimen might be composed of a trip through a virtual obstacle course, where the user has the sensation of climbing over objects, pushing objects aside, pulling his body up a pole, and the like, and the pace through the course, as well as the path chosen, may be determined by the user
- the major objective of the present invention is to provide a computer controlled exercise machine which creates the illusion that the user is present in a visually-represented three- dimensional space containing objects with which the user can interact with sufficient exertion to enhance the user's health
- Figure 1 is a perspective view of an embodiment ofthe invention, including a booth-like enclosure, exercise apparatus with ball screw actuators, sensors, computer, treadmill, and an image display helmet.
- Figure 2 is a diagrammatic drawing of a user wearing a head-mounted display helmet and gripping a handle of an exercise apparatus.
- Figure 3 is a perspective view of a head-mounted display and an array of sensors utilized to detect its location and orientation in space
- Figure 4 is a drawing of one sensor assembly utilized to detect head-mounted display (helmet) position
- Figure 5 is a perspective view of a 3-axis load-sensing device utilized to detect forces applied between the user and the machine structure via a handle. The position of a cover is indicated by the broken line
- Figure 6 is a perspective view of an additional embodiment of the invention including a booth-like enclosure, exercise apparatus with boom and actuators, sensors for both handle position measurement and leg position measurement, and the computer.
- Figure 7 is a diagrammatic top view of an array of two video cameras oriented in the north/south direction
- Figure 8 is a rectilinear map of the combined view area of both ofthe two video cameras illustrated in Figure 7, with the outlined presentation field ofthe video display screen shown in two different positions relative to the border ofthe overall camera view area.
- Figure 9 is a block diagram of the preferred embodiment ofthe present invention.
- Figure 10 is the computer program ofthe preferred embodiment ofthe present invention for using to superimpose a computer generated image ofa scene containing virtual objects over the position ofthe parts ofthe exercise machine over against which the user applies force, such as the handles.
- the program controls the actuators of the machine to apply forces and motions to the handles which correspond with the image and would be appropriate to move a virtual object ofthe mass and inertia which the image represents
- Figure 1 illustrates a typical embodiment ofthe invention in which a hybrid between "virtual reality” and genuine reality is used
- the user sees both a computer- displayed image of a virtual environment, and some especially visible objects within the real environment, including the user's arms and hands, as well as the handles ofthe apparatus.
- the user may see only a computer display and may not directly view any part of his body or his real environment
- the embodiment illustrated in Figure 1 is also simplified in that only two handles are provided, and the user either stands or runs on a treadmill or sits on a seat More complex embodiments which also include active interfaces for the feet, such as boots attached to servo-controlled mechanisms may be provided. Thus, pedaling, climbing, and other actions can be simulated.
- the user 2 stands on a treadmill 4 within a generally darkened booth 6.
- the user wears a display helmet 8 having a partially transparent projection screen 10 before his eyes, and a television projection system 12 mounted to the helmet.
- Audio speakers may also be provided in the helmet as well as a microphone to pick up the user's voice in embodiments where voice commands are recognized by the computer.
- various motors and sensors are illustrated in Figure 1 which are connected to the computer via cables which are not included in the drawing. Details ofthe computer controls are also omitted for simplicity.
- the television projection system utilizes fiber optics to transfer light beams from the main unit located remotely together with the computer 13, to the helmet 8 via an optic cable 14. Separate images may be projected to each eye to provide stereoscopic 3-D images.
- the partially transparent projection screen may be a half mirrored sheet of glass or other transparent material. Special coatings may be used to enhance the clarity ofthe image.
- the user wears an exercise suit similar to a body stocking with gloves, which has a fluorescent surface highly visible under a particular wavelength oflight, such as ultraviolet, shined on the user by a lamp 16.
- the handles of the exercise machine are also visible under this wavelength, but the remainder of the apparatus, including the structural frame members, servo ⁇ motors, force sensors, and the walls ofthe enclosure are all light- absorbent and essentially invisible to the user.
- the computer system which monitors the exact position of the helmet and the positions ofthe machine handles.
- the apparatus includes multi-axis, servo-controlled, actuators for each ofthe two handles, which are capable of positioning the handles to accommodate a wide range of motion for the exercising user, and are capable of applying high loads to the handles and measuring the applied forces.
- the major components ofthe actuator for the left handle 18 include a frame 20, slide 22, ball screw 24, ball nut 26, and servo-motor with an encoder to measure shaft rotational position 28, for the x-axis; a frame 30, slide 32, ball screw 34, ball nut 36, and servo motor with an encoder to measure shaft rotational position 38, for the y-axis; a frame 40, slide 42, ball screw 44, ball nut 46, and servo motor with an encoder to measure shaft rotational position 48, for the z-axis; and high torque servo-motors with encoders and gear heads, 50 and 52, for two additional axes of rotational motion ofthe handle.
- a multi-axis force transducer 54 (see Figure 5) provides the computer 13 with a measurement ofthe magnitude and direction of force applied by the user to the handle (which also represents the force applied by the machine to the user).
- three strain gages 53 are mounted on a slotted block such that x-, y-, and z- axis forces are sensed.
- a similar frame with servo-motors and sensors generally indicated at 58 is also provided for the right handle 60.
- the user may walk or run on the treadmill 4, or may sit on a seat, 64 which can be swiveled into or out of position by servo-motor 66.
- the partially transparent screen 10 receives an image projected onto it from the fiber optic cable 14 through a lens 68, via two mirrors 70 and 72.
- Light-ray path 74 from the screen to the eye coincides with the final portion of light ray path 76 from the hand to the eye.
- the light travels a distance between the lens and the screen of approximately 18 inches which is a typical distance from an average adult's eye to his hand when the elbow is partially bent.
- Such a position, represented in Figure 2 is a very common one when utilizing the arms for relatively heavy lifting. Thus, during use ofthe exercise machine, the user's hands will be relatively close to this distance from his eye much ofthe time.
- the handle ofthe machine When the user looks at his own hand, his eye will focus on it.
- the handle ofthe machine will be at this distance, ifthe user is holding it, and so it will also be in focus. Since the part of the image projected on the screen which represents the part of the virtual environment that the user is touching with his hand is superimposed at approximately the same focal distance, the image can also be presented in good focus.
- the lens in front ofthe fiber optic cable can also be moved by a servo-motor to adjust the focus to the real distance between the handle and the screen, because this distance is calculable by the computer from helmet position and handle position information.
- Figure 3 illustrates the system utilized to provide the control information to the computer to permit calculation ofthe exact helmet position, and therefore to permit proper superimposition ofthe projected image with the real image viewed through the partially transparent screen.
- Three arrays of three transducers are utilized as illustrated generally at I, II, and III, in Figure 3.
- Each of these transducers is comprised of a cable wound around a spring-retracted take-up spool which is in turn mounted to a multiple turn potentiometer or to a digital encoder, as shown in Figure 4.
- Figures 1, 3, and 6 only the take-up spools are shown, for clarity.
- the cable 78 passes through an eyelet 80 and is wrapped around a spool 82 loaded by a band spring 84.
- the length between the end of the cable (off the spool) and the eyelet is a function ofthe rotational position ofthe spool and the number of wraps of cable on the spool.
- the potentiometer 86 attached to the spool thus gives an analog signal proportional to the length of cable beyond the eyelet.
- Use of an A to D converter or suitable digital encoder in place ofthe potentiometer provides an appropriate input signal to the computer.
- the three transducers in array I, 88, 90, and 92 are rigidly attached to the ceiling ofthe exercise booth as shown in Figure 1. Their three eyelets, designated as points A, B, and C, thus form a triangle of known dimensions and orientation with relation to the frame ofthe exercise apparatus.
- the three cables are attached to a common point on the back of the helmet designated as point D. Since distances ab, be and ac are fixed and are known, and since the transducer output signals permit calculation of lengths ad, bd, and cd, a tetrahedron with vertices ABCD is completely defined with the lengths of all sides known and the angles of one side (abc) known.
- the type of mechanism used to apply the forces of exercise and the type of sensors used to detect the forces from moment to moment at the interfaces between the user and the machine may also be varied.
- Figure 6 illustrates an embodiment ofthe invention in which mechanisms and actuators to apply forces to the handles are configured to present a minimal amount of hardware within the user's visual field, and thus be effectively hidden from the user's attention.
- a hollow boom 94 supports the handle 96 mounted on rod 98 passing therethrough via a force-sensing array 54 which may be similar to that illustrated in Figure 5
- the boom passes through a rectangular slot 100 in a wall of the exercise booth which restricts rotational motion ofthe boom around its long axis
- the rod, passing through the hollow core ofthe boom is mounted on rotary bearings at each end and is connected to the shaft ofa servo-motor 99 mounted on the opposite end ofthe boom
- the boom is supported by six actuator cables 101, 102, 103, 104, 105, and 106 which pass through wear-resistant eyelets mounted to the wall ofthe booth Three ofthe actuator cables are each fixed near one end ofthe boom, the remaining three are fixed near the other end ofthe boom, and each cable
- the six actuator cables are led via the eyelets to each of six actuator motor spools 131, 132, 133, 134, 135, and 136
- the rotary position of each spool is detected by a separate sensor such as a motor shaft encoder which provides the computer with control signals proportional to the length of each cable which are maintained taut
- the motors and encoders for each cable attached to each spool are similar to those shown for cable 105 (motor 114, encoder 116) but most ofthese have not been included in the drawing for clarity
- the physical arrangement ofthe boom, cables and eyelets defines two tetrahedra having a common side and vertices opposite the common side which lie near the ends ofthe boom Given that the lengths ofthe sides ofthe common bases ofthe two tetrahedra are known and all the lengths ofthe other sides are determined (by the control signals) the position ofthe boom in three-dimensional space is calculated by the computer Via the servo ⁇ motors and this frame, cable, and boom mechanism, and information from the force
- FIG. 9 illustrates the basic components ofthe computer and its interface with the sensors and the actuators ofthe exercise machine
- the central processing unit (CPU) 51 conducts the main computing functions ofthe computer and is connected both to memory 53 which stores data about both the virtual environment with virtual objects and about the real environment including the physical positions, motions and forces applied to and by the exercise machine structure and its actuators
- the CPU receives inputs from various sensors including handle position sensors 55, handle force sensors 57, and helmet position sensors 59 which offer appropriate signal conditioning by signal conditioners 61, 63, and 65, which may include analog-to-digital converters, amplifiers, or certain preprocessors Based on the data from the sensors, by making appropriate calculations according to the system's programs, the CPU sends the necessary output signals to the video controller 69 which in turn drives the video display 71 to present the image to the user
- the basic computer program which is used superimposes an image of a scene containing virtual objects ofthe position of the parts of the exercise machine against which the user applies force, such as the handles
- the program controls the actuators of the machine to apply forces and motions to the handles which correspond with the image and would be appropriate to move a virtual object of the mass and inertia which the object represents
- the program utilizes data from the sensors described above which permits computation ofthe position ofthe handles and the forces applied thereby and thereto Data from the sensors ofthe position of the helmet in relation to the physical structure ofthe exercise machine is obtained, which is used to compute the vantage point of the user (corresponding to the physical position of the display screen which is mounted in the helmet)
- the program generates the necessary control and display signals to maintain superimposition ofthe computer image and the handle, and to match the force applied by the exercise machines actuators to the appropriate force to move the virtual object
- Many variations of computer program organization and function may be employed, depending upon the exact hardware used with each embodiment of the invention or particular virtual reality environment and objects which the software writer
- the embodiment shown in Figure 6 includes two position- sensing arrays of cables, spools, and potentiometers, which are provided to sense the positions of the user's feet. These arrays, generally indicated at 140, and 142, are attached to fasteners 144, 146, on the lateral sides ofthe user's shoes. As the user walks or runs on the treadmill, the length of his stride and frequency is calculated, permitting the computer to monitor his speed, compute the distance he travels, and implement other functions, such as automatically adjusting the speed ofthe treadmill to the speed ofthe user, so that he does not need to pay attention to the treadmill and has a greater sense that he is walking on the surface ofa virtual environment as presented by the headset visual display.
- the virtual environment viewed by the user may be a composite of photographic images with superimposed computer-generated images, and additionally may incorporate superimposed images of the user's body and parts of the exercise machine which are directly viewed.
- the embodiment of Figure 6 utilizes a video camera 148, mounted to the headset so as to receive an image ofthe field of view in front ofthe user via light reaching the lens as indicated by arrows 149.
- the user wears fluorescent clothing preferentially illuminated compared to the background and mechanism via ultraviolet light form lamp 16.
- other wavelengths and special filters may be used with the camera, 148.
- the image ofthe user's arms, hands, body, legs, and feet thus obtained is superimposed by the computer, over other images on the video display screen, 150, shown here inside a helmet, to create the complete virtual environment scene. In this way, no sensors are required on the user to provide position information for an image ofthe user's body on the screen.
- the video image ofthe user moves in perfect synchrony with him, which helps to provide a very realistic virtual reality illusion with minimal computational requirements compared to other methods, such as animated real-time images representing the user's body presented on the screen.
- a device to provide appropriate photographic images of remote locations which may be used in combination with computer graphics, includes video cameras 152, and 154 which are equipped with sensors to monitor the position while the video image is recorded and to make a simultaneous recording of this position information to orient the image on the screen when it is replayed.
- the camera may be equipped with a compass 156 and a level detector 158 which record the orientation ofthe camera's field of view in reference to magnetic north and to the level ofthe horizon.
- the camera may alternatively be equipped with a gyroscopic position-sensing system or other system such as a satellite locator receiver to record its position in space.
- the video image recorded is a wide angle image, preferentially as close to the full 360 degree field of view as is practical with advanced optics, including multiple cameras if necessary.
- the recorded field of view ofthe scene is thus much wider than the field of view presented on the screen during playback.
- the field of view ofthe playback display is no wider than the normal human field of view and may be considerably narrower.
- the cameras 152 and 154 are equipped with wide angle lenses 160 and 162 which each provide a field of view 150 degrees wide by 150 degrees high.
- the two combined images thus represent a field 300 degrees by 150 degrees. This is represented by the large rectangle 164 of Figure 8 which is a flat map ofthe field.
- the playback and viewing device includes a head-mounted display and computer system to match the portion ofthe recorded image displayed to the actual orientation in space ofthe head- mounted display screen.
- the correspondence ofthe displayed image to the orientation ofthe head mounted screen provides the user with the view that he would have had if he had been at the actual location where the camera was situated when the video was recorded and moved his head to look in different directions.
- the horizon is represented by line 166 in Figure 8, and with the camera array directed north as illustrated in Figure 7, the position ofthe north direction is indicated by line 168 on the map ofthe camera view field in Figure 8.
- the position information encoded by the compass and level detector permit the computer to locate reference points on the camera field map with regard to the horizontal/vertical and north/south directions.
- the small solid line rectangle 170, in Figure 8 corresponds to a viewing screen angular field of 160 degrees by 100 degrees, somewhat less than the normal human combined right-eye/left-eye total angular field of about 180 degrees by 115 degrees.
- the computer system utilizes position-sensor information to ascertain the actual physical orientation ofthe screen in the helmet, and when the helmet is level and facing due north, the computer will present an image recorded by a camera in the same position (level and facing north) as indicated by rectangle 170. If the user then turns his head upward and to the left the computer will present an image on the screen corresponding to the portion ofthe map represented by dotted rectangle 172.
- the computer also is provided with software permitting the orientation of image to be adjusted to the north/south alignment of the booth within which the machine is contained such that absolute direction is unimportant during use. In systems where the replayed camera-recorded video is only part of the overall image perceived by the viewer, a computer- generated image of an object may be superimposed on the recorded scene.
- the position ofthe computer-generated object image in the visual field as well as the perimeter ofthe image are contained in computer memory for each frame ofa motion picture output by the computer. This position and perimeter information is utilized to turn off the portion of the recorded video that would have been displayed within the perimeter of the computer- generated object and thus avoid projecting two different overlaying images onto the viewing screen simultaneously.
- the result is a camera- recorded background with computer-generated objects filling in voids of matching size, position, and shape, deleted from the camera recorded image.
- the "foreground" objects may originate from camera-recorded images and the background may be computer-generated, or a combination may be used, such as in embodiments where a video camera receives the image ofthe user's body.
- Examples of "virtual reality" exercise machines using camera-recorded images and camera- recorded images combined with computer-generated images include bicycle ride and bicycle race configurations.
- a very wide angle video camera with orientation sensors and recorder is mounted on a vehicle and driven along a road at a constant speed with no other traffic while the camera is photographing the scene. Ifthe speed is a typical bicycling speed and the video is recorded at a somewhat high number of frames per second, such as 30, then by varying the playback speed it will appear that the bicycle is traveling at a rate of speed which can be either more or less than the actual speed ofthe vehicle when the recording was made.
- the rate at which the pedals ofa stationary exercise machine are rotated represents the rate at which a real bicycle would travel while pedaling.
- the computer is programmed to vary the playback rate ofthe video recording as a function ofthe pedal speed such that the user perceives motion at the speed he is pedaling. Algorithms to represent coasting may be incorporated. With the head-mounted video display the user can turn his head and see the appropriate scene as if he were looking in the particular direction his head is oriented while riding a bicycle on a real road. This creates a very realistic interactive "virtual reality" effect.
- the simulated trip made along the road would not require simulated steering. For example, if the road ran due north, and the person on the exercise machine held his head straight (i.e. not turned to either side) as the video recording was played back, he would face due north along the center of the road during the entire trip. However, if the road along which the video recording was made was curved, say, 90 degrees and headed east, as the person exercising viewed the portion of the tape corresponding to the curve, he would have to turn his head 90 degrees to the right retain a view along the center ofthe road. The apparent forward direction of his bicycle would remain north, and it would seem that he should travel off the side of the road.
- a simulated steering mechanism and control system is provided.
- the position of the exercise bicycle handle bars are measured by a sensor that provides a control signal to the computer.
- the north/south orientation ofthe image displayed is modified according to this control signal to permit the exerciser to follow the road and maintain the centered alignment of his body and the exercise bicycle with an apparent course down the center ofthe road.
- the computer may calculate the radius of curvature that the path ofa real bicycle would make at a given speed and adjust the rate of modification ofthe visual image to correspond to this.
- a bicycle race exercise machine the same video recording and system as described in the preceding paragraphs can be enhanced by providing a computer-generated image ofa competing bicycle rider superimposed on the roadway and background scenery.
- the user will thus see the competitive image either ahead or behind by turning his head around (provided a wide enough field of view is photographed).
- the user may "compete" against a computer-simulated competitor, or two exercise machines with two competing individuals may be interconnected, such that the position ofthe image ofthe competitive rider that each sees corresponds to the appropriate position represented by the complete history of pedal rotation ofthe real competitor since the beginning of the race. This creates a "real" race in a virtual environment. Innumerable variations of this are possible utilizing machines which simulate many kinds of physical activity in many settings.
Abstract
This invention relates to computer (13) controlled exercise machines, and provides the user (2) with a wide variety of interactive exercise options controlled by the software. A 'virtual reality' environment or a hybrid of virtual and real environment is provided which permits the user to perform significant physical exertion by applying forces to the machine while viewing images on a head mounted display (12, 150). The invention includes means for the user to view his own hands and body superimposed over a computer generated image of objects that are not actually present while maintaining parts of the exercise machine that the user physically contacts, such as a handle (18, 60, 96), superimposed over the computer generated image. As the user exerts forces against the machine (such as handle) he perceives that he is exerting forces against the objects the images represent. The invention includes video camera (148) and computer means, adapted to record images from the real world, which may be combined with computer generated images while retaining proper spatial orientation to produce a composite 'virtual reality' environment.
Description
INTERACTIVE COMPUTER CONTROLLED EXERCISE SYSTEM FIELD OF THE INVENTION
This invention relates to computer controlled exercise machines and provides the user with a wide variety of interactive exercise options controlled by software. A virtual reality environment or a hybrid of virtual and real environments is provided which permits the user to perform significant physical exertion by applying forces to the machine while viewing images in such a fashion that the user perceives that he is exerting forces against the objects the images represent. The system may incorporate advanced image presentation methods, such as computer graphics, holography, or 3-D motion picture projection, and may include interactive entertainment, competitive, or motivational software, together with sound, vibration, and other special effects to enhance the sense of realism experienced by the user.
BACKGROUND OF THE INVENTION
The positive benefits of exercise are well known and regular exercise is increasingly becoming a part of the modern lifestyle. Particularly in urban and suburban settings, exercise available in health clubs is often repetitive and boring. Attention and motivation are key to success but are often neglected or not creatively approached by designers of exercise equipment. Concurrently, the fields of robotics, simulation, computer graphics, and virtual reality, are all advancing rapidly. These are being applied in entertainment, industrial, military, space, and rehabilitation medicine applications, but have received little attention in conditioning and sports medicine.
Generally, in interactive robotics using such methods as remotely actuated mechanisms or in simulators, controls incorporating force sensing and position detection may be used. Virtual reality systems applied to simulations such as preparation for space missions or strictly for entertainment also use force and position detectors. But these inventions differ fundamentally from exercise machines in that they are not specifically adapted to apply the type of load to the user which results in the most beneficial conditioning effect. In exercise machines, exertion to the point of fatigue, as well as to the limits of strength, are important to the achievement of health benefits by the user. Exercise machines are designed to maximize the safe strain on the user. By contrast, interactive robotics and machine controls are designed to minimize strain on the operator and to reduce fatigue thereby increasing worker output and accuracy.
Computer controlled exercise machines which utilize a virtual reality or hybrid environment must be robust, must be capable of applying loads up to the usual limits of human strength, and
must be suited to many repetitive cycles Delicate sensors susceptible to flex fatigue failure are undesirable Components must be adapted to withstand exposure to sweat without corrosion or electrical failure Skin abrasion and blister formation should be avoided Typical virtual reality components such as sensor-equipped gloves and body suits, and heavy visual display helmets are not well adapted to exercise applications
U S Patent No 5, 184,319 entitled FORCE FEEDBACK AND TEXTURES SIMULATING INTERFACE DEVICE discloses a force-actuating device which transmits force to a pressure-sensing body part and also includes sensors to measure the actual force applied to the body part as well as additional sensors to measure joint angles These components may be incorporated into a "feedback glove" permitting texture information and other simulation of grasping in a virtual reality system The force transmitters and sensors are intended for relatively low loads, such as at finger joints, and no applications for exercise are anticipated The invention is principally concerned with transfer of information at the man-machine interface rather than the transfer of significant conditioning forces at the interface as are required for effective exercise
U S Patent No 5,227,985 entitled COMPUTER VISION SYSTEM FOR POSITION MONITORING IN THREE DIMENSIONS USING NON-COPLANER LIGHT SOURCES ATTACHED TO A MONITORED OBJECT, relates to " monitoring the position and orientation of an object while it is displaced by an operator, who is enabled to interactively control various devices such as teleoperated mechanisms or graphics displays presented to his eyes while his motions are sensed and computed " The invention finds application in "communication methods with the virtual world", and although it could be incorporated as a component of a computer controlled exercise machine the patent makes no mention of and does not anticipate this use
U.S Patent No 4,934,694, entitled COMPUTER CONTROLLED EXERCISE MACHINE, discloses a device to apply computer controlled forces to the user for the purpose of exercise and to measure performance Although the device includes a display, no aspect of virtual reality is incorporated
A number of additional prior art patents disclose exercise inventions having computer control but do not include the concept of virtual reality or any perception that the user is exerting exercise force to manipulate the image of an apparent object in his visual field These include U S Patent No 4,235,437, entitled ROBOTIC EXERCISE MACHINE AND METHOD, and U.S Patent No 4,919,418 entitled COMPUTERIZED DRIVE MECHANISM FOR EXERCISE, PHYSICAL THERAPY AND REHABILITATION
Additional prior art patents incorporate games displayed on a video monitor together with
an exercise apparatus to increase motivation and interest during exercise or conversely, to include useful exercise in game playing. Among these are U.S. Patent No. 4,408,613, entitled
INTERACTIVE EXERCISE DEVICE, U.S. Patent No. 4,489,938, entitled VIDEO COMPUTER
SYSTEM CONTROLLED BY OPERATION OF ELASTOMERIC EXERCISE APPARATUS,
U.S. Patent No. 4,542,897, entitled EXERCISE CYCLE WITH INTERACTIVE AMUSEMENT
DEVICE, and U.S. Patent No. 4,558,864 entitled HANDGRIP EXERCISING, COMPUTER
GAME CONTROLLER.
SUMMARY OF THE INVENTION
Exercise machines ofthe present invention range from relatively simple means of creating the illusion in the mind of the user that he is exercising in a computer-presented virtual reality environment to more complex highly interactive systems in which the user can control the "place" in the virtual world that he chooses to "go" and the activities that he chooses to do in that "place". The former includes a virtual reality exercise bicycle in which the user controls the speed and direction of his trip on a virtual road via the handle bars, pedals and brakes, or a treadmill on which the user runs in a virtual race.
The term "virtual reality" as used in the present technological vernacular is not succinctly defined. It generally refers to computer systems which create the illusion that the user is actually present in a tangible physical environment and can interact with objects which appear to be present when in actuality the images viewed are of objects which are not physically present and are displayed using information present in the computer. As used in this patent document, the term "virtual reality" is used to refer to any illusionary environment which visually creates the false impression that the individual is actually present in a three-dimensional space and can interact with objects appearing to be present. The image presented may be derived from computer graphics, photography, or art work, so long as the false illusion is created that the individual is interactively present within the space represented by the image. As used in this patent document, the term "hybrid virtual reality" refers to the inclusion of real objects in addition to images representing objects that are not physically present, both within view ofthe individual experiencing the "hybrid virtual reality". For example, a system where the user both sees his real hand and sees the image of an object not actually present that appears to be held in his hand is an example of "hybrid virtual reality".
Interactive "virtual reality" or "hybrid virtual reality" exercise machines of the present invention provide a visual image of an illusory environment with which the user interacts by
applying force to the machine structure. The position of "virtual" objects presented in the visual field by a computer via a display such as a head-mounted video screen, the orientation ofthe user with respect to the "virtual" objects, and the forces applied between the user and the machine structure at various locations, are all monitored by the computer. As the user applies forces to the machine, the parts ofthe machine structure are physically moved, creating a changing arrangement of machine elements during use. The forces required of the user to cause these changes are controlled by the computer in such a way that the user perceives that he is interacting with the
"virtual" objects presented in the visual field and moving within the visual environment. Other components may be included to increase the sense of realism, such as loudspeakers to provide environmental sounds. The computer may utilize a CD-ROM device or other high capacity memory device with advanced graphics, and may also incorporate 3-D images.
One preferred embodiment ofthe invention is intended for installation in health clubs and gymnasiums. It includes a booth-like enclosure which contains an exercise mechanism and provides an externally darkened setting permitting images projected on a partially transparent visor to be clearly viewed. The user wears a lightweight exercise suit which has a fluorescent surface illuminated by ultraviolet light. Thus, the user sees both the image projected on the visor and his own body in superimposed alignment. The inside ofthe booth is a black-light-absorbent surface and, other than his own body, only the "handle grip" portions ofthe machine structure are visible to the user through the visor. The software, sensors, and computational algorithms are such that the computer matches the overlay of the projected image with the real positions of the exercise machine "handles" such that the handles appear to be part ofthe picture. There may be multiple such real physical objects which are part ofthe movable machine structure against which the user can exert forces, but other than them, in this embodiment, the user cannot "touch" the virtual object. In this sense, the device is a hybrid between actual reality and virtual reality.
As an example, to envision how an image incorporating such a "handle" is composed, imagine a simple view of a wall with a door within a frame. In the projected image, a picture ofa door handle is superimposed in the visual field over a cylindrical extension ofthe exercise machine, connected to a powered actuator. The headset and visor the user is wearing includes sensors which permit the computer to track its exact position and alter the projected image so that even when the user moves his head, the image appears stationary, with the door handle remaining in the same apparent position. The user reaches out grasps the handle and sees his hand grasp the handle. He simultaneously feels the handle because he is touching a real physical object. Then, as the user pulls the handle to open the door, the machine provides a particular force of resistance and path of
motion ofthe handle, which is kept in correspondence to a changing computer-generated image of the scene with the door swinging open on its hinges The user perceives that he has opened a virtual door, when actually he has only pulled on a portion ofthe machine's mechanism
The exercise machine may have many interchangeable "handles" of different size, shape, hardness, and texture, that are held in a storage area and may be removably affixed to the machine utilizing a mechanism similar to a tool changer such as used with a computer controlled milling machine This is accomplished without the attention ofthe user, and permits an appropriate surface to be contacted by the user to exert the necessary forces for the exercise while reasonably well matching the surface represented in the image Thus, during the exercise session, as the user reaches, grasps, and manipulates objects in the displayed image, the real objects he is contacting and moving provide the computer with position and force information necessary to update the display image As a further example, an exercise regimen might be composed of a trip through a virtual obstacle course, where the user has the sensation of climbing over objects, pushing objects aside, pulling his body up a pole, and the like, and the pace through the course, as well as the path chosen, may be determined by the user
Numerous other embodiments ofthe invention are possible, including devices which utilize full virtual reality in which the user sees only a computer display and does not directly also see either parts of his own body or any other real physical objects, hybrid embodiments where the user sees a video camera image of his own body superimposed over computer-generated images, and embodiments where two or more users participate simultaneously in a computer-mediated physical competition such as virtual bicycle racing or virtual team sports
OBJECTS OF THE INVENTION
Thus, the major objective of the present invention is to provide a computer controlled exercise machine which creates the illusion that the user is present in a visually-represented three- dimensional space containing objects with which the user can interact with sufficient exertion to enhance the user's health
Additional objects ofthe invention include the following,
(a) To provide relatively inexpensive exercise machines capable of implementing a great variety of exercise regimens through software,
(b) To provide "virtual reality" exercise machines and software capable of permitting interactive decision making by the user during the course ofthe exercise including the user's choice
of which objects in an image to manipulate or where within a "virtual environment" the user chooses to look or to go,
(c) To provide "virtual reality" computer controlled exercise machines capable of matching the strength and exertion level available to the physical capability ofthe user,
(d) To provide "virtual reality" computer controlled exercise machines capable of monitoring the physical performance ofthe user during a workout,
(e) To provide exercise machines utilizing numerically controlled multiple axis actuators to simulate muscular actions such as lifting, pushing, pulling or twisting objects by hand, and to simulate other actions ofthe body such as walking, running, or pedaling,
(f) To provide computer controlled exercise machines by which two or more individuals can compete in real time simulated physical competition such as bicycle racing,
(g) To provide computer controlled exercise machines equipped to stimulate several senses including vision, touch, sound, etc. and thereby enhance the user's perception that computer generated images are real objects,
(h) To provide computer controlled exercise machines utilizing a combination between "virtual reality" and actual reality thereby achieving many ofthe visual effects available with entirely computer-generated environments while applying the high forces necessary for physical conditioning utilizing rugged mechanical components capable of withstanding repeated use,
(i) To provide sensor and actuator components of "virtual reality" exercise machines well suited to the functional interface between the user and the computer under the stress and strain of exercise.
(j) To provide attention-getting, emotionally stimulating mental activity integrated with physical exercise tasks, including decision-making, satisfaction of curiosity, competition, game playing, and excitement,
(k) To provide photographic means of recording and integrating video camera images into "virtual reality" environments, such as those which may be used with exercise machines.
BRIEF DESCRIPTION OF THE DRAWINGS
Figure 1 is a perspective view of an embodiment ofthe invention, including a booth-like enclosure, exercise apparatus with ball screw actuators, sensors, computer, treadmill, and an image display helmet.
Figure 2 is a diagrammatic drawing of a user wearing a head-mounted display helmet and gripping a handle of an exercise apparatus.
Figure 3 is a perspective view of a head-mounted display and an array of sensors utilized to detect its location and orientation in space
Figure 4 is a drawing of one sensor assembly utilized to detect head-mounted display (helmet) position
Figure 5 is a perspective view of a 3-axis load-sensing device utilized to detect forces applied between the user and the machine structure via a handle. The position of a cover is indicated by the broken line
Figure 6 is a perspective view of an additional embodiment of the invention including a booth-like enclosure, exercise apparatus with boom and actuators, sensors for both handle position measurement and leg position measurement, and the computer.
Figure 7 is a diagrammatic top view of an array of two video cameras oriented in the north/south direction
Figure 8 is a rectilinear map of the combined view area of both ofthe two video cameras illustrated in Figure 7, with the outlined presentation field ofthe video display screen shown in two different positions relative to the border ofthe overall camera view area.
Figure 9 is a block diagram ofthe preferred embodiment ofthe present invention.
Figure 10 is the computer program ofthe preferred embodiment ofthe present invention for using to superimpose a computer generated image ofa scene containing virtual objects over the position ofthe parts ofthe exercise machine over against which the user applies force, such as the handles. The program controls the actuators of the machine to apply forces and motions to the handles which correspond with the image and would be appropriate to move a virtual object ofthe mass and inertia which the image represents
SOME EMBODIMENTS OF THE INVENTION Figure 1 illustrates a typical embodiment ofthe invention in which a hybrid between "virtual reality" and genuine reality is used In this hybrid mode, the user sees both a computer- displayed image of a virtual environment, and some especially visible objects within the real environment, including the user's arms and hands, as well as the handles ofthe apparatus. In other embodiments utilizing more complete "virtual reality", the user may see only a computer display and may not directly view any part of his body or his real environment The embodiment illustrated in Figure 1 is also simplified in that only two handles are provided, and the user either stands or runs on a treadmill or sits on a seat More complex embodiments which also include active interfaces for the feet, such as boots attached to servo-controlled mechanisms may be provided. Thus, pedaling,
climbing, and other actions can be simulated.
Referring to Figure 1, the user 2 stands on a treadmill 4 within a generally darkened booth 6. The user wears a display helmet 8 having a partially transparent projection screen 10 before his eyes, and a television projection system 12 mounted to the helmet. Audio speakers may also be provided in the helmet as well as a microphone to pick up the user's voice in embodiments where voice commands are recognized by the computer. Generally, various motors and sensors are illustrated in Figure 1 which are connected to the computer via cables which are not included in the drawing. Details ofthe computer controls are also omitted for simplicity.
In the present embodiment, the television projection system utilizes fiber optics to transfer light beams from the main unit located remotely together with the computer 13, to the helmet 8 via an optic cable 14. Separate images may be projected to each eye to provide stereoscopic 3-D images. The partially transparent projection screen may be a half mirrored sheet of glass or other transparent material. Special coatings may be used to enhance the clarity ofthe image.
The user wears an exercise suit similar to a body stocking with gloves, which has a fluorescent surface highly visible under a particular wavelength oflight, such as ultraviolet, shined on the user by a lamp 16. The handles of the exercise machine are also visible under this wavelength, but the remainder of the apparatus, including the structural frame members, servo¬ motors, force sensors, and the walls ofthe enclosure are all light- absorbent and essentially invisible to the user. Thus, when an image of proper intensity is projected on the screen ofthe helmet, the user sees it superimposed on the real objects in his visual field which are specially illuminated and fluorescent. The projected images and the real images are maintained in proper superimposed alignment by the computer system, which monitors the exact position of the helmet and the positions ofthe machine handles.
Still referring to Figure 1 the apparatus includes multi-axis, servo-controlled, actuators for each ofthe two handles, which are capable of positioning the handles to accommodate a wide range of motion for the exercising user, and are capable of applying high loads to the handles and measuring the applied forces. The major components ofthe actuator for the left handle 18 include a frame 20, slide 22, ball screw 24, ball nut 26, and servo-motor with an encoder to measure shaft rotational position 28, for the x-axis; a frame 30, slide 32, ball screw 34, ball nut 36, and servo motor with an encoder to measure shaft rotational position 38, for the y-axis; a frame 40, slide 42, ball screw 44, ball nut 46, and servo motor with an encoder to measure shaft rotational position 48, for the z-axis; and high torque servo-motors with encoders and gear heads, 50 and 52, for two additional axes of rotational motion ofthe handle. A multi-axis
force transducer 54 (see Figure 5) provides the computer 13 with a measurement ofthe magnitude and direction of force applied by the user to the handle (which also represents the force applied by the machine to the user). In this configuration, three strain gages 53 are mounted on a slotted block such that x-, y-, and z- axis forces are sensed.
A similar frame with servo-motors and sensors generally indicated at 58 is also provided for the right handle 60.
During exercise, the user may walk or run on the treadmill 4, or may sit on a seat, 64 which can be swiveled into or out of position by servo-motor 66.
Referring to Figure 2, details ofthe helmet and function ofthe hybrid virtual reality mode are illustrated. The partially transparent screen 10 receives an image projected onto it from the fiber optic cable 14 through a lens 68, via two mirrors 70 and 72. Light-ray path 74 from the screen to the eye coincides with the final portion of light ray path 76 from the hand to the eye. The light travels a distance between the lens and the screen of approximately 18 inches which is a typical distance from an average adult's eye to his hand when the elbow is partially bent. Such a position, represented in Figure 2 is a very common one when utilizing the arms for relatively heavy lifting. Thus, during use ofthe exercise machine, the user's hands will be relatively close to this distance from his eye much ofthe time. When the user looks at his own hand, his eye will focus on it. The handle ofthe machine will be at this distance, ifthe user is holding it, and so it will also be in focus. Since the part of the image projected on the screen which represents the part of the virtual environment that the user is touching with his hand is superimposed at approximately the same focal distance, the image can also be presented in good focus. The lens in front ofthe fiber optic cable can also be moved by a servo-motor to adjust the focus to the real distance between the handle and the screen, because this distance is calculable by the computer from helmet position and handle position information.
Figure 3 illustrates the system utilized to provide the control information to the computer to permit calculation ofthe exact helmet position, and therefore to permit proper superimposition ofthe projected image with the real image viewed through the partially transparent screen. Three arrays of three transducers are utilized as illustrated generally at I, II, and III, in Figure 3. Each of these transducers is comprised of a cable wound around a spring-retracted take-up spool which is in turn mounted to a multiple turn potentiometer or to a digital encoder, as shown in Figure 4. In Figures 1, 3, and 6, only the take-up spools are shown, for clarity. As illustrated in Figure 4, the cable 78 passes through an eyelet 80 and is wrapped around a spool 82 loaded by a band spring 84. As the cable is pulled out or permitted to be retracted by the spring, the length between the end of
the cable (off the spool) and the eyelet is a function ofthe rotational position ofthe spool and the number of wraps of cable on the spool. The potentiometer 86 attached to the spool thus gives an analog signal proportional to the length of cable beyond the eyelet. Use of an A to D converter or suitable digital encoder in place ofthe potentiometer provides an appropriate input signal to the computer.
Referring again to Figure 3, the three transducers in array I, 88, 90, and 92, are rigidly attached to the ceiling ofthe exercise booth as shown in Figure 1. Their three eyelets, designated as points A, B, and C, thus form a triangle of known dimensions and orientation with relation to the frame ofthe exercise apparatus. The three cables are attached to a common point on the back of the helmet designated as point D. Since distances ab, be and ac are fixed and are known, and since the transducer output signals permit calculation of lengths ad, bd, and cd, a tetrahedron with vertices ABCD is completely defined with the lengths of all sides known and the angles of one side (abc) known. By simple trigonometry and solid geometry, this permits the location of point D to be exactly calculated in three-dimensional space with relation to the exercise machine frame. The cables from transducer array II likewise are connected to the top front ofthe helmet at point E, and the cables of array III are connected to the helmet at point F. Thus, the transducer signals from all three arrays, taken together with information defining the position of all the eyelets permits computation of the exact position and orientation of the plane DEF through the helmet. This, together with the geometry ofthe helmet screen 10 in relation to plane DEF, permits computation ofthe position ofthe helmet screen.
Many other methods are available to compute the position ofthe helmet including light- beam measuring methods, and the use of helmets connected to rigid booms with position sensors. The present method employing an array of cables has the advantage that the spring forces on the cables can be selected so as to support most of the weight of the helmet, and the user can then move his head readily with little resistance.
Just as the specific device and method used to compute the position ofthe visual display screen can be varied without departing from the scope ofthe invention, the type of mechanism used to apply the forces of exercise, and the type of sensors used to detect the forces from moment to moment at the interfaces between the user and the machine may also be varied.
Figure 6 illustrates an embodiment ofthe invention in which mechanisms and actuators to apply forces to the handles are configured to present a minimal amount of hardware within the user's visual field, and thus be effectively hidden from the user's attention. Referring to the mechanism actuating the user's left handle in Figure 6, a hollow boom 94 supports the handle 96
mounted on rod 98 passing therethrough via a force-sensing array 54 which may be similar to that illustrated in Figure 5 The boom passes through a rectangular slot 100 in a wall of the exercise booth which restricts rotational motion ofthe boom around its long axis The rod, passing through the hollow core ofthe boom, is mounted on rotary bearings at each end and is connected to the shaft ofa servo-motor 99 mounted on the opposite end ofthe boom The boom is supported by six actuator cables 101, 102, 103, 104, 105, and 106 which pass through wear-resistant eyelets mounted to the wall ofthe booth Three ofthe actuator cables are each fixed near one end ofthe boom, the remaining three are fixed near the other end ofthe boom, and each cable is also attached to a spool mounted to a servo-motor For example, cable 106 is fixed to the boom at 108, passes through eyelet 1 10, and is wrapped on spool 131 connected to servo-motor 114 having encoder
1 16 In addition to the six actuator cables, an additional three position-sensing cables may be utilized paralleling the portions of each actuator cable between the boom and eyelet These sensing cables are represented by a dashed lines 1 18, 120, and 122 Each is connected to a spring loaded take-up spool and potentiometer as illustrated in Figure 4, shown at 124, 126, and 128 These positioning sensors provide a redundant control method to assure that there is no slack in the actuator cables
The six actuator cables are led via the eyelets to each of six actuator motor spools 131, 132, 133, 134, 135, and 136 The rotary position of each spool is detected by a separate sensor such as a motor shaft encoder which provides the computer with control signals proportional to the length of each cable which are maintained taut The motors and encoders for each cable attached to each spool are similar to those shown for cable 105 (motor 114, encoder 116) but most ofthese have not been included in the drawing for clarity The physical arrangement ofthe boom, cables and eyelets defines two tetrahedra having a common side and vertices opposite the common side which lie near the ends ofthe boom Given that the lengths ofthe sides ofthe common bases ofthe two tetrahedra are known and all the lengths ofthe other sides are determined (by the control signals) the position ofthe boom in three-dimensional space is calculated by the computer Via the servo¬ motors and this frame, cable, and boom mechanism, and information from the force sensors attached to the handle at 54, software in the computer can be run to move the handle to any position within its operational range, and the force and direction of force applied by the boom in each position can be controlled
The position of another boom for an actuator mechanism for the right handle is illustrated at 138 but the cables, spools, servo-motors and sensors are not shown in the drawing to avoid confusion with the components ofthe left handle actuator mechanism
Figure 9 illustrates the basic components ofthe computer and its interface with the sensors and the actuators ofthe exercise machine The central processing unit (CPU) 51 conducts the main computing functions ofthe computer and is connected both to memory 53 which stores data about both the virtual environment with virtual objects and about the real environment including the physical positions, motions and forces applied to and by the exercise machine structure and its actuators The CPU receives inputs from various sensors including handle position sensors 55, handle force sensors 57, and helmet position sensors 59 which offer appropriate signal conditioning by signal conditioners 61, 63, and 65, which may include analog-to-digital converters, amplifiers, or certain preprocessors Based on the data from the sensors, by making appropriate calculations according to the system's programs, the CPU sends the necessary output signals to the video controller 69 which in turn drives the video display 71 to present the image to the user The CPU also sends and receives servo control signals from servomotors which actuate the exercise machine via the servomotor interface 73
Referring to Figure 10, the basic computer program which is used superimposes an image of a scene containing virtual objects ofthe position of the parts of the exercise machine against which the user applies force, such as the handles The program controls the actuators of the machine to apply forces and motions to the handles which correspond with the image and would be appropriate to move a virtual object of the mass and inertia which the object represents The program utilizes data from the sensors described above which permits computation ofthe position ofthe handles and the forces applied thereby and thereto Data from the sensors ofthe position of the helmet in relation to the physical structure ofthe exercise machine is obtained, which is used to compute the vantage point of the user (corresponding to the physical position of the display screen which is mounted in the helmet) The program generates the necessary control and display signals to maintain superimposition ofthe computer image and the handle, and to match the force applied by the exercise machines actuators to the appropriate force to move the virtual object Many variations of computer program organization and function may be employed, depending upon the exact hardware used with each embodiment of the invention or particular virtual reality environment and objects which the software writer wishes to provide As an illustration, intended to be illustrative ofthe function of a typical program provided to run the exercise machine, is the example given in the SUMMARY OF THE INVENTION section, where opening a virtual door is described In that example, the basic sequential steps ofthe computer program, as illustrated in figure 10, are as follows 75 Start
77 Set clock to t=0 sec.
79 Read the computer memory data file of the virtual objects (the door, door frame and other details ofthe scene) from computer memory including data concerning geometry, mass and motion.
81 Sense the three dimensional position ofthe exercise machine handle at t=0 sec.
83 Sense the three dimensional position and orientation of the helmet mounted display screen at time t=0 sec.
85 Compute the perspective view of the virtual objects as seen from the helmet position relative to the exercise machine handle at t=0 sec.
87 Display the perspective view ofthe virtual object (door handle) on the helmet screen visually superimposed over the real three dimensional position ofthe exercise machine handle at t=0 sec.
(Note that because the user will see his or her own hand through the display screen actually grasping the machine handle, it will appear that the user is grasping a virtual object [door handle].)
89 Sense the forces applied to the exercise machine handle by the user at t=0 sec. in the various axes measured.
91 Compute the velocity, acceleration, and direction ofthe motion that the sensed forces would impart to the virtual objects if they were real.
93 Compute the position, velocity and acceleration that the virtual objects would reach in a finite time interval (about 1/30 sec. For one frame) between t=0 and t=.03 sec.
95 Update the computer memory data file.
97 Compute the force the object would exert on the handle and other virtual objects at time t=.03 sec.
Including the force vectors in the direction of each axis.
99 Compute and send servo command signals to the actuator motors to move the actuator handle to the position the part ofthe virtual object corresponding to the corresponding to the machine handle would reach at time t=.03 sec. calculated in step 9 above and apply the forces calculates in step 10 above.
111 Sense the three dimensional position and orientation ofthe helmet mounted display screen at time t=.03 sec.
113 Compute the perspective view ofthe virtual object (door handle) on the helmet screen visually superimposed over the real three dimensional position ofthe exercise machine handle at t=.03 sec.
(Note that because the user will see his or her hand through the display screen, and the user's hand will move together with the exercise machine handle between t=0 sec. and t=.03 sec. It will appear that the user is moving the virtual object [door handle].)
77 Reset the clock to t=0 sec.
Repeat steps 79-1 15 or stop.
The embodiment shown in Figure 6 includes two position- sensing arrays of cables, spools, and potentiometers, which are provided to sense the positions of the user's feet. These arrays, generally indicated at 140, and 142, are attached to fasteners 144, 146, on the lateral sides ofthe user's shoes. As the user walks or runs on the treadmill, the length of his stride and frequency is calculated, permitting the computer to monitor his speed, compute the distance he travels, and implement other functions, such as automatically adjusting the speed ofthe treadmill to the speed ofthe user, so that he does not need to pay attention to the treadmill and has a greater sense that he is walking on the surface ofa virtual environment as presented by the headset visual display.
In some embodiments ofthe invention, the virtual environment viewed by the user may be a composite of photographic images with superimposed computer-generated images, and additionally may incorporate superimposed images of the user's body and parts of the exercise machine which are directly viewed.
The embodiment of Figure 6 utilizes a video camera 148, mounted to the headset so as to receive an image ofthe field of view in front ofthe user via light reaching the lens as indicated by arrows 149. The user wears fluorescent clothing preferentially illuminated compared to the background and mechanism via ultraviolet light form lamp 16. Alternatively, other wavelengths and special filters may be used with the camera, 148. The image ofthe user's arms, hands, body, legs, and feet thus obtained is superimposed by the computer, over other images on the video display screen, 150, shown here inside a helmet, to create the complete virtual environment scene. In this way, no sensors are required on the user to provide position information for an image ofthe user's body on the screen. The video image ofthe user, of course, moves in perfect synchrony with him, which helps to provide a very realistic virtual reality illusion with minimal computational requirements compared to other methods, such as animated real-time images representing the user's body presented on the screen.
A device to provide appropriate photographic images of remote locations which may be used in combination with computer graphics, includes video cameras 152, and 154 which are equipped with sensors to monitor the position while the video image is recorded and to make a simultaneous recording of this position information to orient the image on the screen when it is replayed. The camera may be equipped with a compass 156 and a level detector 158 which record the orientation ofthe camera's field of view in reference to magnetic north and to the level ofthe horizon. The camera may alternatively be equipped with a gyroscopic position-sensing system or
other system such as a satellite locator receiver to record its position in space.
To achieve the "virtual reality" effect which in this case represents the perception by the user that he is present in a three-dimensional place represented by the images he sees and can interact there, the video image recorded is a wide angle image, preferentially as close to the full 360 degree field of view as is practical with advanced optics, including multiple cameras if necessary. The recorded field of view ofthe scene is thus much wider than the field of view presented on the screen during playback. The field of view ofthe playback display is no wider than the normal human field of view and may be considerably narrower. In Figure 7, the cameras 152 and 154 are equipped with wide angle lenses 160 and 162 which each provide a field of view 150 degrees wide by 150 degrees high. The two combined images thus represent a field 300 degrees by 150 degrees. This is represented by the large rectangle 164 of Figure 8 which is a flat map ofthe field.
The playback and viewing device includes a head-mounted display and computer system to match the portion ofthe recorded image displayed to the actual orientation in space ofthe head- mounted display screen. Thus, the correspondence ofthe displayed image to the orientation ofthe head mounted screen provides the user with the view that he would have had if he had been at the actual location where the camera was situated when the video was recorded and moved his head to look in different directions.
With the camera level, the horizon is represented by line 166 in Figure 8, and with the camera array directed north as illustrated in Figure 7, the position ofthe north direction is indicated by line 168 on the map ofthe camera view field in Figure 8. With the camera array pointed in any direction, the position information encoded by the compass and level detector permit the computer to locate reference points on the camera field map with regard to the horizontal/vertical and north/south directions. The small solid line rectangle 170, in Figure 8 corresponds to a viewing screen angular field of 160 degrees by 100 degrees, somewhat less than the normal human combined right-eye/left-eye total angular field of about 180 degrees by 115 degrees. The computer system utilizes position-sensor information to ascertain the actual physical orientation ofthe screen in the helmet, and when the helmet is level and facing due north, the computer will present an image recorded by a camera in the same position (level and facing north) as indicated by rectangle 170. If the user then turns his head upward and to the left the computer will present an image on the screen corresponding to the portion ofthe map represented by dotted rectangle 172. The computer also is provided with software permitting the orientation of image to be adjusted to the north/south alignment of the booth within which the machine is contained such that absolute direction is unimportant during use.
In systems where the replayed camera-recorded video is only part of the overall image perceived by the viewer, a computer- generated image of an object may be superimposed on the recorded scene. The position ofthe computer-generated object image in the visual field as well as the perimeter ofthe image are contained in computer memory for each frame ofa motion picture output by the computer. This position and perimeter information is utilized to turn off the portion of the recorded video that would have been displayed within the perimeter of the computer- generated object and thus avoid projecting two different overlaying images onto the viewing screen simultaneously. The result is a camera- recorded background with computer-generated objects filling in voids of matching size, position, and shape, deleted from the camera recorded image. Alternatively, the "foreground" objects may originate from camera-recorded images and the background may be computer-generated, or a combination may be used, such as in embodiments where a video camera receives the image ofthe user's body.
Examples of "virtual reality" exercise machines using camera-recorded images and camera- recorded images combined with computer-generated images include bicycle ride and bicycle race configurations. For example, to create the visual background, a very wide angle video camera with orientation sensors and recorder is mounted on a vehicle and driven along a road at a constant speed with no other traffic while the camera is photographing the scene. Ifthe speed is a typical bicycling speed and the video is recorded at a somewhat high number of frames per second, such as 30, then by varying the playback speed it will appear that the bicycle is traveling at a rate of speed which can be either more or less than the actual speed ofthe vehicle when the recording was made. The rate at which the pedals ofa stationary exercise machine are rotated represents the rate at which a real bicycle would travel while pedaling. The computer is programmed to vary the playback rate ofthe video recording as a function ofthe pedal speed such that the user perceives motion at the speed he is pedaling. Algorithms to represent coasting may be incorporated. With the head-mounted video display the user can turn his head and see the appropriate scene as if he were looking in the particular direction his head is oriented while riding a bicycle on a real road. This creates a very realistic interactive "virtual reality" effect.
Ifthe road on which the video camera recording is made is entirely straight, the simulated trip made along the road would not require simulated steering. For example, ifthe road ran due north, and the person on the exercise machine held his head straight (i.e. not turned to either side) as the video recording was played back, he would face due north along the center of the road during the entire trip. However, ifthe road along which the video recording was made was curved, say, 90 degrees and headed east, as the person exercising viewed the portion of the tape
corresponding to the curve, he would have to turn his head 90 degrees to the right retain a view along the center ofthe road. The apparent forward direction of his bicycle would remain north, and it would seem that he should travel off the side of the road. To avoid this, a simulated steering mechanism and control system is provided. The position of the exercise bicycle handle bars are measured by a sensor that provides a control signal to the computer. The north/south orientation ofthe image displayed is modified according to this control signal to permit the exerciser to follow the road and maintain the centered alignment of his body and the exercise bicycle with an apparent course down the center ofthe road. The computer may calculate the radius of curvature that the path ofa real bicycle would make at a given speed and adjust the rate of modification ofthe visual image to correspond to this. Thus, if the exerciser were stopped and not pedaling at all and he turned the handle bars, the image would appear stationary unless he turned his head. But if the exerciser were peddling and turned the handle bars without turning his head, the image would move as ifthe bicycle were turning.
In a bicycle race exercise machine, the same video recording and system as described in the preceding paragraphs can be enhanced by providing a computer-generated image ofa competing bicycle rider superimposed on the roadway and background scenery. The user will thus see the competitive image either ahead or behind by turning his head around (provided a wide enough field of view is photographed). The user may "compete" against a computer-simulated competitor, or two exercise machines with two competing individuals may be interconnected, such that the position ofthe image ofthe competitive rider that each sees corresponds to the appropriate position represented by the complete history of pedal rotation ofthe real competitor since the beginning of the race. This creates a "real" race in a virtual environment. Innumerable variations of this are possible utilizing machines which simulate many kinds of physical activity in many settings.
The information disclosed in the description of the present invention is intended to be representative of the principles that I have described. It will thus be seen that the objects of the invention set forth above and those made apparent from the preceding description are efficiently obtained and that certain changes may be made in the above articles and constructions without departing from the scope of the invention. It is intended that all matter contained in the above description or shown in the accompanying drawings shall be interpreted as illustrative but not in a limiting sense. It is also understood that the following claims are intended to cover all of the generic and specific features ofthe invention herein described and all statements ofthe scope ofthe invention which as a matter of language might be said to fall there between.
Claims
I claim
1 An interactive computer controlled exercise system adapted to apply multiple forces with motion to the body of the user corresponding with the computer generated visual images of an environment presented on a display, said system comprising: a a movable powered structure means having movable components connected to a fixed support frame means adapted to apply forces to the body ofthe user in conjunction with forces applied by the user to the movable components of said movable powered structure means, so as to provide health benefit to the user during exercise with said movable components of said movable powered structure means, b means for sensing the force and position of said movable components of said movable powered structure means and forces applied thereto by the user and to provide the sensed signals to a computer means, c means for detecting and measuring the position and orientation ofthe user's head and to provide the detected signals to said computer means, d powered actuator means for altering the three dimensional configuration and motion of said movable powered structure means and for applying forces of variable magnitude and direction to the body ofthe user in conjunction with corresponding forces applied by the user to said movable components of said movable powered structure means; e visual display means for presenting virtual images superimposed over the position of said movable powered structure means in real time maintained in visual perspective from the vantage point ofthe user when the user moves, said virtual images comprise a pictorial representation ofa virtual environment with virtual objects oriented and movable relative to the user such that if real objects existed in place of said virtual environment, contact between said virtual objects and the user would apply forces to the user's body similar to the forces applied to the user's body by said movable components of said movable powered structure means, and f said computer means, responsive to the sensed signals from said sensing means and the detected signals from said detecting means, for calculating said three dimensional configuration and motion of said movable powered structure means and forces applied thereto by the user and the position of the user's head in relation to said fixed support frame means, thereby to control said power actuator means and said visual display means
2 The exercise system of claim 1, wherein said computer means comprises memory means and program means, said computer means adapted to a store data for defining said virtual objects in said memory and present virtual images of said virtual objects on said visual display means, b receive said sensed signals from said sensing means and said detected signals from said detecting means, and compute said three dimensional configuration and motion of said movable powered structure means, the direction and magnitude of forces applied thereto by the user, and the position and orientation of said user's head relative to said fixed support frame, c control the motion of said movable powered structure means via said power actuator means, and present a virtual image of said virtual objects on said visual display means as seen from said user's vantage point superimposed over the position of said movable components of said movable powered structure means, so as to create the realistic illusion that said user is moving virtual objects when said user applies force to said movable components of said movable powered structure means during exercise
3 The exercise system of claim 2 further comprises head-mounted video camera means adapted to photograph the hands and body of said user and some movable components of said movable powered structure means which said user may contract, and said computer means adapted to superimpose the image of said user from said video camera on said virtual images on said virtual environment while at least one of said movable components of the movable powered structure means is moved by said power actuator means
4 The exercise system of claim 1 wherein said visual display means comprises a head-mounted partially transparent visual display screen
5 The exercise system of claim 3 further comprises light source means by which the user's arms and hands are preferentially illuminated so as to be more clearly visible than non-illuminated parts ofthe exercise system that lay within the user's visual field
6 The exercise system of claim 1 further comprises wide angle video camera means having at least one camera adapted to record an image encompassing a high proportion ofthe field of view surrounding said camera in all directions, and sensor and control signal recording means for measuring and recording reference information simultaneously with the recording of the image, wherein said computer means superimposes said recorded video image from said wide-angle video camera means with said virtual images of said virtual environment of said virtual display means
7 A "virtual reality" exercise machine comprised of a. visual display and computer means including hardware and software adapted to present a "virtual reality" environment scene to the user, b. mechanical means to provide the user physical exercise sufficient to benefit health while the user views said "virtual reality" scene, c. sensor means to provide information to the computer concerning position and movement ofthe user, and, d. computer control means to alter the "virtual reality" scene in correspondence with physical movements ofthe user while the user is exercising.
8. The "virtual reality" exercise machine of claim 7 in which said visual display is head- mounted, said mechanical means include a leg-operated actuator or treadmill, said sensor means monitor movements ofthe legs or feet, and said computer control means alter the "virtual reality" scene in correspondence to simulated locomotion ofthe user or simulated propulsion ofthe user in a vehicle or on a bicycle.
9. The "virtual reality" exercise machine of claim 7 in which said visual display is head- mounted, said mechanical means include an actuator adapted to be grasped by the user's hand, said sensor means monitor movements of said actuator and also monitor forces applied between the user's hand and said actuator, and said computer control means alter the "virtual reality" scene in correspondence to forces applied by the user to said actuator.
10. The "virtual reality" exercise machine of claim 7 in which force-sensing means and force- producing actuator means are included, and said computer-control means adjusts the magnitude and direction of force application necessary by the user to move "virtual objects" displayed in the visual field according to software run during the course ofthe exercise session.
11. The "virtual reality" exercise machine of claim 10 including means to maintain the forces applied by the machine within a range determined by preprogrammed exercise regimen software according to the strength and endurance level ofthe individual user.
12. Any computer game software adapted for use on a virtual reality exercise machine.
13. The computer game software of claim 12 adapted for interactive use with a virtual reality exercise machine in which the user perceives a virtual reality environment via a head-mounted display and must exert sufficient force to said machine to effect objects in the virtual environment and thereby play the game, whereby the user achieves physical exercise sufficient to benefit health.
14. The software of claim 12 for use with exercise machines whereby two or more individuals compete, in which said "virtual reality" exercise machines are networked together such that the image each user sees may be altered by any other user.
15 A "virtual reality" video camera and computer controlled system adapted to record photographic images and thereafter replay said images via a head mounted display, comprising a wide-angle video camera means including one or more cameras adapted to record a image encompassing a high proportion ofthe field of view surrounding said camera in all directions, b sensor and control signal recording means to measure and record reference information simultaneously with the recording ofthe image whereby the orientation ofthe image with respect to the horizontal direction and with respect to a north/south reference direction is labeled for recognition by a computer system during playback, c head-mounted visual video display means having a viewing screen before the user's eyes angularly extending approximately to the periphery ofthe normal human field of view or somewhat less, d head-orientation sensing and signal-generating means whereby the orientation ofthe viewing screen with respect to the horizontal direction and with respect to a north/south reference direction is input into a computer system, e computer image presentation control means whereby the recorded horizontal direction and north/south reference direction of the video camera recorded image is matched to a corresponding reference orientation ofthe head mounted display screen such that the recorded image presented on the screen at all times permits the user to see the portion ofthe overall scene which would have been within his visual field had his head been generally located where the camera was when the video was recorded and had his head been orientated with regard to the horizontal and north/south reference directions the same way it is oriented as the video is played back
16 The "virtual reality" exercise machine of claim 7 in which said "virtual reality" environment scene includes video images recorded and replayed utilizing the video camera and computer control system of claim 15
17 The method of creating a perceptual interface between a user and a computer comprised of a monitoring the position of a real object which the user can move via sensors which provide position information to the computer, b computing the position of said object within a coordinate system representing the space surrounding the real object, c scaling and representing said coordinate system on a computer display screen and presenting an image representing the way other objects would appear on said screen if they were within the real space surrounding said real object, d. superimposing a direct visual image or real time video image ofthe user's hands or other parts ofthe user's body on the coordinate position corresponding to said real object represented on the computer screen, such that the position ofthe image ofthe user's body relative to said coordinate position on the screen corresponds to the position ofthe user's body relative to said real object in the real physical space surrounding it.
18. The method of creating a perceptual interface between a computer and a user of claim 21 in which said real object which the user can move is affixed to the user's body.
19. A computer controlled exercise machine in which the perceptual interface method of claim 17 is utilized and means are included to provide the user with physical exertion while viewing said computer display and thereby obtain health benefit.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/513,671 | 1995-08-04 | ||
US08/513,671 US5577981A (en) | 1994-01-19 | 1995-08-04 | Virtual reality exercise machine and computer controlled video system |
Publications (1)
Publication Number | Publication Date |
---|---|
WO1997005926A1 true WO1997005926A1 (en) | 1997-02-20 |
Family
ID=24044221
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US1996/012369 WO1997005926A1 (en) | 1995-08-04 | 1996-07-29 | Interactive computer controlled exercise system |
Country Status (2)
Country | Link |
---|---|
US (1) | US5577981A (en) |
WO (1) | WO1997005926A1 (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7335134B1 (en) | 2004-02-23 | 2008-02-26 | Lavelle Richard | Exercise and game controller apparatus and method |
US7497807B2 (en) | 2003-07-15 | 2009-03-03 | Cube X Incorporated | Interactive computer simulation enhanced exercise machine |
WO2011025466A1 (en) * | 2009-08-25 | 2011-03-03 | Sefik Isik | SINGLE OR MULTIPLE SPORTS UNIT WITH 2D or 3D DISPLAY (FILM), SOUND AND AIR SUPPORT |
WO2013034981A3 (en) * | 2011-09-08 | 2013-06-06 | Offshore Incorporations (Cayman) Limited, | System and method for visualizing synthetic objects withinreal-world video clip |
US8620146B1 (en) | 2008-03-28 | 2013-12-31 | Theresa Coleman | Picture-in-picture video system for virtual exercise, instruction and entertainment |
CN104069633A (en) * | 2014-07-11 | 2014-10-01 | 上海体育学院 | Virtual reality generation system for table tennis ball serving |
CN106139519A (en) * | 2016-07-29 | 2016-11-23 | 韩莹光 | A kind of universal treadmill of mixed reality and application process thereof |
US10088911B2 (en) | 2016-12-30 | 2018-10-02 | Manuel Saez | Programmable electronic helmet |
US10810798B2 (en) | 2015-06-23 | 2020-10-20 | Nautilus, Inc. | Systems and methods for generating 360 degree mixed reality environments |
Families Citing this family (526)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5889670A (en) * | 1991-10-24 | 1999-03-30 | Immersion Corporation | Method and apparatus for tactilely responsive user interface |
US8352400B2 (en) | 1991-12-23 | 2013-01-08 | Hoffberg Steven M | Adaptive pattern recognition based controller apparatus and method and human-factored interface therefore |
US6801008B1 (en) | 1992-12-02 | 2004-10-05 | Immersion Corporation | Force feedback system and actuator power management |
US5629594A (en) | 1992-12-02 | 1997-05-13 | Cybernet Systems Corporation | Force feedback system |
US7345672B2 (en) | 1992-12-02 | 2008-03-18 | Immersion Corporation | Force feedback system and actuator power management |
US6437771B1 (en) | 1995-01-18 | 2002-08-20 | Immersion Corporation | Force feedback device including flexure member between actuator and user object |
US5734373A (en) | 1993-07-16 | 1998-03-31 | Immersion Human Interface Corporation | Method and apparatus for controlling force feedback interface systems utilizing a host computer |
US5739811A (en) | 1993-07-16 | 1998-04-14 | Immersion Human Interface Corporation | Method and apparatus for controlling human-computer interface systems providing force feedback |
US5805140A (en) | 1993-07-16 | 1998-09-08 | Immersion Corporation | High bandwidth force feedback interface using voice coils and flexures |
JP3030490B2 (en) * | 1994-03-18 | 2000-04-10 | 富士通株式会社 | Synchronous control method and device |
US5623582A (en) | 1994-07-14 | 1997-04-22 | Immersion Human Interface Corporation | Computer interface or control input device for laparoscopic surgical instrument and other elongated mechanical objects |
US6310604B1 (en) * | 1994-07-19 | 2001-10-30 | Asahi Kasei Kabushiki Kaisha | Virtual reality and telereality system |
US5890906A (en) * | 1995-01-20 | 1999-04-06 | Vincent J. Macri | Method and apparatus for tutorial, self and assisted instruction directed to simulated preparation, training and competitive play and entertainment |
AUPN298995A0 (en) * | 1995-05-16 | 1995-06-08 | Phillips, Scott | Vr walker |
US5913727A (en) * | 1995-06-02 | 1999-06-22 | Ahdoot; Ned | Interactive movement and contact simulation game |
US7824310B1 (en) | 1995-06-22 | 2010-11-02 | Shea Michael J | Exercise apparatus providing mental activity for an exerciser |
US5999168A (en) * | 1995-09-27 | 1999-12-07 | Immersion Corporation | Haptic accelerator for force feedback computer peripherals |
US6142913A (en) * | 1995-10-11 | 2000-11-07 | Ewert; Bruce | Dynamic real time exercise video apparatus and method |
US5754023A (en) * | 1995-10-26 | 1998-05-19 | Cybernet Systems Corporation | Gyro-stabilized platforms for force-feedback applications |
USRE39906E1 (en) | 1995-10-26 | 2007-11-06 | Immersion Corporation | Gyro-stabilized platforms for force-feedback applications |
US6430997B1 (en) * | 1995-11-06 | 2002-08-13 | Trazer Technologies, Inc. | System and method for tracking and assessing movement skills in multidimensional space |
US6308565B1 (en) | 1995-11-06 | 2001-10-30 | Impulse Technology Ltd. | System and method for tracking and assessing movement skills in multidimensional space |
US6639581B1 (en) | 1995-11-17 | 2003-10-28 | Immersion Corporation | Flexure mechanism for interface device |
US8508469B1 (en) | 1995-12-01 | 2013-08-13 | Immersion Corporation | Networked applications including haptic feedback |
US6028593A (en) | 1995-12-01 | 2000-02-22 | Immersion Corporation | Method and apparatus for providing simulated physical interactions within computer generated environments |
US6219032B1 (en) * | 1995-12-01 | 2001-04-17 | Immersion Corporation | Method for providing force feedback to a user of an interface device based on interactions of a controlled cursor with graphical elements in a graphical user interface |
GB9525047D0 (en) * | 1995-12-07 | 1996-02-07 | Philips Electronics Nv | Virtual body control device |
US6808472B1 (en) | 1995-12-14 | 2004-10-26 | Paul L. Hickman | Method and apparatus for remote interactive exercise and health equipment |
EP0780783A1 (en) * | 1995-12-23 | 1997-06-25 | Deutsche ITT Industries GmbH | Virtual reality reproducing system with an optical position determination device |
US5961541A (en) * | 1996-01-02 | 1999-10-05 | Ferrati; Benito | Orthopedic apparatus for walking and rehabilitating disabled persons including tetraplegic persons and for facilitating and stimulating the revival of comatose patients through the use of electronic and virtual reality units |
US5872575A (en) * | 1996-02-14 | 1999-02-16 | Digital Media Interactive | Method and system for the creation of and navigation through a multidimensional space using encoded digital video |
US5864333A (en) * | 1996-02-26 | 1999-01-26 | O'heir; Brian S. | Foot force actuated computer input apparatus and method |
US6046712A (en) * | 1996-07-23 | 2000-04-04 | Telxon Corporation | Head mounted communication system for providing interactive visual communications with a remote system |
JP3170638B2 (en) * | 1996-08-08 | 2001-05-28 | 谷 白糸 | Virtual reality experience device |
JP2917128B2 (en) * | 1996-08-08 | 1999-07-12 | 谷 白糸 | Walking experience device |
US5944530A (en) | 1996-08-13 | 1999-08-31 | Ho; Chi Fai | Learning method and system that consider a student's concentration level |
JPH1065996A (en) * | 1996-08-23 | 1998-03-06 | Olympus Optical Co Ltd | Head wearable display device |
JPH10151162A (en) * | 1996-11-25 | 1998-06-09 | Mitsubishi Electric Corp | Patient's bedside wellness system |
JP3469410B2 (en) * | 1996-11-25 | 2003-11-25 | 三菱電機株式会社 | Wellness system |
US5698782A (en) * | 1997-01-09 | 1997-12-16 | Ministry Of Solicitor General And Correctional Services | Hand and arm strength testing apparatus to be used in a constable selection process |
BR9807750A (en) * | 1997-02-24 | 2001-10-02 | Thermo Information Solutions I | Method and apparatus to operate the superposition of effects generated by the computer to position itself on a living image |
IL120507A (en) * | 1997-03-24 | 2001-06-14 | Keytron Electronics & Technolo | Exercise monitoring system |
US6050924A (en) | 1997-04-28 | 2000-04-18 | Shea; Michael J. | Exercise system |
US6425764B1 (en) | 1997-06-09 | 2002-07-30 | Ralph J. Lamson | Virtual reality immersion therapy for treating psychological, psychiatric, medical, educational and self-help problems |
US5883628A (en) * | 1997-07-03 | 1999-03-16 | International Business Machines Corporation | Climability: property for objects in 3-D virtual environments |
CA2242170C (en) * | 1997-08-04 | 2002-11-05 | Alfred Vaino Aho | Three dimensional virtual reality enhancement techniques |
US6522312B2 (en) * | 1997-09-01 | 2003-02-18 | Canon Kabushiki Kaisha | Apparatus for presenting mixed reality shared among operators |
US6050822A (en) * | 1997-10-01 | 2000-04-18 | The United States Of America As Represented By The Secretary Of The Army | Electromagnetic locomotion platform for translation and total immersion of humans into virtual environments |
US7190392B1 (en) | 1997-10-23 | 2007-03-13 | Maguire Jr Francis J | Telepresence system and active/passive mode display for use therein |
US6104382A (en) * | 1997-10-31 | 2000-08-15 | Immersion Corporation | Force feedback transmission mechanisms |
US6020875A (en) * | 1997-10-31 | 2000-02-01 | Immersion Corporation | High fidelity mechanical transmission system and interface device |
US6281651B1 (en) | 1997-11-03 | 2001-08-28 | Immersion Corporation | Haptic pointing devices |
DE29720110U1 (en) | 1997-11-13 | 1998-01-02 | Wittenbecher Rainer Dr Ing | Training device |
US6667741B1 (en) * | 1997-12-24 | 2003-12-23 | Kabushiki Kaisha Sega Enterprises | Image generating device and image generating method |
US6625299B1 (en) | 1998-04-08 | 2003-09-23 | Jeffrey Meisner | Augmented reality technology |
US6429846B2 (en) | 1998-06-23 | 2002-08-06 | Immersion Corporation | Haptic feedback for touchpads and other touch controls |
US6184868B1 (en) | 1998-09-17 | 2001-02-06 | Immersion Corp. | Haptic feedback control devices |
JP4487054B2 (en) * | 1998-12-17 | 2010-06-23 | バイオフィリア研究所有限会社 | Lower limb function training device |
US20040152058A1 (en) * | 2002-06-11 | 2004-08-05 | Browne H. Lee | Video instructional system and method for teaching motor skills |
US7904187B2 (en) | 1999-02-01 | 2011-03-08 | Hoffberg Steven M | Internet appliance system and method |
US7015950B1 (en) | 1999-05-11 | 2006-03-21 | Pryor Timothy R | Picture taking method and apparatus |
US6162189A (en) * | 1999-05-26 | 2000-12-19 | Rutgers, The State University Of New Jersey | Ankle rehabilitation system |
JP2000350865A (en) * | 1999-06-11 | 2000-12-19 | Mr System Kenkyusho:Kk | Game device for composite real space, image processing method therefor and program storage medium |
US6244988B1 (en) | 1999-06-28 | 2001-06-12 | David H. Delman | Interactive exercise system and attachment module for same |
US6545661B1 (en) | 1999-06-21 | 2003-04-08 | Midway Amusement Games, Llc | Video game system having a control unit with an accelerometer for controlling a video game |
US8029415B2 (en) | 1999-07-08 | 2011-10-04 | Icon Ip, Inc. | Systems, methods, and devices for simulating real world terrain on an exercise device |
US7166062B1 (en) | 1999-07-08 | 2007-01-23 | Icon Ip, Inc. | System for interaction with exercise device |
US7537546B2 (en) | 1999-07-08 | 2009-05-26 | Icon Ip, Inc. | Systems and methods for controlling the operation of one or more exercise devices and providing motivational programming |
US7985164B2 (en) * | 1999-07-08 | 2011-07-26 | Icon Ip, Inc. | Methods and systems for controlling an exercise apparatus using a portable data storage device |
US7628730B1 (en) | 1999-07-08 | 2009-12-08 | Icon Ip, Inc. | Methods and systems for controlling an exercise apparatus using a USB compatible portable remote device |
US6447424B1 (en) | 2000-02-02 | 2002-09-10 | Icon Health & Fitness Inc | System and method for selective adjustment of exercise apparatus |
US6135928A (en) * | 1999-08-20 | 2000-10-24 | Butterfield; Anthony | Virtual reality equipment |
US6749432B2 (en) * | 1999-10-20 | 2004-06-15 | Impulse Technology Ltd | Education system challenging a subject's physiologic and kinesthetic systems to synergistically enhance cognitive function |
US6822635B2 (en) | 2000-01-19 | 2004-11-23 | Immersion Corporation | Haptic interface for laptop computers and other portable devices |
US7328119B1 (en) * | 2000-03-07 | 2008-02-05 | Pryor Timothy R | Diet and exercise planning and motivation including apparel purchases based on future appearance |
US20070281828A1 (en) | 2000-03-21 | 2007-12-06 | Rice Michael J P | Games controllers |
GB0006672D0 (en) * | 2000-03-21 | 2000-05-10 | Rice Michael J P | Improvements relating to controllers |
JP3404651B2 (en) * | 2000-03-30 | 2003-05-12 | 財団法人理工学振興会 | 3D input device |
US6428449B1 (en) * | 2000-05-17 | 2002-08-06 | Stanford Apseloff | Interactive video system responsive to motion and voice command |
US20060122474A1 (en) * | 2000-06-16 | 2006-06-08 | Bodymedia, Inc. | Apparatus for monitoring health, wellness and fitness |
GB2368293A (en) * | 2000-08-30 | 2002-05-01 | Christos Farmakis | Doing aerobics through the Internet |
KR20020025325A (en) * | 2000-09-28 | 2002-04-04 | 편동현 | Local Simulation Health System using Internet |
US7112168B2 (en) | 2000-12-15 | 2006-09-26 | Icon Ip, Inc. | Selectively dynamic exercise platform |
USD493500S1 (en) | 2000-12-15 | 2004-07-27 | William T. Dalebout | Top surface of an exercise device |
DE10105965B4 (en) * | 2001-02-09 | 2004-06-09 | Peter-Raphael Von Buengner | Device and method for deriving electrical signals from a physical or physiological activity of a test person |
US8306635B2 (en) * | 2001-03-07 | 2012-11-06 | Motion Games, Llc | Motivation and enhancement of physical and mental exercise, rehabilitation, health and social interaction |
JP4424869B2 (en) * | 2001-03-16 | 2010-03-03 | 浜松ホトニクス株式会社 | Stride measuring device |
JP4009433B2 (en) * | 2001-03-29 | 2007-11-14 | 株式会社コナミデジタルエンタテインメント | GAME DEVICE, GAME PROGRAM, AND GAME SYSTEM |
JP3848101B2 (en) * | 2001-05-17 | 2006-11-22 | シャープ株式会社 | Image processing apparatus, image processing method, and image processing program |
IL143255A (en) | 2001-05-20 | 2015-09-24 | Simbionix Ltd | Endoscopic ultrasonography simulation |
US6743154B2 (en) | 2001-06-01 | 2004-06-01 | Neil B. Epstein | Omnidirectional moving surface |
US7588516B2 (en) * | 2001-07-23 | 2009-09-15 | Southwest Research Institute | Virtual reality system locomotion interface utilizing a pressure-sensing mat |
US7520836B2 (en) * | 2001-07-23 | 2009-04-21 | Southwest Research Institute | Virtual reality system locomotion interface utilizing a pressure-sensing mat attached to movable base structure |
US7381152B2 (en) * | 2001-07-23 | 2008-06-03 | Southwest Research Institute | Virtual reality system locomotion interface utilizing a pressure-sensing mat |
US7736272B2 (en) * | 2001-08-21 | 2010-06-15 | Pantometrics, Ltd. | Exercise system with graphical feedback and method of gauging fitness progress |
US6921351B1 (en) | 2001-10-19 | 2005-07-26 | Cybergym, Inc. | Method and apparatus for remote interactive exercise and health equipment |
JP2003126291A (en) * | 2001-10-19 | 2003-05-07 | Konami Co Ltd | Exercising aid control method and exercising aid apparatus |
US6990639B2 (en) | 2002-02-07 | 2006-01-24 | Microsoft Corporation | System and process for controlling electronic components in a ubiquitous computing environment using multimodal integration |
ES2192482B1 (en) * | 2002-03-13 | 2005-02-16 | Miguel Jimenez Laso | GYM AND SPORTS DEVICE WITH STEREOSCOPIC PROJECTION SCREEN. |
US6902513B1 (en) | 2002-04-02 | 2005-06-07 | Mcclure Daniel R. | Interactive fitness equipment |
US8066621B2 (en) * | 2002-08-15 | 2011-11-29 | Carlson Alan L | Exercise apparatus having a user interface which can move arcuately in three dimensions |
US7104926B2 (en) * | 2002-08-15 | 2006-09-12 | Dynabolic Gym Equipment | Exercising machine for working muscles that support the spine |
USD489778S1 (en) | 2002-10-18 | 2004-05-11 | Reebok International Ltd. | Portion of an exercise device |
AU2003283887A1 (en) * | 2002-11-20 | 2004-06-15 | John Hansen Ryall | Instruction method using virtual apparatus |
AU2003297716A1 (en) | 2002-12-08 | 2004-06-30 | Immersion Corporation | Methods and systems for providing haptic messaging to handheld communication devices |
US8059088B2 (en) | 2002-12-08 | 2011-11-15 | Immersion Corporation | Methods and systems for providing haptic messaging to handheld communication devices |
US8830161B2 (en) | 2002-12-08 | 2014-09-09 | Immersion Corporation | Methods and systems for providing a virtual touch haptic effect to handheld communication devices |
US6985240B2 (en) * | 2002-12-23 | 2006-01-10 | International Business Machines Corporation | Method and apparatus for retrieving information about an object of interest to an observer |
JP2004213350A (en) * | 2002-12-27 | 2004-07-29 | Seiko Epson Corp | Inner force sense presenting device and image correcting method |
JP2004246498A (en) * | 2003-02-12 | 2004-09-02 | Fanuc Ltd | Position control device |
US20050130741A1 (en) * | 2003-02-18 | 2005-06-16 | Raj Pandian | Human power conversion system based on video game play with exercise bicycle |
US8745541B2 (en) | 2003-03-25 | 2014-06-03 | Microsoft Corporation | Architecture for controlling a computer using hand gestures |
US7665041B2 (en) | 2003-03-25 | 2010-02-16 | Microsoft Corporation | Architecture for controlling a computer using hand gestures |
KR100526741B1 (en) * | 2003-03-26 | 2005-11-08 | 김시학 | Tension Based Interface System for Force Feedback and/or Position Tracking and Surgically Operating System for Minimally Incising the affected Part Using the Same |
AU2003901669A0 (en) * | 2003-04-01 | 2003-05-01 | Colin Richard Carbis | Hypnosis in a virtual environment |
US7046151B2 (en) * | 2003-07-14 | 2006-05-16 | Michael J. Dundon | Interactive body suit and interactive limb covers |
US7850456B2 (en) | 2003-07-15 | 2010-12-14 | Simbionix Ltd. | Surgical simulation device, system and method |
US20050054492A1 (en) * | 2003-07-15 | 2005-03-10 | Neff John D. | Exercise device for under a desk |
US7497812B2 (en) * | 2003-07-15 | 2009-03-03 | Cube X, Incorporated | Interactive computer simulation enhanced exercise machine |
US7231834B2 (en) * | 2003-07-28 | 2007-06-19 | Hamamatsu Photonics K. K. | Stride measuring apparatus |
US20050075213A1 (en) * | 2003-10-06 | 2005-04-07 | Arick Thomas P. | Exercise device independent, variable display rate visual exercise system |
FI117308B (en) * | 2004-02-06 | 2006-08-31 | Nokia Corp | gesture Control |
JP3904562B2 (en) * | 2004-02-18 | 2007-04-11 | 株式会社ソニー・コンピュータエンタテインメント | Image display system, recording medium, and program |
JP3851907B2 (en) * | 2004-02-18 | 2006-11-29 | 株式会社ソニー・コンピュータエンタテインメント | Image display system and video game system |
US20050216243A1 (en) * | 2004-03-02 | 2005-09-29 | Simon Graham | Computer-simulated virtual reality environments for evaluation of neurobehavioral performance |
US7224326B2 (en) * | 2004-03-03 | 2007-05-29 | Volo, Llc | Virtual reality system |
US7658695B1 (en) | 2004-04-27 | 2010-02-09 | Performance Health Technologies, Inc. | Position monitoring displays |
US7625316B1 (en) | 2004-04-27 | 2009-12-01 | Performance Health Technologies, Inc. | Position monitoring system |
US7503878B1 (en) * | 2004-04-27 | 2009-03-17 | Performance Health Technologies, Inc. | Position monitoring device |
GB0416042D0 (en) | 2004-07-17 | 2004-08-18 | Koninkl Philips Electronics Nv | Exercise system and method |
US20060017654A1 (en) * | 2004-07-23 | 2006-01-26 | Romo Justin R | Virtual reality interactivity system and method |
WO2006014810A2 (en) | 2004-07-29 | 2006-02-09 | Kevin Ferguson | A human movement measurement system |
US6912900B1 (en) * | 2004-07-30 | 2005-07-05 | Seoul National University Industry Foundation | Muscular motion simulating system |
WO2006023647A1 (en) * | 2004-08-18 | 2006-03-02 | Sarnoff Corporation | Systeme and method for monitoring training environment |
TWM261246U (en) * | 2004-08-27 | 2005-04-11 | Fitness Authority Ind Co Ltd | Electrically operated handstand machines |
US20060236120A1 (en) * | 2005-04-14 | 2006-10-19 | Ibm Corporation | Method and apparatus employing stress detection for highly secure communication |
US20060236121A1 (en) * | 2005-04-14 | 2006-10-19 | Ibm Corporation | Method and apparatus for highly secure communication |
US7918732B2 (en) * | 2005-05-06 | 2011-04-05 | Milton Charles Van Noland | Manifold compatibility electronic omni axis human interface |
US20060258458A1 (en) * | 2005-05-13 | 2006-11-16 | Addington David R | System and method for interfacing a simulation device with a gaming device |
US20060287089A1 (en) * | 2005-05-13 | 2006-12-21 | Addington David R | System and method for interfacing a simulation device with a gaming device |
US7864168B2 (en) * | 2005-05-25 | 2011-01-04 | Impulse Technology Ltd. | Virtual reality movement system |
US20070021207A1 (en) * | 2005-07-25 | 2007-01-25 | Ned Ahdoot | Interactive combat game between a real player and a projected image of a computer generated player or a real player with a predictive method |
US20070021199A1 (en) * | 2005-07-25 | 2007-01-25 | Ned Ahdoot | Interactive games with prediction method |
US20070171199A1 (en) * | 2005-08-05 | 2007-07-26 | Clement Gosselin | Locomotion simulation apparatus, system and method |
US20100197462A1 (en) * | 2005-09-07 | 2010-08-05 | Bvp Holding, Inc. | 3-d, interactive exercise analysis, gaming, and physical therapy system |
WO2007013833A1 (en) * | 2005-09-15 | 2007-02-01 | Oleg Stanilasvovich Rurin | Method and system for visualising virtual three-dimensional objects |
US20070136093A1 (en) * | 2005-10-11 | 2007-06-14 | Rankin Innovations, Inc. | Methods, systems, and programs for health and wellness management |
US7697827B2 (en) | 2005-10-17 | 2010-04-13 | Konicek Jeffrey C | User-friendlier interfaces for a camera |
US20070117081A1 (en) * | 2005-10-31 | 2007-05-24 | Ford John H | System and Method for Delivering Information to Optimize Information Retention |
US20070109259A1 (en) * | 2005-11-11 | 2007-05-17 | Xin Liu | Exploring platform for virtual environment |
US7509215B2 (en) * | 2005-12-12 | 2009-03-24 | Microsoft Corporation | Augmented navigation system |
US7780573B1 (en) * | 2006-01-31 | 2010-08-24 | Carmein David E E | Omni-directional treadmill with applications |
US20080110115A1 (en) * | 2006-11-13 | 2008-05-15 | French Barry J | Exercise facility and method |
US8054289B2 (en) | 2006-12-01 | 2011-11-08 | Mimic Technologies, Inc. | Methods, apparatus, and article for force feedback based on tension control and tracking through cables |
TWI321465B (en) * | 2006-12-29 | 2010-03-11 | Ind Tech Res Inst | Automatic evaluation method and system of cardio-respiratory fitness |
US8500451B2 (en) | 2007-01-16 | 2013-08-06 | Simbionix Ltd. | Preoperative surgical simulation |
US8543338B2 (en) | 2007-01-16 | 2013-09-24 | Simbionix Ltd. | System and method for performing computerized simulations for image-guided procedures using a patient specific model |
US8005238B2 (en) | 2007-03-22 | 2011-08-23 | Microsoft Corporation | Robust adaptive beamforming with enhanced noise suppression |
US7909741B2 (en) * | 2007-03-27 | 2011-03-22 | Dhkl, Inc. | Devices, systems and methods for receiving, recording and displaying information relating to physical exercise |
US8117137B2 (en) | 2007-04-19 | 2012-02-14 | Microsoft Corporation | Field-programmable gate array based accelerator system |
JP5427343B2 (en) | 2007-04-20 | 2014-02-26 | 任天堂株式会社 | Game controller |
US8005237B2 (en) | 2007-05-17 | 2011-08-23 | Microsoft Corp. | Sensor array beamformer post-processor |
US20080310707A1 (en) * | 2007-06-15 | 2008-12-18 | Microsoft Corporation | Virtual reality enhancement using real world data |
US8629976B2 (en) * | 2007-10-02 | 2014-01-14 | Microsoft Corporation | Methods and systems for hierarchical de-aliasing time-of-flight (TOF) systems |
JP5427346B2 (en) | 2007-10-05 | 2014-02-26 | 任天堂株式会社 | Load detection program, load detection device, load detection system, and load detection method |
JP5080196B2 (en) * | 2007-10-09 | 2012-11-21 | 任天堂株式会社 | Program, information processing apparatus, information processing system, and information processing method |
JP4382844B2 (en) * | 2007-10-31 | 2009-12-16 | 任天堂株式会社 | Weighting machine for adjustment and weighting method for adjustment |
US8127235B2 (en) | 2007-11-30 | 2012-02-28 | International Business Machines Corporation | Automatic increasing of capacity of a virtual space in a virtual world |
US10204525B1 (en) | 2007-12-14 | 2019-02-12 | JeffRoy H. Tillis | Suggestion-based virtual sessions engaging the mirror neuron system |
US20090164919A1 (en) | 2007-12-24 | 2009-06-25 | Cary Lee Bates | Generating data for managing encounters in a virtual world environment |
US20090166684A1 (en) * | 2007-12-26 | 2009-07-02 | 3Dv Systems Ltd. | Photogate cmos pixel for 3d cameras having reduced intra-pixel cross talk |
US8264505B2 (en) | 2007-12-28 | 2012-09-11 | Microsoft Corporation | Augmented reality and filtering |
US9649469B2 (en) | 2008-04-24 | 2017-05-16 | The Invention Science Fund I Llc | Methods and systems for presenting a combination treatment |
US9026369B2 (en) | 2008-04-24 | 2015-05-05 | The Invention Science Fund I, Llc | Methods and systems for presenting a combination treatment |
US9239906B2 (en) | 2008-04-24 | 2016-01-19 | The Invention Science Fund I, Llc | Combination treatment selection methods and systems |
US8615407B2 (en) * | 2008-04-24 | 2013-12-24 | The Invention Science Fund I, Llc | Methods and systems for detecting a bioactive agent effect |
US8930208B2 (en) | 2008-04-24 | 2015-01-06 | The Invention Science Fund I, Llc | Methods and systems for detecting a bioactive agent effect |
US8606592B2 (en) | 2008-04-24 | 2013-12-10 | The Invention Science Fund I, Llc | Methods and systems for monitoring bioactive agent use |
US20090270687A1 (en) * | 2008-04-24 | 2009-10-29 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Methods and systems for modifying bioactive agent use |
US9282927B2 (en) | 2008-04-24 | 2016-03-15 | Invention Science Fund I, Llc | Methods and systems for modifying bioactive agent use |
US8876688B2 (en) | 2008-04-24 | 2014-11-04 | The Invention Science Fund I, Llc | Combination treatment modification methods and systems |
US8682687B2 (en) | 2008-04-24 | 2014-03-25 | The Invention Science Fund I, Llc | Methods and systems for presenting a combination treatment |
US7801686B2 (en) * | 2008-04-24 | 2010-09-21 | The Invention Science Fund I, Llc | Combination treatment alteration methods and systems |
US9662391B2 (en) | 2008-04-24 | 2017-05-30 | The Invention Science Fund I Llc | Side effect ameliorating combination therapeutic products and systems |
US9560967B2 (en) | 2008-04-24 | 2017-02-07 | The Invention Science Fund I Llc | Systems and apparatus for measuring a bioactive agent effect |
US9449150B2 (en) | 2008-04-24 | 2016-09-20 | The Invention Science Fund I, Llc | Combination treatment selection methods and systems |
US7974787B2 (en) | 2008-04-24 | 2011-07-05 | The Invention Science Fund I, Llc | Combination treatment alteration methods and systems |
US9064036B2 (en) | 2008-04-24 | 2015-06-23 | The Invention Science Fund I, Llc | Methods and systems for monitoring bioactive agent use |
US20090280901A1 (en) * | 2008-05-09 | 2009-11-12 | Dell Products, Lp | Game controller device and methods thereof |
US8385557B2 (en) * | 2008-06-19 | 2013-02-26 | Microsoft Corporation | Multichannel acoustic echo reduction |
US8325909B2 (en) | 2008-06-25 | 2012-12-04 | Microsoft Corporation | Acoustic echo suppression |
US8203699B2 (en) | 2008-06-30 | 2012-06-19 | Microsoft Corporation | System architecture design for time-of-flight system having reduced differential pixel size, and time-of-flight systems so designed |
US9144709B2 (en) * | 2008-08-22 | 2015-09-29 | Alton Reich | Adaptive motor resistance video game exercise apparatus and method of use thereof |
US8639666B2 (en) * | 2008-09-05 | 2014-01-28 | Cast Group Of Companies Inc. | System and method for real-time environment tracking and coordination |
US20100062818A1 (en) * | 2008-09-09 | 2010-03-11 | Apple Inc. | Real-time interaction with a virtual competitor while performing an exercise routine |
US7901325B2 (en) * | 2008-09-12 | 2011-03-08 | Joe Henderson | Athletic training device |
US8131659B2 (en) | 2008-09-25 | 2012-03-06 | Microsoft Corporation | Field-programmable gate array based accelerator system |
US8301638B2 (en) | 2008-09-25 | 2012-10-30 | Microsoft Corporation | Automated feature selection based on rankboost for ranking |
JP4752896B2 (en) * | 2008-10-27 | 2011-08-17 | ブラザー工業株式会社 | Portable device |
US20100279822A1 (en) * | 2008-11-01 | 2010-11-04 | Ford John Hajime | Systems and methods for optimizing one or more audio tracks to a video stream |
JP5361349B2 (en) | 2008-11-28 | 2013-12-04 | 任天堂株式会社 | Information processing apparatus, computer program, information processing system, and information processing method |
US20100156760A1 (en) * | 2008-12-19 | 2010-06-24 | At&T Intellectual Property I, L.P. | Motion controlled multimedia content viewing method and system |
JP5806443B2 (en) | 2008-12-26 | 2015-11-10 | 任天堂株式会社 | Biological information management system |
US8681321B2 (en) * | 2009-01-04 | 2014-03-25 | Microsoft International Holdings B.V. | Gated 3D camera |
EP2211213A2 (en) * | 2009-01-21 | 2010-07-28 | Sergio Lara Pereira Monteiro | Method for transferring images with incoherent randomly arranged fiber optical bundle and for displaying images with randomly arranged pixels |
US8577085B2 (en) * | 2009-01-30 | 2013-11-05 | Microsoft Corporation | Visual target tracking |
US20100199231A1 (en) | 2009-01-30 | 2010-08-05 | Microsoft Corporation | Predictive determination |
US8294767B2 (en) | 2009-01-30 | 2012-10-23 | Microsoft Corporation | Body scan |
US7996793B2 (en) | 2009-01-30 | 2011-08-09 | Microsoft Corporation | Gesture recognizer system architecture |
US8487938B2 (en) * | 2009-01-30 | 2013-07-16 | Microsoft Corporation | Standard Gestures |
US8682028B2 (en) * | 2009-01-30 | 2014-03-25 | Microsoft Corporation | Visual target tracking |
US8588465B2 (en) | 2009-01-30 | 2013-11-19 | Microsoft Corporation | Visual target tracking |
US8577084B2 (en) * | 2009-01-30 | 2013-11-05 | Microsoft Corporation | Visual target tracking |
US20100199228A1 (en) * | 2009-01-30 | 2010-08-05 | Microsoft Corporation | Gesture Keyboarding |
US8267781B2 (en) | 2009-01-30 | 2012-09-18 | Microsoft Corporation | Visual target tracking |
US8565476B2 (en) * | 2009-01-30 | 2013-10-22 | Microsoft Corporation | Visual target tracking |
US8295546B2 (en) | 2009-01-30 | 2012-10-23 | Microsoft Corporation | Pose tracking pipeline |
US8565477B2 (en) * | 2009-01-30 | 2013-10-22 | Microsoft Corporation | Visual target tracking |
US8448094B2 (en) | 2009-01-30 | 2013-05-21 | Microsoft Corporation | Mapping a natural input device to a legacy system |
JP5271121B2 (en) | 2009-03-09 | 2013-08-21 | 任天堂株式会社 | Information processing program, information processing apparatus, information processing system, and information processing method |
US8773355B2 (en) | 2009-03-16 | 2014-07-08 | Microsoft Corporation | Adaptive cursor sizing |
US8988437B2 (en) * | 2009-03-20 | 2015-03-24 | Microsoft Technology Licensing, Llc | Chaining animations |
US9256282B2 (en) | 2009-03-20 | 2016-02-09 | Microsoft Technology Licensing, Llc | Virtual object manipulation |
US8251874B2 (en) | 2009-03-27 | 2012-08-28 | Icon Health & Fitness, Inc. | Exercise systems for simulating real world terrain |
JP5436909B2 (en) | 2009-03-30 | 2014-03-05 | 任天堂株式会社 | Information processing program, information processing apparatus, information processing system, and information processing method |
US9313376B1 (en) | 2009-04-01 | 2016-04-12 | Microsoft Technology Licensing, Llc | Dynamic depth power equalization |
US9898675B2 (en) * | 2009-05-01 | 2018-02-20 | Microsoft Technology Licensing, Llc | User movement tracking feedback to improve tracking |
US9498718B2 (en) | 2009-05-01 | 2016-11-22 | Microsoft Technology Licensing, Llc | Altering a view perspective within a display environment |
US9377857B2 (en) | 2009-05-01 | 2016-06-28 | Microsoft Technology Licensing, Llc | Show body position |
US8253746B2 (en) * | 2009-05-01 | 2012-08-28 | Microsoft Corporation | Determine intended motions |
US8660303B2 (en) * | 2009-05-01 | 2014-02-25 | Microsoft Corporation | Detection of body and props |
US8340432B2 (en) | 2009-05-01 | 2012-12-25 | Microsoft Corporation | Systems and methods for detecting a tilt angle from a depth image |
US20100277470A1 (en) * | 2009-05-01 | 2010-11-04 | Microsoft Corporation | Systems And Methods For Applying Model Tracking To Motion Capture |
US9015638B2 (en) | 2009-05-01 | 2015-04-21 | Microsoft Technology Licensing, Llc | Binding users to a gesture based system and providing feedback to the users |
US8649554B2 (en) * | 2009-05-01 | 2014-02-11 | Microsoft Corporation | Method to control perspective for a camera-controlled computer |
US8638985B2 (en) | 2009-05-01 | 2014-01-28 | Microsoft Corporation | Human body pose estimation |
US8181123B2 (en) | 2009-05-01 | 2012-05-15 | Microsoft Corporation | Managing virtual port associations to users in a gesture-based computing environment |
US8503720B2 (en) | 2009-05-01 | 2013-08-06 | Microsoft Corporation | Human body pose estimation |
US8942428B2 (en) * | 2009-05-01 | 2015-01-27 | Microsoft Corporation | Isolate extraneous motions |
US20100295771A1 (en) * | 2009-05-20 | 2010-11-25 | Microsoft Corporation | Control of display objects |
US20100302233A1 (en) * | 2009-05-26 | 2010-12-02 | Holland David Ames | Virtual Diving System and Method |
US8509479B2 (en) * | 2009-05-29 | 2013-08-13 | Microsoft Corporation | Virtual object |
US8320619B2 (en) | 2009-05-29 | 2012-11-27 | Microsoft Corporation | Systems and methods for tracking a model |
US9182814B2 (en) | 2009-05-29 | 2015-11-10 | Microsoft Technology Licensing, Llc | Systems and methods for estimating a non-visible or occluded body part |
US8693724B2 (en) | 2009-05-29 | 2014-04-08 | Microsoft Corporation | Method and system implementing user-centric gesture control |
US9383823B2 (en) | 2009-05-29 | 2016-07-05 | Microsoft Technology Licensing, Llc | Combining gestures beyond skeletal |
US8542252B2 (en) | 2009-05-29 | 2013-09-24 | Microsoft Corporation | Target digitization, extraction, and tracking |
US8625837B2 (en) * | 2009-05-29 | 2014-01-07 | Microsoft Corporation | Protocol and format for communicating an image from a camera to a computing environment |
US20100306716A1 (en) * | 2009-05-29 | 2010-12-02 | Microsoft Corporation | Extending standard gestures |
US8379101B2 (en) | 2009-05-29 | 2013-02-19 | Microsoft Corporation | Environment and/or target segmentation |
US8856691B2 (en) * | 2009-05-29 | 2014-10-07 | Microsoft Corporation | Gesture tool |
US20100302138A1 (en) * | 2009-05-29 | 2010-12-02 | Microsoft Corporation | Methods and systems for defining or modifying a visual representation |
US9400559B2 (en) * | 2009-05-29 | 2016-07-26 | Microsoft Technology Licensing, Llc | Gesture shortcuts |
US8744121B2 (en) | 2009-05-29 | 2014-06-03 | Microsoft Corporation | Device for identifying and tracking multiple humans over time |
US8418085B2 (en) | 2009-05-29 | 2013-04-09 | Microsoft Corporation | Gesture coach |
US8487871B2 (en) | 2009-06-01 | 2013-07-16 | Microsoft Corporation | Virtual desktop coordinate transformation |
US20100309197A1 (en) * | 2009-06-08 | 2010-12-09 | Nvidia Corporation | Interaction of stereoscopic objects with physical objects in viewing area |
KR20100138725A (en) * | 2009-06-25 | 2010-12-31 | 삼성전자주식회사 | Method and apparatus for processing virtual world |
KR100938922B1 (en) * | 2009-06-26 | 2010-01-27 | (주)이랜서 | Treadmill system and device method of the same |
US8390680B2 (en) | 2009-07-09 | 2013-03-05 | Microsoft Corporation | Visual representation expression based on player expression |
US9159151B2 (en) * | 2009-07-13 | 2015-10-13 | Microsoft Technology Licensing, Llc | Bringing a visual representation to life via learned input from the user |
US8264536B2 (en) * | 2009-08-25 | 2012-09-11 | Microsoft Corporation | Depth-sensitive imaging via polarization-state mapping |
US9141193B2 (en) | 2009-08-31 | 2015-09-22 | Microsoft Technology Licensing, Llc | Techniques for using human gestures to control gesture unaware programs |
US8330134B2 (en) | 2009-09-14 | 2012-12-11 | Microsoft Corporation | Optical fault monitoring |
US8508919B2 (en) * | 2009-09-14 | 2013-08-13 | Microsoft Corporation | Separation of electrical and optical components |
US8760571B2 (en) * | 2009-09-21 | 2014-06-24 | Microsoft Corporation | Alignment of lens and image sensor |
US8428340B2 (en) * | 2009-09-21 | 2013-04-23 | Microsoft Corporation | Screen space plane identification |
US8976986B2 (en) * | 2009-09-21 | 2015-03-10 | Microsoft Technology Licensing, Llc | Volume adjustment based on listener position |
US9014546B2 (en) | 2009-09-23 | 2015-04-21 | Rovi Guides, Inc. | Systems and methods for automatically detecting users within detection regions of media devices |
JP5161182B2 (en) | 2009-09-28 | 2013-03-13 | 任天堂株式会社 | Information processing program and information processing apparatus |
JP5610735B2 (en) | 2009-09-29 | 2014-10-22 | 任天堂株式会社 | Information processing program, information processing apparatus, information processing method, and information processing system |
JP5496591B2 (en) * | 2009-09-30 | 2014-05-21 | 任天堂株式会社 | Information processing program and information processing apparatus |
US8452087B2 (en) | 2009-09-30 | 2013-05-28 | Microsoft Corporation | Image selection techniques |
US8723118B2 (en) * | 2009-10-01 | 2014-05-13 | Microsoft Corporation | Imager for constructing color and depth images |
US20110083108A1 (en) * | 2009-10-05 | 2011-04-07 | Microsoft Corporation | Providing user interface feedback regarding cursor position on a display screen |
US7961910B2 (en) | 2009-10-07 | 2011-06-14 | Microsoft Corporation | Systems and methods for tracking a model |
US8963829B2 (en) | 2009-10-07 | 2015-02-24 | Microsoft Corporation | Methods and systems for determining and tracking extremities of a target |
US8564534B2 (en) | 2009-10-07 | 2013-10-22 | Microsoft Corporation | Human tracking system |
US8867820B2 (en) | 2009-10-07 | 2014-10-21 | Microsoft Corporation | Systems and methods for removing a background of an image |
US9400548B2 (en) * | 2009-10-19 | 2016-07-26 | Microsoft Technology Licensing, Llc | Gesture personalization and profile roaming |
US20110099476A1 (en) * | 2009-10-23 | 2011-04-28 | Microsoft Corporation | Decorating a display environment |
US8988432B2 (en) * | 2009-11-05 | 2015-03-24 | Microsoft Technology Licensing, Llc | Systems and methods for processing an image for target tracking |
US8843857B2 (en) | 2009-11-19 | 2014-09-23 | Microsoft Corporation | Distance scalable no touch computing |
US9244533B2 (en) * | 2009-12-17 | 2016-01-26 | Microsoft Technology Licensing, Llc | Camera navigation for presentations |
US20110150271A1 (en) | 2009-12-18 | 2011-06-23 | Microsoft Corporation | Motion detection using depth images |
US20110151974A1 (en) * | 2009-12-18 | 2011-06-23 | Microsoft Corporation | Gesture style recognition and reward |
US8320621B2 (en) | 2009-12-21 | 2012-11-27 | Microsoft Corporation | Depth projector system with integrated VCSEL array |
CN101780321B (en) * | 2009-12-30 | 2012-01-25 | 永春至善体育用品有限公司 | Method for making high-presence virtual reality of exercise fitness equipment, and interactive system and method based on virtual reality |
US9019201B2 (en) * | 2010-01-08 | 2015-04-28 | Microsoft Technology Licensing, Llc | Evolving universal gesture sets |
US8631355B2 (en) | 2010-01-08 | 2014-01-14 | Microsoft Corporation | Assigning gesture dictionaries |
US9268404B2 (en) * | 2010-01-08 | 2016-02-23 | Microsoft Technology Licensing, Llc | Application gesture interpretation |
US8933884B2 (en) * | 2010-01-15 | 2015-01-13 | Microsoft Corporation | Tracking groups of users in motion capture system |
US8334842B2 (en) | 2010-01-15 | 2012-12-18 | Microsoft Corporation | Recognizing user intent in motion capture system |
US8676581B2 (en) | 2010-01-22 | 2014-03-18 | Microsoft Corporation | Speech recognition analysis via identification information |
US8265341B2 (en) | 2010-01-25 | 2012-09-11 | Microsoft Corporation | Voice-body identity correlation |
US8864581B2 (en) * | 2010-01-29 | 2014-10-21 | Microsoft Corporation | Visual based identitiy tracking |
US8891067B2 (en) | 2010-02-01 | 2014-11-18 | Microsoft Corporation | Multiple synchronized optical sources for time-of-flight range finding systems |
US8687044B2 (en) * | 2010-02-02 | 2014-04-01 | Microsoft Corporation | Depth camera compatibility |
US8619122B2 (en) * | 2010-02-02 | 2013-12-31 | Microsoft Corporation | Depth camera compatibility |
US8717469B2 (en) * | 2010-02-03 | 2014-05-06 | Microsoft Corporation | Fast gating photosurface |
US8499257B2 (en) * | 2010-02-09 | 2013-07-30 | Microsoft Corporation | Handles interactions for human—computer interface |
US8659658B2 (en) * | 2010-02-09 | 2014-02-25 | Microsoft Corporation | Physical interaction zone for gesture-based user interfaces |
US8633890B2 (en) * | 2010-02-16 | 2014-01-21 | Microsoft Corporation | Gesture detection based on joint skipping |
US8928579B2 (en) * | 2010-02-22 | 2015-01-06 | Andrew David Wilson | Interacting with an omni-directionally projected display |
US8422769B2 (en) | 2010-03-05 | 2013-04-16 | Microsoft Corporation | Image segmentation using reduced foreground training data |
US8655069B2 (en) * | 2010-03-05 | 2014-02-18 | Microsoft Corporation | Updating image segmentation following user input |
US8411948B2 (en) | 2010-03-05 | 2013-04-02 | Microsoft Corporation | Up-sampling binary images for segmentation |
US20110223995A1 (en) | 2010-03-12 | 2011-09-15 | Kevin Geisner | Interacting with a computer based application |
US20110221755A1 (en) * | 2010-03-12 | 2011-09-15 | Kevin Geisner | Bionic motion |
US8279418B2 (en) | 2010-03-17 | 2012-10-02 | Microsoft Corporation | Raster scanning for depth detection |
US8213680B2 (en) * | 2010-03-19 | 2012-07-03 | Microsoft Corporation | Proxy training data for human body tracking |
US8514269B2 (en) * | 2010-03-26 | 2013-08-20 | Microsoft Corporation | De-aliasing depth images |
US20110234481A1 (en) * | 2010-03-26 | 2011-09-29 | Sagi Katz | Enhancing presentations using depth sensing cameras |
US8523667B2 (en) * | 2010-03-29 | 2013-09-03 | Microsoft Corporation | Parental control settings based on body dimensions |
US8605763B2 (en) | 2010-03-31 | 2013-12-10 | Microsoft Corporation | Temperature measurement and control for laser and light-emitting diodes |
US9098873B2 (en) | 2010-04-01 | 2015-08-04 | Microsoft Technology Licensing, Llc | Motion-based interactive shopping environment |
US9646340B2 (en) | 2010-04-01 | 2017-05-09 | Microsoft Technology Licensing, Llc | Avatar-based virtual dressing room |
US8351651B2 (en) | 2010-04-26 | 2013-01-08 | Microsoft Corporation | Hand-location post-process refinement in a tracking system |
US8379919B2 (en) | 2010-04-29 | 2013-02-19 | Microsoft Corporation | Multiple centroid condensation of probability distribution clouds |
US8284847B2 (en) | 2010-05-03 | 2012-10-09 | Microsoft Corporation | Detecting motion for a multifunction sensor device |
US8498481B2 (en) | 2010-05-07 | 2013-07-30 | Microsoft Corporation | Image segmentation using star-convexity constraints |
US8885890B2 (en) | 2010-05-07 | 2014-11-11 | Microsoft Corporation | Depth map confidence filtering |
US8457353B2 (en) | 2010-05-18 | 2013-06-04 | Microsoft Corporation | Gestures and gesture modifiers for manipulating a user-interface |
US8803888B2 (en) | 2010-06-02 | 2014-08-12 | Microsoft Corporation | Recognition system for sharing information |
US8751215B2 (en) | 2010-06-04 | 2014-06-10 | Microsoft Corporation | Machine based sign language interpreter |
US9008355B2 (en) | 2010-06-04 | 2015-04-14 | Microsoft Technology Licensing, Llc | Automatic depth camera aiming |
US9557574B2 (en) | 2010-06-08 | 2017-01-31 | Microsoft Technology Licensing, Llc | Depth illumination and detection optics |
US8330822B2 (en) | 2010-06-09 | 2012-12-11 | Microsoft Corporation | Thermally-tuned depth camera light source |
US9384329B2 (en) | 2010-06-11 | 2016-07-05 | Microsoft Technology Licensing, Llc | Caloric burn determination from body movement |
US8749557B2 (en) | 2010-06-11 | 2014-06-10 | Microsoft Corporation | Interacting with user interface via avatar |
US8675981B2 (en) | 2010-06-11 | 2014-03-18 | Microsoft Corporation | Multi-modal gender recognition including depth data |
US8982151B2 (en) | 2010-06-14 | 2015-03-17 | Microsoft Technology Licensing, Llc | Independently processing planes of display data |
US8670029B2 (en) | 2010-06-16 | 2014-03-11 | Microsoft Corporation | Depth camera illuminator with superluminescent light-emitting diode |
US8558873B2 (en) | 2010-06-16 | 2013-10-15 | Microsoft Corporation | Use of wavefront coding to create a depth image |
US8296151B2 (en) | 2010-06-18 | 2012-10-23 | Microsoft Corporation | Compound gesture-speech commands |
US8381108B2 (en) | 2010-06-21 | 2013-02-19 | Microsoft Corporation | Natural user input for driving interactive stories |
US8416187B2 (en) | 2010-06-22 | 2013-04-09 | Microsoft Corporation | Item navigation using motion-capture data |
TWI436646B (en) * | 2010-08-19 | 2014-05-01 | Univ Nat Taiwan Science Tech | Device with displayed image capable of moving based on body state of exerciser and performing method thereof |
US9075434B2 (en) | 2010-08-20 | 2015-07-07 | Microsoft Technology Licensing, Llc | Translating user motion into multiple object responses |
US8613666B2 (en) | 2010-08-31 | 2013-12-24 | Microsoft Corporation | User selection and navigation based on looped motions |
US20120058824A1 (en) | 2010-09-07 | 2012-03-08 | Microsoft Corporation | Scalable real-time motion recognition |
US8437506B2 (en) | 2010-09-07 | 2013-05-07 | Microsoft Corporation | System for fast, probabilistic skeletal tracking |
US8988508B2 (en) | 2010-09-24 | 2015-03-24 | Microsoft Technology Licensing, Llc. | Wide angle field of view active illumination imaging system |
US8681255B2 (en) | 2010-09-28 | 2014-03-25 | Microsoft Corporation | Integrated low power depth camera and projection device |
US8548270B2 (en) | 2010-10-04 | 2013-10-01 | Microsoft Corporation | Time-of-flight depth imaging |
US9484065B2 (en) | 2010-10-15 | 2016-11-01 | Microsoft Technology Licensing, Llc | Intelligent determination of replays based on event identification |
US8592739B2 (en) | 2010-11-02 | 2013-11-26 | Microsoft Corporation | Detection of configuration changes of an optical element in an illumination system |
US20120108909A1 (en) * | 2010-11-03 | 2012-05-03 | HeadRehab, LLC | Assessment and Rehabilitation of Cognitive and Motor Functions Using Virtual Reality |
US8866889B2 (en) | 2010-11-03 | 2014-10-21 | Microsoft Corporation | In-home depth camera calibration |
US8667519B2 (en) | 2010-11-12 | 2014-03-04 | Microsoft Corporation | Automatic passive and anonymous feedback system |
US10726861B2 (en) | 2010-11-15 | 2020-07-28 | Microsoft Technology Licensing, Llc | Semi-private communication in open environments |
US9349040B2 (en) | 2010-11-19 | 2016-05-24 | Microsoft Technology Licensing, Llc | Bi-modal depth-image analysis |
US10234545B2 (en) | 2010-12-01 | 2019-03-19 | Microsoft Technology Licensing, Llc | Light source module |
US9440134B2 (en) * | 2010-12-06 | 2016-09-13 | Full-Swing Golf, Inc. | Microsoft kinect |
US10086249B2 (en) | 2010-12-06 | 2018-10-02 | Full-Swing Golf, Inc. | Method and apparatus for simulated golf |
US8553934B2 (en) | 2010-12-08 | 2013-10-08 | Microsoft Corporation | Orienting the position of a sensor |
US8618405B2 (en) | 2010-12-09 | 2013-12-31 | Microsoft Corp. | Free-space gesture musical instrument digital interface (MIDI) controller |
US8408706B2 (en) | 2010-12-13 | 2013-04-02 | Microsoft Corporation | 3D gaze tracker |
US9171264B2 (en) | 2010-12-15 | 2015-10-27 | Microsoft Technology Licensing, Llc | Parallel processing machine learning decision tree training |
US8884968B2 (en) | 2010-12-15 | 2014-11-11 | Microsoft Corporation | Modeling an object from image data |
US8920241B2 (en) | 2010-12-15 | 2014-12-30 | Microsoft Corporation | Gesture controlled persistent handles for interface guides |
US8448056B2 (en) | 2010-12-17 | 2013-05-21 | Microsoft Corporation | Validation analysis of human target |
US8803952B2 (en) | 2010-12-20 | 2014-08-12 | Microsoft Corporation | Plural detector time-of-flight depth mapping |
US9821224B2 (en) | 2010-12-21 | 2017-11-21 | Microsoft Technology Licensing, Llc | Driving simulator control with virtual skeleton |
US9848106B2 (en) | 2010-12-21 | 2017-12-19 | Microsoft Technology Licensing, Llc | Intelligent gameplay photo capture |
US9823339B2 (en) | 2010-12-21 | 2017-11-21 | Microsoft Technology Licensing, Llc | Plural anode time-of-flight sensor |
US8385596B2 (en) | 2010-12-21 | 2013-02-26 | Microsoft Corporation | First person shooter control with virtual skeleton |
US8994718B2 (en) | 2010-12-21 | 2015-03-31 | Microsoft Technology Licensing, Llc | Skeletal control of three-dimensional virtual world |
US9123316B2 (en) | 2010-12-27 | 2015-09-01 | Microsoft Technology Licensing, Llc | Interactive content creation |
US8488888B2 (en) | 2010-12-28 | 2013-07-16 | Microsoft Corporation | Classification of posture states |
US8401242B2 (en) | 2011-01-31 | 2013-03-19 | Microsoft Corporation | Real-time camera tracking using depth maps |
US8587583B2 (en) | 2011-01-31 | 2013-11-19 | Microsoft Corporation | Three-dimensional environment reconstruction |
US9247238B2 (en) | 2011-01-31 | 2016-01-26 | Microsoft Technology Licensing, Llc | Reducing interference between multiple infra-red depth cameras |
US8401225B2 (en) | 2011-01-31 | 2013-03-19 | Microsoft Corporation | Moving object segmentation using depth images |
US8724887B2 (en) | 2011-02-03 | 2014-05-13 | Microsoft Corporation | Environmental modifications to mitigate environmental factors |
CN106964150B (en) * | 2011-02-11 | 2021-03-02 | 漳州市爵晟电子科技有限公司 | Action positioning point control system and sleeve type positioning point control equipment thereof |
US8942917B2 (en) | 2011-02-14 | 2015-01-27 | Microsoft Corporation | Change invariant scene recognition by an agent |
US8497838B2 (en) | 2011-02-16 | 2013-07-30 | Microsoft Corporation | Push actuation of interface controls |
US9551914B2 (en) | 2011-03-07 | 2017-01-24 | Microsoft Technology Licensing, Llc | Illuminator with refractive optical element |
US9067136B2 (en) | 2011-03-10 | 2015-06-30 | Microsoft Technology Licensing, Llc | Push personalization of interface controls |
US8571263B2 (en) | 2011-03-17 | 2013-10-29 | Microsoft Corporation | Predicting joint positions |
US9470778B2 (en) | 2011-03-29 | 2016-10-18 | Microsoft Technology Licensing, Llc | Learning from high quality depth measurements |
US9298287B2 (en) | 2011-03-31 | 2016-03-29 | Microsoft Technology Licensing, Llc | Combined activation for natural user interface systems |
US9842168B2 (en) | 2011-03-31 | 2017-12-12 | Microsoft Technology Licensing, Llc | Task driven user intents |
US10642934B2 (en) | 2011-03-31 | 2020-05-05 | Microsoft Technology Licensing, Llc | Augmented conversational understanding architecture |
US9760566B2 (en) | 2011-03-31 | 2017-09-12 | Microsoft Technology Licensing, Llc | Augmented conversational understanding agent to identify conversation context between two humans and taking an agent action thereof |
US8824749B2 (en) | 2011-04-05 | 2014-09-02 | Microsoft Corporation | Biometric recognition |
US8503494B2 (en) | 2011-04-05 | 2013-08-06 | Microsoft Corporation | Thermal management system |
US8620113B2 (en) | 2011-04-25 | 2013-12-31 | Microsoft Corporation | Laser diode modes |
US9259643B2 (en) | 2011-04-28 | 2016-02-16 | Microsoft Technology Licensing, Llc | Control of separate computer game elements |
US8702507B2 (en) | 2011-04-28 | 2014-04-22 | Microsoft Corporation | Manual and camera-based avatar control |
US10671841B2 (en) | 2011-05-02 | 2020-06-02 | Microsoft Technology Licensing, Llc | Attribute state classification |
US8888331B2 (en) | 2011-05-09 | 2014-11-18 | Microsoft Corporation | Low inductance light source module |
US9064006B2 (en) | 2012-08-23 | 2015-06-23 | Microsoft Technology Licensing, Llc | Translating natural language utterances to keyword search queries |
US9137463B2 (en) | 2011-05-12 | 2015-09-15 | Microsoft Technology Licensing, Llc | Adaptive high dynamic range camera |
US8788973B2 (en) | 2011-05-23 | 2014-07-22 | Microsoft Corporation | Three-dimensional gesture controlled avatar configuration interface |
US8760395B2 (en) | 2011-05-31 | 2014-06-24 | Microsoft Corporation | Gesture recognition techniques |
US8526734B2 (en) | 2011-06-01 | 2013-09-03 | Microsoft Corporation | Three-dimensional background removal for vision system |
US9594430B2 (en) | 2011-06-01 | 2017-03-14 | Microsoft Technology Licensing, Llc | Three-dimensional foreground selection for vision system |
US9098110B2 (en) | 2011-06-06 | 2015-08-04 | Microsoft Technology Licensing, Llc | Head rotation tracking from depth-based center of mass |
US9724600B2 (en) | 2011-06-06 | 2017-08-08 | Microsoft Technology Licensing, Llc | Controlling objects in a virtual environment |
US8929612B2 (en) | 2011-06-06 | 2015-01-06 | Microsoft Corporation | System for recognizing an open or closed hand |
US8897491B2 (en) | 2011-06-06 | 2014-11-25 | Microsoft Corporation | System for finger recognition and tracking |
US9013489B2 (en) | 2011-06-06 | 2015-04-21 | Microsoft Technology Licensing, Llc | Generation of avatar reflecting player appearance |
US8597142B2 (en) | 2011-06-06 | 2013-12-03 | Microsoft Corporation | Dynamic camera based practice mode |
US9208571B2 (en) | 2011-06-06 | 2015-12-08 | Microsoft Technology Licensing, Llc | Object digitization |
US10796494B2 (en) | 2011-06-06 | 2020-10-06 | Microsoft Technology Licensing, Llc | Adding attributes to virtual representations of real-world objects |
US9597587B2 (en) | 2011-06-08 | 2017-03-21 | Microsoft Technology Licensing, Llc | Locational node device |
US8786730B2 (en) | 2011-08-18 | 2014-07-22 | Microsoft Corporation | Image exposure using exclusion regions |
US9557836B2 (en) | 2011-11-01 | 2017-01-31 | Microsoft Technology Licensing, Llc | Depth image compression |
US9117281B2 (en) | 2011-11-02 | 2015-08-25 | Microsoft Corporation | Surface segmentation from RGB and depth images |
US9582178B2 (en) | 2011-11-07 | 2017-02-28 | Immersion Corporation | Systems and methods for multi-pressure interaction on touch-sensitive surfaces |
US8854426B2 (en) | 2011-11-07 | 2014-10-07 | Microsoft Corporation | Time-of-flight camera with guided light |
CN103108197A (en) | 2011-11-14 | 2013-05-15 | 辉达公司 | Priority level compression method and priority level compression system for three-dimensional (3D) video wireless display |
US8724906B2 (en) | 2011-11-18 | 2014-05-13 | Microsoft Corporation | Computing pose and/or shape of modifiable entities |
US8509545B2 (en) | 2011-11-29 | 2013-08-13 | Microsoft Corporation | Foreground subject detection |
US20130137076A1 (en) * | 2011-11-30 | 2013-05-30 | Kathryn Stone Perez | Head-mounted display based education and instruction |
US8635637B2 (en) | 2011-12-02 | 2014-01-21 | Microsoft Corporation | User interface presenting an animated avatar performing a media reaction |
US8803800B2 (en) | 2011-12-02 | 2014-08-12 | Microsoft Corporation | User interface control based on head orientation |
US20130162536A1 (en) * | 2011-12-05 | 2013-06-27 | Vaibhav Narayan Khedkar | Systems and methods for enabling or accessing or viewing components, involving an input assembly and a screen |
US9100685B2 (en) | 2011-12-09 | 2015-08-04 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US8879831B2 (en) | 2011-12-15 | 2014-11-04 | Microsoft Corporation | Using high-level attributes to guide image processing |
US8630457B2 (en) | 2011-12-15 | 2014-01-14 | Microsoft Corporation | Problem states for pose tracking pipeline |
US8971612B2 (en) | 2011-12-15 | 2015-03-03 | Microsoft Corporation | Learning image processing tasks from scene reconstructions |
US8811938B2 (en) | 2011-12-16 | 2014-08-19 | Microsoft Corporation | Providing a user interface experience based on inferred vehicle state |
US9342139B2 (en) | 2011-12-19 | 2016-05-17 | Microsoft Technology Licensing, Llc | Pairing a computing device to a user |
US9168418B2 (en) * | 2011-12-30 | 2015-10-27 | Lawrence G. Adamchick | Portable physical therapy/rehabilitation/exercise device, system and method |
US9339691B2 (en) | 2012-01-05 | 2016-05-17 | Icon Health & Fitness, Inc. | System and method for controlling an exercise device |
US9720089B2 (en) | 2012-01-23 | 2017-08-01 | Microsoft Technology Licensing, Llc | 3D zoom imager |
US9829715B2 (en) | 2012-01-23 | 2017-11-28 | Nvidia Corporation | Eyewear device for transmitting signal and communication method thereof |
US9375629B2 (en) * | 2012-02-13 | 2016-06-28 | Gusto Technologies, Inc. | Method and apparatus for visual simulation of exercise |
US8898687B2 (en) | 2012-04-04 | 2014-11-25 | Microsoft Corporation | Controlling a media program based on a media reaction |
US9210401B2 (en) | 2012-05-03 | 2015-12-08 | Microsoft Technology Licensing, Llc | Projected visual cues for guiding physical movement |
CA2775700C (en) | 2012-05-04 | 2013-07-23 | Microsoft Corporation | Determining a future portion of a currently presented media program |
JP6018707B2 (en) | 2012-06-21 | 2016-11-02 | マイクロソフト コーポレーション | Building an avatar using a depth camera |
US9836590B2 (en) | 2012-06-22 | 2017-12-05 | Microsoft Technology Licensing, Llc | Enhanced accuracy of user presence status determination |
US8779908B2 (en) * | 2012-07-16 | 2014-07-15 | Shmuel Ur | System and method for social dancing |
US9696427B2 (en) | 2012-08-14 | 2017-07-04 | Microsoft Technology Licensing, Llc | Wide angle depth detection |
US20150255002A1 (en) * | 2012-08-28 | 2015-09-10 | HAI Logan Gym, LLC | Personal training system and related exercise facility and method |
US9318030B2 (en) | 2013-08-28 | 2016-04-19 | HAI Logan Gym, LLC | Personal training system and method |
US20140363800A1 (en) * | 2012-08-28 | 2014-12-11 | HAI Logan Gym, LLC | Personal training system and method |
BR112015004174A2 (en) | 2012-08-28 | 2017-12-05 | Hai Logan Gym Llc | personnel training system, exercise facility, and process for generating and conducting a custom exercise regime |
US11191996B2 (en) | 2012-08-31 | 2021-12-07 | Blue Goji Llc | Body joystick for interacting with virtual reality or mixed reality machines or software applications |
US9849333B2 (en) | 2012-08-31 | 2017-12-26 | Blue Goji Llc | Variable-resistance exercise machine with wireless communication for smart device control and virtual reality applications |
US11612786B2 (en) * | 2012-08-31 | 2023-03-28 | Blue Goji Llc | System and method for targeted neurological therapy using brainwave entrainment with passive treatment |
US9578224B2 (en) | 2012-09-10 | 2017-02-21 | Nvidia Corporation | System and method for enhanced monoimaging |
US8882310B2 (en) | 2012-12-10 | 2014-11-11 | Microsoft Corporation | Laser die light source module with low inductance |
US9857470B2 (en) | 2012-12-28 | 2018-01-02 | Microsoft Technology Licensing, Llc | Using photometric stereo for 3D environment modeling |
GB2509551A (en) * | 2013-01-08 | 2014-07-09 | Sony Comp Entertainment Europe | Detecting potential cable tangling or wrapping. |
US11052288B1 (en) | 2013-01-19 | 2021-07-06 | Bertec Corporation | Force measurement system |
US9251590B2 (en) | 2013-01-24 | 2016-02-02 | Microsoft Technology Licensing, Llc | Camera pose estimation for 3D reconstruction |
US9052746B2 (en) | 2013-02-15 | 2015-06-09 | Microsoft Technology Licensing, Llc | User center-of-mass and mass distribution extraction using depth images |
US9940553B2 (en) | 2013-02-22 | 2018-04-10 | Microsoft Technology Licensing, Llc | Camera/object pose from predicted coordinates |
US9135516B2 (en) | 2013-03-08 | 2015-09-15 | Microsoft Technology Licensing, Llc | User body angle, curvature and average extremity positions extraction using depth images |
US9092657B2 (en) | 2013-03-13 | 2015-07-28 | Microsoft Technology Licensing, Llc | Depth image processing |
US9274606B2 (en) | 2013-03-14 | 2016-03-01 | Microsoft Technology Licensing, Llc | NUI video conference controls |
EP2969058B1 (en) | 2013-03-14 | 2020-05-13 | Icon Health & Fitness, Inc. | Strength training apparatus with flywheel and related methods |
US9953213B2 (en) | 2013-03-27 | 2018-04-24 | Microsoft Technology Licensing, Llc | Self discovery of autonomous NUI devices |
JP6138566B2 (en) * | 2013-04-24 | 2017-05-31 | 川崎重工業株式会社 | Component mounting work support system and component mounting method |
US9442186B2 (en) | 2013-05-13 | 2016-09-13 | Microsoft Technology Licensing, Llc | Interference reduction for TOF systems |
US20160232809A1 (en) * | 2013-08-28 | 2016-08-11 | HAI Logan Gym, LLC | Personal training system and related exercise facility and method |
US9462253B2 (en) | 2013-09-23 | 2016-10-04 | Microsoft Technology Licensing, Llc | Optical modules that reduce speckle contrast and diffraction artifacts |
US9443310B2 (en) | 2013-10-09 | 2016-09-13 | Microsoft Technology Licensing, Llc | Illumination modules that emit structured light |
US9674563B2 (en) | 2013-11-04 | 2017-06-06 | Rovi Guides, Inc. | Systems and methods for recommending content |
US9769459B2 (en) | 2013-11-12 | 2017-09-19 | Microsoft Technology Licensing, Llc | Power efficient laser diode driver circuit and method |
US9508385B2 (en) | 2013-11-21 | 2016-11-29 | Microsoft Technology Licensing, Llc | Audio-visual project generator |
KR101563736B1 (en) * | 2013-12-24 | 2015-11-06 | 전자부품연구원 | Apparatus and Method for Mapping Position Information to Virtual Resources |
EP3974036A1 (en) | 2013-12-26 | 2022-03-30 | iFIT Inc. | Magnetic resistance mechanism in a cable machine |
US9971491B2 (en) | 2014-01-09 | 2018-05-15 | Microsoft Technology Licensing, Llc | Gesture library for natural user input |
US10935788B2 (en) | 2014-01-24 | 2021-03-02 | Nvidia Corporation | Hybrid virtual 3D rendering approach to stereovision |
WO2015138339A1 (en) | 2014-03-10 | 2015-09-17 | Icon Health & Fitness, Inc. | Pressure sensor to quantify work |
US9821465B2 (en) * | 2014-05-06 | 2017-11-21 | Kenneth Dean Stephens, Jr. | Enhanced environment simulator for proxy robot handlers |
WO2015191445A1 (en) | 2014-06-09 | 2015-12-17 | Icon Health & Fitness, Inc. | Cable system incorporated into a treadmill |
JP6344084B2 (en) * | 2014-06-20 | 2018-06-20 | オムロン株式会社 | Worker terminal for robot operation |
WO2015195965A1 (en) | 2014-06-20 | 2015-12-23 | Icon Health & Fitness, Inc. | Post workout massage device |
US9746984B2 (en) * | 2014-08-19 | 2017-08-29 | Sony Interactive Entertainment Inc. | Systems and methods for providing feedback to a user while interacting with content |
US10391361B2 (en) | 2015-02-27 | 2019-08-27 | Icon Health & Fitness, Inc. | Simulating real-world terrain on an exercise device |
KR101850028B1 (en) * | 2015-03-06 | 2018-05-30 | 한국전자통신연구원 | Device for virtual tour experience and method thereof |
US10379604B2 (en) | 2015-04-10 | 2019-08-13 | Virzoom, Inc. | Virtual reality exercise game |
US20160346597A1 (en) * | 2015-05-12 | 2016-12-01 | Sean O'Mara | In-floor treadmill assembly |
US10880346B2 (en) * | 2015-05-27 | 2020-12-29 | Google Llc | Streaming spherical video |
US20170050102A1 (en) * | 2015-08-22 | 2017-02-23 | FissionFit, LLC | Modular fitness activity facility and method for providing same |
US11301045B1 (en) * | 2015-08-22 | 2022-04-12 | Bertec Corporation | Measurement system that includes at least one measurement assembly, a visual display device, and at least one data processing device |
US10341339B2 (en) | 2015-10-21 | 2019-07-02 | Harman International Industries, Incorporated | Techniques for hearable authentication |
US10212994B2 (en) | 2015-11-02 | 2019-02-26 | Icon Health & Fitness, Inc. | Smart watch band |
US10948514B2 (en) * | 2015-11-30 | 2021-03-16 | Harman International Industries, Incorporated | Center of gravity shifting force device |
FR3046855B1 (en) * | 2016-01-18 | 2018-07-06 | Vincent Brom | PILOTAGE EQUIPMENT BY TELEGUIDAGE OF A VEHICLE AND STEERING METHOD IMPLEMENTED BY SAID EQUIPMENT |
US10412280B2 (en) | 2016-02-10 | 2019-09-10 | Microsoft Technology Licensing, Llc | Camera with light valve over sensor array |
US10257932B2 (en) | 2016-02-16 | 2019-04-09 | Microsoft Technology Licensing, Llc. | Laser diode chip on printed circuit board |
US9906981B2 (en) | 2016-02-25 | 2018-02-27 | Nvidia Corporation | Method and system for dynamic regulation and control of Wi-Fi scans |
CN105786184A (en) * | 2016-03-02 | 2016-07-20 | 广州聚变网络科技有限公司 | Virtual reality system |
RU2617972C1 (en) * | 2016-03-09 | 2017-04-28 | Публичное акционерное общество "Межрегиональная распределительная сетевая компания Центра" | Simulator for operational and maintenance staff on the basis of virtual reality models of transformer substation |
US10462452B2 (en) | 2016-03-16 | 2019-10-29 | Microsoft Technology Licensing, Llc | Synchronizing active illumination cameras |
US10625137B2 (en) | 2016-03-18 | 2020-04-21 | Icon Health & Fitness, Inc. | Coordinated displays in an exercise device |
US10493349B2 (en) | 2016-03-18 | 2019-12-03 | Icon Health & Fitness, Inc. | Display on exercise device |
ES2574621B1 (en) * | 2016-03-18 | 2017-03-28 | Juan Carlos CANEDA GONZÁLEZ | Sports machine for neck and cervical rehabilitation |
US10272317B2 (en) | 2016-03-18 | 2019-04-30 | Icon Health & Fitness, Inc. | Lighted pace feature in a treadmill |
CN109414612B (en) * | 2016-04-19 | 2022-07-12 | S·萨米特 | Virtual reality haptic systems and devices |
WO2018195344A1 (en) * | 2017-04-19 | 2018-10-25 | Scott Summit | Virtual reality haptic system and apparatus |
JP7140465B2 (en) * | 2016-06-10 | 2022-09-21 | 任天堂株式会社 | Game program, information processing device, information processing system, game processing method |
US20180021627A1 (en) * | 2016-07-20 | 2018-01-25 | Strive VR, LLC | Interactive and Dynamic Fitness System |
US10198855B2 (en) * | 2016-07-20 | 2019-02-05 | Colopl, Inc. | Method of providing virtual space, method of providing virtual experience, system and medium for implementing the methods |
US10671705B2 (en) | 2016-09-28 | 2020-06-02 | Icon Health & Fitness, Inc. | Customizing recipe recommendations |
CN106485973A (en) * | 2016-10-21 | 2017-03-08 | 上海申电教育培训有限公司 | Electric power skills training platform based on virtual reality technology |
US9983684B2 (en) | 2016-11-02 | 2018-05-29 | Microsoft Technology Licensing, Llc | Virtual affordance display at virtual target |
TWI648081B (en) | 2016-12-05 | 2019-01-21 | 美商愛康運動與健康公司 | Pull rope resistance mechanism in treadmill |
IT201600124198A1 (en) * | 2016-12-07 | 2018-06-07 | Michele Presicci | "INTERACTIVE SYSTEM TO BE APPLIED TO GYMNASTER EQUIPMENT, BASED ON THE INCREASED REALITY" |
US11712601B2 (en) | 2016-12-09 | 2023-08-01 | MILLZ, Inc. | Exercise device |
US10780358B1 (en) * | 2017-03-22 | 2020-09-22 | Intuitive Research And Technology Corporation | Virtual reality arena system |
US10319109B2 (en) * | 2017-03-31 | 2019-06-11 | Honda Motor Co., Ltd. | Interaction with physical objects as proxy objects representing virtual objects |
US10222860B2 (en) * | 2017-04-14 | 2019-03-05 | International Business Machines Corporation | Enhanced virtual scenarios for safety concerns |
US10719125B2 (en) * | 2017-05-09 | 2020-07-21 | Microsoft Technology Licensing, Llc | Object and environment tracking via shared sensor |
WO2019013808A1 (en) * | 2017-07-14 | 2019-01-17 | Hewlett-Packard Development Company, L.P. | Virtual reality headset stands |
US10070123B1 (en) * | 2017-08-14 | 2018-09-04 | Oculus Vr, Llc | Apparatuses, systems, and methods for characterizing and calibrating displays |
WO2019046942A1 (en) * | 2017-09-05 | 2019-03-14 | Technologies Aperium Inc. | Navigating in a virtual environment |
US11058960B2 (en) * | 2017-11-17 | 2021-07-13 | Ifly Holdings, Llc | Interactive modular sensor system for indoor skydiving wind tunnels |
US11903712B2 (en) * | 2018-06-08 | 2024-02-20 | International Business Machines Corporation | Physiological stress of a user of a virtual reality environment |
US11311778B2 (en) | 2018-08-07 | 2022-04-26 | Interactive Strength, Inc. | Interactive exercise machine support and mounting system |
CN109272801A (en) * | 2018-11-20 | 2019-01-25 | 深圳供电局有限公司 | A kind of power distribution network operation simulation wire examination method and system based on VR technology |
RU2715148C1 (en) * | 2019-07-24 | 2020-02-25 | Акционерное общество «Центральная пригородная пассажирская компания» | Switching vr-simulator and method of training and assessing competency for cashiers-controllers using it |
US11298578B2 (en) | 2020-01-31 | 2022-04-12 | Interactive Strength, Inc. | Positionable arm with quick release for an interactive exercise machine |
CN111632370B (en) * | 2020-05-20 | 2024-02-27 | 东北石油大学 | Hand resistance processing method and device in virtual reality |
WO2022221173A1 (en) * | 2021-04-11 | 2022-10-20 | Khurana Vikas | Device and process for trapping and compressing organisms |
US20230398426A1 (en) * | 2022-06-10 | 2023-12-14 | Mario McKay | Integrated physical therapy and virtual reality display system and methods of use |
US11794053B1 (en) | 2023-03-14 | 2023-10-24 | Stephen H. Woodyard, III | System and method for full body isometric machine |
US11911651B1 (en) | 2023-08-10 | 2024-02-27 | Barron Associates, Inc. | System, device and method for electronically mediated upper extremity therapy |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4735410A (en) * | 1986-08-13 | 1988-04-05 | Mizuno Corporation | Rowing machine |
US5185561A (en) * | 1991-07-23 | 1993-02-09 | Digital Equipment Corporation | Torque motor as a tactile feedback device in a computer system |
US5213555A (en) * | 1990-02-27 | 1993-05-25 | Hood Robert L | Exercise equipment information, communication and display system |
US5308296A (en) * | 1992-07-16 | 1994-05-03 | Donald Eckstein | Interactive video and exercise apparatus |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4408613A (en) * | 1981-10-02 | 1983-10-11 | Aerobitronics, Inc. | Interactive exercise device |
US4489938A (en) * | 1983-06-20 | 1984-12-25 | Darzinskis Kazimir R | Video computer system controlled by operation of elastomeric exercise apparatus |
DE3404539A1 (en) * | 1984-02-09 | 1985-08-22 | Frans van der 6450 Hanau Stel | Training appliance for stationary physical training |
US4558864A (en) * | 1984-06-13 | 1985-12-17 | Medwedeff Marion C | Handgrip exercising, computer game controller |
US4934694A (en) * | 1985-12-06 | 1990-06-19 | Mcintosh James L | Computer controlled exercise system |
DE3784903T2 (en) * | 1986-12-18 | 1993-06-24 | Michael Anthony Smithard | LEARNING DEVICE. |
US4919418A (en) * | 1988-01-27 | 1990-04-24 | Miller Jan W | Computerized drive mechanism for exercise, physical therapy and rehabilitation |
US5229756A (en) * | 1989-02-07 | 1993-07-20 | Yamaha Corporation | Image control apparatus |
US5184319A (en) * | 1990-02-02 | 1993-02-02 | Kramer James F | Force feedback and textures simulating interface device |
US5227985A (en) * | 1991-08-19 | 1993-07-13 | University Of Maryland | Computer vision system for position monitoring in three dimensions using non-coplanar light sources attached to a monitored object |
US5320538A (en) * | 1992-09-23 | 1994-06-14 | Hughes Training, Inc. | Interactive aircraft training system and method |
US5362069A (en) * | 1992-12-03 | 1994-11-08 | Heartbeat Corporation | Combination exercise device/video game |
-
1995
- 1995-08-04 US US08/513,671 patent/US5577981A/en not_active Expired - Lifetime
-
1996
- 1996-07-29 WO PCT/US1996/012369 patent/WO1997005926A1/en active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4735410A (en) * | 1986-08-13 | 1988-04-05 | Mizuno Corporation | Rowing machine |
US5213555A (en) * | 1990-02-27 | 1993-05-25 | Hood Robert L | Exercise equipment information, communication and display system |
US5185561A (en) * | 1991-07-23 | 1993-02-09 | Digital Equipment Corporation | Torque motor as a tactile feedback device in a computer system |
US5308296A (en) * | 1992-07-16 | 1994-05-03 | Donald Eckstein | Interactive video and exercise apparatus |
Non-Patent Citations (1)
Title |
---|
FISHER, S.S. et al., "Virtual Environment Display System", ACM 1986 Workshop on Interactive 3D Graphics, October 23 and 24, 11 pages. * |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7497807B2 (en) | 2003-07-15 | 2009-03-03 | Cube X Incorporated | Interactive computer simulation enhanced exercise machine |
US7335134B1 (en) | 2004-02-23 | 2008-02-26 | Lavelle Richard | Exercise and game controller apparatus and method |
US8620146B1 (en) | 2008-03-28 | 2013-12-31 | Theresa Coleman | Picture-in-picture video system for virtual exercise, instruction and entertainment |
WO2011025466A1 (en) * | 2009-08-25 | 2011-03-03 | Sefik Isik | SINGLE OR MULTIPLE SPORTS UNIT WITH 2D or 3D DISPLAY (FILM), SOUND AND AIR SUPPORT |
WO2013034981A3 (en) * | 2011-09-08 | 2013-06-06 | Offshore Incorporations (Cayman) Limited, | System and method for visualizing synthetic objects withinreal-world video clip |
US9586141B2 (en) | 2011-09-08 | 2017-03-07 | Paofit Holdings Pte. Ltd. | System and method for visualizing synthetic objects within real-world video clip |
US10828570B2 (en) | 2011-09-08 | 2020-11-10 | Nautilus, Inc. | System and method for visualizing synthetic objects within real-world video clip |
CN104069633A (en) * | 2014-07-11 | 2014-10-01 | 上海体育学院 | Virtual reality generation system for table tennis ball serving |
US10810798B2 (en) | 2015-06-23 | 2020-10-20 | Nautilus, Inc. | Systems and methods for generating 360 degree mixed reality environments |
CN106139519A (en) * | 2016-07-29 | 2016-11-23 | 韩莹光 | A kind of universal treadmill of mixed reality and application process thereof |
US10088911B2 (en) | 2016-12-30 | 2018-10-02 | Manuel Saez | Programmable electronic helmet |
Also Published As
Publication number | Publication date |
---|---|
US5577981A (en) | 1996-11-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5577981A (en) | Virtual reality exercise machine and computer controlled video system | |
US20070171199A1 (en) | Locomotion simulation apparatus, system and method | |
US5388990A (en) | Virtual reality flight control display with six-degree-of-freedom controller and spherical orientation overlay | |
RU2109336C1 (en) | Method and device for immersing user into virtual world | |
US5436638A (en) | Image display method and apparatus with means for yoking viewpoint orienting muscles of a user | |
KR101914140B1 (en) | Virtual reality experience apparatus | |
US5913727A (en) | Interactive movement and contact simulation game | |
US10004940B2 (en) | Exercising bicycle | |
US8892219B2 (en) | Motivation and enhancement of physical and mental exercise, rehabilitation, health and social interaction | |
US6741911B2 (en) | Natural robot control | |
JP4871270B2 (en) | System and method for operating in a virtual three-dimensional space and system for selecting operations via a visualization system | |
KR102161646B1 (en) | System and method for interworking virtual reality and indoor exercise machine | |
JP3862348B2 (en) | Motion capture system | |
WO2020218368A1 (en) | Exercise equipment | |
JPH1118025A (en) | Image display device | |
CN105664469B (en) | Kinematic system and its implementation with simulated scenario | |
WO2006017926A1 (en) | Locomotion simulation system and method | |
WO2018195344A1 (en) | Virtual reality haptic system and apparatus | |
RU2109337C1 (en) | Method and device for inserting user into virtual world | |
Iwata | Haptic interfaces | |
TW201733639A (en) | System of exercise healthy machine with virtual reality transmitting by cloud server | |
Bouguila et al. | Realizing a new step-in-place locomotion interface for virtual environment with large display system | |
WO2011119118A1 (en) | A haptic system, a method of forming a haptic system and a method of controlling a haptic system | |
US11887259B2 (en) | Method, system, and apparatus for full-body tracking with magnetic fields in virtual reality and augmented reality applications | |
JP3190026B1 (en) | Humanoid robot experience presentation device and master-slave control device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): CA JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
NENP | Non-entry into the national phase |
Ref country code: CA |
|
122 | Ep: pct application non-entry in european phase |