US6795041B2 - Mixed reality realizing system - Google Patents

Mixed reality realizing system Download PDF

Info

Publication number
US6795041B2
US6795041B2 US10/189,127 US18912702A US6795041B2 US 6795041 B2 US6795041 B2 US 6795041B2 US 18912702 A US18912702 A US 18912702A US 6795041 B2 US6795041 B2 US 6795041B2
Authority
US
United States
Prior art keywords
target object
exhibit
coordinate
display
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime, expires
Application number
US10/189,127
Other versions
US20040004583A1 (en
Inventor
Kenji Ogawa
Toshiro Saeki
Haruyo Tamura
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Hitachi Zosen Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to JP2001097782A priority Critical patent/JP3957468B2/en
Application filed by Hitachi Zosen Corp filed Critical Hitachi Zosen Corp
Priority to US10/189,127 priority patent/US6795041B2/en
Assigned to HITACHI ZOSEN CORPORATION reassignment HITACHI ZOSEN CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OGAWA, KENJI, SAEKI, TOSHIRO, TAMURA, HARUYO
Publication of US20040004583A1 publication Critical patent/US20040004583A1/en
Application granted granted Critical
Publication of US6795041B2 publication Critical patent/US6795041B2/en
Assigned to MICROSOFT TECHNOLOGY LICENSING, LLC reassignment MICROSOFT TECHNOLOGY LICENSING, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MICROSOFT CORPORATION
Assigned to MICROSOFT CORPORATION reassignment MICROSOFT CORPORATION MERGER (SEE DOCUMENT FOR DETAILS). Assignors: HITACHI ZOSEN CORP
Adjusted expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0138Head-up displays characterised by optical features comprising image capture systems, e.g. camera
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • G02B2027/0187Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye

Definitions

  • the present invention relates to a mixed reality realizing system for use in exhibition facilities like a museum for example, which provides viewers with information on exhibits together with the exhibits they are viewing, by integrating a virtual space with a real space.
  • the electronic guidance system using the PDA is capable of providing viewers with interactive information corresponding to their interests by combining voices and images. Therefore, this system is considered to be the best one in the present circumstances.
  • the present invention has an object to provide a mixed reality realizing system capable of producing a display effect of giving reality sense and immersion sense.
  • the mixed reality realizing system of the present invention comprises identification signal transmitters located around a target object to transmit identification signals of the target object, a plurality of position indicators for indicating a position of the target object, and a portable type terminal device comprising a head mount display (HMD), a controller and a computer,
  • HMD head mount display
  • the head mount display of the portable type terminal device comprises: a video display; an identification signal receiver for receiving the identification signals of the target object transmitted from the identification signal transmitters; and a camera for shooting the plurality of position indicators
  • the computer comprises: a storage for storing, in advance, data including coordinate positions of the plurality of position indicators on an image shot by the camera from a reference position previously determined based on the target object and contents relating to the target object; a control operation unit for receiving the signals from the identification signal receiver and retrieving the contents relating to the target object from the storage; a coordinate position recognizing unit for recognizing, in real time, coordinate positions of the plurality of position indicators on the image shot by the camera; a position-information operation unit for calculating a visible position and a real position of the target object based on the plurality of coordinate positions obtained from the position indicators; and a content superimposing unit for superimposing the contents relating to the target object on the target object on the video display of the head mount display based on the information on the plurality of positions obtained from the position-information
  • the position-information operation unit calculates a difference between each coordinate position of the position indicator on the image registered in advance in the storage of the computer, and a coordinate position recognized in real time by the coordinate position recognizing unit, thereby to superimpose in real time the contents relating to the target object on the target object based on the calculation result.
  • the HMD associated with the portable type terminal device is provided with the camera for shooting the position indicators placed around the target object on a vertical plane to effect superimposed display by positioning the contents relating to the target object on a real image of the target object in real time, so that without using a special three dimensional position-and-attitude sensor-or the like, that is to say, with only such a simple arrangement as the position indicator and the camera combined, it is possible to provide a viewer with interactive information full of rich sources in which the target object and its related contents are integrated together. Thus, it is possible to give an advanced immersion sense to the viewer, as well as to enhance the effect of exhibition when the target object is an exhibit, for example.
  • an optical see-through type head mount display which includes a half mirror for permeating a reflected light from the target object toward a visible position of the target object, and which does not output a background image shot by the camera to the video display.
  • a video see-through type head mount display which includes a both-sided mirror located midway along a reflected light course from the target object toward the visual position, for refracting a reflected light from the target object in other directions than the one toward the visual position, and which outputs a background image shot by the camera to the video display.
  • FIG. 1 is a block diagram schematically showing a mixed reality realizing system according to a preferred embodiment of the invention
  • FIG. 2 is a plan view illustrating an optical positional relationship of an HMD in the system
  • FIG. 3 is an elevation illustrating an example for realizing a mixed reality for a planar exhibit by using the system
  • FIG. 4 is a plan view illustrating the same example
  • FIG. 5 shows an image viewed by the HMD of the system
  • FIG. 6 is a flow chart showing operating processes by the system.
  • a mixed reality realizing system of the invention comprises four infrared LEDs 1 located around an exhibit (an example of a target object) in a vertical plane including the exhibit, and a portable type terminal device (for example, one which is wearable on a body or one which can be carried in a pocket) 5 comprising a single eye type head mount display (hereinafter, referred to as HMD) 2 , a controller 3 and a computer 4 .
  • Each infrared LED 1 is used as an identification signal transmitter for transmitting an identification signal 6 of an exhibit to identify the exhibit, and also as a position indicator for indicating a position of the exhibit.
  • At least one infrared LED 1 is required as the identification signal transmitter, and at least three infrared LEDs are required as the position indicator. If one out of three infrared LEDs used as position indicators is used also as an identification signal transmitter, it may be enough to provide total three infrared LEDs. Thus, by using infrared LEDs, one LED can be used in common as the identification signal transmitter and as the position indicator. As a matter of course, the identification signal transmitter and the position indicator may be used separately as different devices.
  • the HMD 2 of the portable type terminal device 5 comprises a video display (a liquid crystal or CRT display may be used) 7 , a sound transmitter (an ear phone is generally used but a small-size speaker may be used) 8 , an infrared light receiver 9 for receiving an infrared light identification signal 6 transmitted from the infrared LED 1 to identify a specific exhibit, and a CCD camera (an example of shooting means, and a CMOS camera may be used instead) 10 for shooting a plurality of infrared LEDs 1 .
  • the infrared light receiver 9 represents an example of the identification signal receiver.
  • a filter is provided for passing only light waves having the same wave length as of the light emitted from a light source, so that the light serving as the identification signal can be recognized and the position-indicating light of the position indicator can be recognized by performing the same image processing as for the case where the infrared light is used.
  • the CCD camera 10 is mounted on the HMD 2 to allow shooting a back ground that is the same as what a viewer H sees through the HMD 2 , and as the HMD 2 mounted with this CCD camera 10 , either one of an optical see-through type display or a video-see-through type display may be used.
  • a mirror M mounted to the HMD 2 is a half mirror.
  • the half mirror M is placed at a position on a reflected light P toward an eye of the viewer H from an exhibit A displayed vertically on a vertical plane (wall surface) W, and the CCD camera 10 and the video display 7 are optically placed on both sides of the half mirror M through lenses 16 and 17 , respectively.
  • the background image shot by the CCD camera 10 is not output to the video display 7 . Yet, it is indispensable to shoot the infrared LED 1 by the CCD camera 10 to determine the display position and size of a superimposition image to be displayed on the video display 7 .
  • a mirror M mounted to the HMD 2 is a both sided mirror. That is, the reflected light P toward the eye of the viewer H from the exhibit A displayed in a vertical posture on a vertical plane W is refracted at an right angle by the both sided mirror M when the background image is shot by the CCD camera 10 , and the background image thus shot is output to the video display 7 and is superimposed with a predetermined superimposition image.
  • the computer 4 of the portable type terminal device 5 comprises a storage 11 for registering in advance contents (this sort of contents includes various kinds of data such as technical data, and may simply be called information) including images, explanatory sentences, sounds and so on, all relating to an exhibit, and data (hereinafter called as reference coordinate data, which will be described later) including a coordinate position on an image of the infrared LED 1 shot by the CCD camera 10 from a reference position predetermined by the exhibit; a control operation unit 12 for receiving a signal from the infrared light receiver 9 and retrieving from the storage 11 the contents relating to the exhibit such as images, explanatory sentences, sounds and so on; a coordinate position recognizing unit 13 for recognizing in real time a coordinate position of the infrared LED 1 on an image shot by the CCD camera 10 ; a position-information operation unit 14 for calculating, based on the recognized coordinate data, a visible position of the exhibit by a viewer wearing the HMD 2 of the portable type terminal device 5 and a position of the exhibit; and
  • an identification signal 6 of an exhibit A is transmitted from one of four infrared LEDs 1 located at four corners of the exhibit A and on a same vertical plane W as the exhibit A hanged in a vertical posture on this vertical plane W which is a wall in an exhibition room R.
  • first reference point a position outside and close to the upper left corner of the exhibit A where one infrared LED 1 is located
  • positions where other three infrared LEDs 1 are located are defined as second, third, and fourth reference points clockwise.
  • step S 21 When an identification signal 6 transmitted from one of the infrared LEDs 1 is received by the infrared light receiver 9 of the HMD 2 (step S 21 ), an output signal from this infrared light receiver 9 is output to the control operation unit 12 in the computer 4 .
  • this control operation unit 12 operates to retrieve from the storage 11 the contents such as images, explanatory sentences or voices all relating to the exhibit and the reference coordinate data (step S 22 ), and the retrieved reference coordinate data is sent to the position-information operation unit 14 while the contents are sent to the content superimposing unit 15 (steps S 23 , S 24 ).
  • the above mentioned reference coordinate data includes, as shown in FIG. 4, distances between the first and fourth reference points (or the second and third reference points) and between the first and second reference points on an image of the infrared LED 1 shot by the CCD camera 10 of the HMD 2 from a reference position BP determined in accordance with the exhibit A, and, as shown in FIG. 5, a coordinate position for superimposing an image SE relating to the exhibit A (a coordinate position in the upper left corner of the image SE when the upper left corner of an image LE shot by the CCD camera 10 from the reference position BP is defined as original point 0 ).
  • step S 25 the coordinate position recognizing unit 13 of the computer 4 .
  • step S 26 the coordinate position is obtained on the image LE of the four infrared LEDs thus shot by the CCD camera (the upper left corner of the image is determined as original point) (step S 26 ). From these coordinate points, each distance on the image between the first and the fourth reference points (or the second and the third reference points) and between the first and the second reference points are obtained.
  • the obtained distances are compared with those distances which have been obtained in advance by the shooting from the reference position BP, respectively between the first and the fourth reference points (or the second and the third reference points) and between the first and the second reference points, whereby the position information on the viewer H relative to the exhibit A, that is to say a distance and an angle, is calculated.
  • the distance measured in advance between the first and the fourth reference points (or the second and the third reference points) is normalized as “1” to calculate the ratio scale (percentage) of the distance ⁇ measured between the first and the fourth reference points (or the second and the third reference points) based on the shooting from the real visible position RP of the viewer H, to the normalized distance.
  • the distance measured in advance between the first and the second reference points is normalized as “1” to calculate the ratio scale (percentage) of the distance ⁇ measured between the first and the second reference points based on the shooting from the real visible position RP of the viewer H, to the normalized distance (step S 27 ).
  • the distance ⁇ may be calculated from the length between the second and the third reference points.
  • a coordinate position is determined for superimposing the image SE relating to the exhibit on the image LE which has been shot from the real visible position RP of the viewer H, and scale alignment of the image SE is performed (step S 28 ). Thereafter, the image SE is superimposed on the exhibit A (step S 29 ).
  • difference between a coordinate position at the first reference point shot from the real visible position RP of the viewer H and a coordinate position of the first reference point shot from the reference position BP is calculated, and a coordinate position (a coordinate position on the image shot from the reference position BP) at the upper left corner of the image SE for superimposition is translated by the amount of the difference.
  • difference between the coordinate position at the upper left corner of the image SE, which has been translated, and the coordinate position at the first reference point shot from the real visible position RP of the viewer H is calculated.
  • a Y component (Y represents a vertical direction) of the difference is multiplied by a times
  • a X component (X represents a horizontal direction) of the difference is multiplied by ⁇ times, thereby to determine the coordinate position of the image SE at its upper left corner with respect to the first reference point shot from the real visible position RP of the viewer H.
  • images registered in advance in the storage 11 are multiplied by a times in both X and Y directions, and further multiplied by ⁇ times in the X direction. Thereafter, the image SE is adjusted so that its coordinate position at the upper left corner corresponds to the coordinate position determined as mentioned above. Thus, the image SE is superimposed in real time on the image LE.
  • the image SE for superimposition is displayed at a position which comes within a range of the exhibit A, but it is not necessarily restricted to it. It is possible to display the image SE at a position which is not overlapping the exhibit A.
  • a display position changing unit (display position changing means) is provided for changing a display position of an image so that a “reference position” for determining a coordinate position at the upper left corner of the image, that is, a “reference position of display” may be changed to a second reference position by this display position changing unit.
  • the display position of the image for superimposition may deviate from within the range of an exhibit to an outside space on the right.
  • the exhibit hanged on the vertical plane has a planar form, but it is possible to realize the same mixed reality as the above described one by setting a virtual vertical plane for a three dimensional exhibit placed in a free space and locating an infrared LED 1 at each four corners of the exhibit. Further, by applying three dimensional graphics to the image SE for superimposition, it becomes possible to superimpose a three dimensional image on a three dimensional exhibit. Furthermore, regardless of presence/absence of an image for superimposition, various guidance, explanations and the like may be transmitted from a sound transmitter, allowing a viewer to have a deeper immersion sense.
  • the system of the present invention is aimed to realize a mixed reality in association with each exhibit in cultural facilities such as a museum, it is contemplated that the system may be applied in different fields such as for inspections of various industrial machines and apparatuses, and medical diagnoses, as well as in the fields of education, entertainment and so forth. That is to say, a target object of the system may be just anything as far as reference information thereon may be provided.

Abstract

According to a mixed reality realizing system of the present invention, when a viewer wears a head mount display (HMD) including a video display and a CCD camera, and views an exhibit using an optical see-through type system, identification signals and position signals transmitted from infrared LEDs located around the exhibit are received to determine a position of the exhibit, and to superimpose contents such as explanations of the exhibit, which have been stored in advance in a portable computer, on a predetermined position on the video display where the exhibit is displayed.

Description

TECHNICAL FIELD OF INVENTION
The present invention relates to a mixed reality realizing system for use in exhibition facilities like a museum for example, which provides viewers with information on exhibits together with the exhibits they are viewing, by integrating a virtual space with a real space.
BACKGROUND OF THE INVENTION
In displaying excavated clay pots at a museum, for example, it is desirable to present to viewers not only exhibits themselves but with information on the exhibits, such as excavated places, conditions of excavation and the like.
Conventionally, in response to such a demand, explanatory panels have been installed or voice guidance systems have been employed to provide relevant information.
In recent year, following the advance in technology of microprocessors, such an electronic guidance system has been developed that uses a portable-type information terminal device or a personal digital assistant (PDA) for combining voices and images to provide viewers with desired information in an interactive manner.
However, when conventional explanatory panels are used, it is very difficult to provide pertinent information corresponding to interests of viewers in an interactive manner, because there are physical restrictions to sizes of the panels, space for placement of the panels and soon. When voice guidance systems are used, there are no physical restrictions but it is impossible to provide visual information such as pictures, images and so on. Consequently, it is impossible to expect such display effects as providing viewers with reality sense and immersion sense.
Contrary to these, the electronic guidance system using the PDA is capable of providing viewers with interactive information corresponding to their interests by combining voices and images. Therefore, this system is considered to be the best one in the present circumstances.
However, when the electronic guidance system is utilized, viewers have to bother to hold the PDA with both hands to operate it, inconveniencing the operation thereof, besides there is difference in display position between an exhibit itself and its image projected by the PDA, necessitating the viewers to frequently move their sight lines between the exhibit and the PDA. Thus, a problem is posed that the viewers cannot afford to fully enjoy appreciating the exhibit.
DISCLOSURE OF THE INVENTION
Therefore, the present invention has an object to provide a mixed reality realizing system capable of producing a display effect of giving reality sense and immersion sense.
In order to achieve this object, the mixed reality realizing system of the present invention comprises identification signal transmitters located around a target object to transmit identification signals of the target object, a plurality of position indicators for indicating a position of the target object, and a portable type terminal device comprising a head mount display (HMD), a controller and a computer,
wherein the head mount display of the portable type terminal device comprises: a video display; an identification signal receiver for receiving the identification signals of the target object transmitted from the identification signal transmitters; and a camera for shooting the plurality of position indicators, and the computer comprises: a storage for storing, in advance, data including coordinate positions of the plurality of position indicators on an image shot by the camera from a reference position previously determined based on the target object and contents relating to the target object; a control operation unit for receiving the signals from the identification signal receiver and retrieving the contents relating to the target object from the storage; a coordinate position recognizing unit for recognizing, in real time, coordinate positions of the plurality of position indicators on the image shot by the camera; a position-information operation unit for calculating a visible position and a real position of the target object based on the plurality of coordinate positions obtained from the position indicators; and a content superimposing unit for superimposing the contents relating to the target object on the target object on the video display of the head mount display based on the information on the plurality of positions obtained from the position-information operation unit, the contents having been registered in advance in the storage, and
wherein the position-information operation unit calculates a difference between each coordinate position of the position indicator on the image registered in advance in the storage of the computer, and a coordinate position recognized in real time by the coordinate position recognizing unit, thereby to superimpose in real time the contents relating to the target object on the target object based on the calculation result.
According to the foregoing, the HMD associated with the portable type terminal device is provided with the camera for shooting the position indicators placed around the target object on a vertical plane to effect superimposed display by positioning the contents relating to the target object on a real image of the target object in real time, so that without using a special three dimensional position-and-attitude sensor-or the like, that is to say, with only such a simple arrangement as the position indicator and the camera combined, it is possible to provide a viewer with interactive information full of rich sources in which the target object and its related contents are integrated together. Thus, it is possible to give an advanced immersion sense to the viewer, as well as to enhance the effect of exhibition when the target object is an exhibit, for example.
Further, according to a preferred embodiment of the present invention, used as the head mount display of the mixed reality realizing system is an optical see-through type head mount display which includes a half mirror for permeating a reflected light from the target object toward a visible position of the target object, and which does not output a background image shot by the camera to the video display.
Still further, according to the preferred embodiment of the present invention, used as the head mount display of the mixed reality realizing system is a video see-through type head mount display which includes a both-sided mirror located midway along a reflected light course from the target object toward the visual position, for refracting a reflected light from the target object in other directions than the one toward the visual position, and which outputs a background image shot by the camera to the video display.
A number of characteristic features and effects of the present invention will be clarified by the preferred embodiment described below with reference to annexed drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram schematically showing a mixed reality realizing system according to a preferred embodiment of the invention;
FIG. 2 is a plan view illustrating an optical positional relationship of an HMD in the system;
FIG. 3 is an elevation illustrating an example for realizing a mixed reality for a planar exhibit by using the system;
FIG. 4 is a plan view illustrating the same example;
FIG. 5 shows an image viewed by the HMD of the system; and
FIG. 6 is a flow chart showing operating processes by the system.
DESCRIPTION OF PREFERRED EMBODIMENTS
As shown in FIG. 1, a mixed reality realizing system of the invention comprises four infrared LEDs 1 located around an exhibit (an example of a target object) in a vertical plane including the exhibit, and a portable type terminal device (for example, one which is wearable on a body or one which can be carried in a pocket) 5 comprising a single eye type head mount display (hereinafter, referred to as HMD) 2, a controller 3 and a computer 4. Each infrared LED 1 is used as an identification signal transmitter for transmitting an identification signal 6 of an exhibit to identify the exhibit, and also as a position indicator for indicating a position of the exhibit. At least one infrared LED 1 is required as the identification signal transmitter, and at least three infrared LEDs are required as the position indicator. If one out of three infrared LEDs used as position indicators is used also as an identification signal transmitter, it may be enough to provide total three infrared LEDs. Thus, by using infrared LEDs, one LED can be used in common as the identification signal transmitter and as the position indicator. As a matter of course, the identification signal transmitter and the position indicator may be used separately as different devices.
The HMD 2 of the portable type terminal device 5 comprises a video display (a liquid crystal or CRT display may be used) 7, a sound transmitter (an ear phone is generally used but a small-size speaker may be used) 8, an infrared light receiver 9 for receiving an infrared light identification signal 6 transmitted from the infrared LED 1 to identify a specific exhibit, and a CCD camera (an example of shooting means, and a CMOS camera may be used instead) 10 for shooting a plurality of infrared LEDs 1. The infrared light receiver 9 represents an example of the identification signal receiver. When a light used in common for both the position indicator and the identification signal transmitter is other than an infrared light, a filter is provided for passing only light waves having the same wave length as of the light emitted from a light source, so that the light serving as the identification signal can be recognized and the position-indicating light of the position indicator can be recognized by performing the same image processing as for the case where the infrared light is used.
The CCD camera 10, as shown in FIG. 2, is mounted on the HMD 2 to allow shooting a back ground that is the same as what a viewer H sees through the HMD 2, and as the HMD 2 mounted with this CCD camera 10, either one of an optical see-through type display or a video-see-through type display may be used.
In the optical see-through type HMD 2, a mirror M mounted to the HMD 2 is a half mirror. In this case, as shown in FIG. 2, the half mirror M is placed at a position on a reflected light P toward an eye of the viewer H from an exhibit A displayed vertically on a vertical plane (wall surface) W, and the CCD camera 10 and the video display 7 are optically placed on both sides of the half mirror M through lenses 16 and 17, respectively.
According to this optical see-through type HMD 2, the background image shot by the CCD camera 10 is not output to the video display 7. Yet, it is indispensable to shoot the infrared LED 1 by the CCD camera 10 to determine the display position and size of a superimposition image to be displayed on the video display 7.
In the video-see-through type HMD 2, a mirror M mounted to the HMD 2 is a both sided mirror. That is, the reflected light P toward the eye of the viewer H from the exhibit A displayed in a vertical posture on a vertical plane W is refracted at an right angle by the both sided mirror M when the background image is shot by the CCD camera 10, and the background image thus shot is output to the video display 7 and is superimposed with a predetermined superimposition image. In this video-see-through type HMD 2, too, it is indispensable to shoot the infrared LED 1 by the CCD camera 10 to determine the display position and size of a superimposition image to be displayed on the video display 7.
The computer 4 of the portable type terminal device 5 comprises a storage 11 for registering in advance contents (this sort of contents includes various kinds of data such as technical data, and may simply be called information) including images, explanatory sentences, sounds and so on, all relating to an exhibit, and data (hereinafter called as reference coordinate data, which will be described later) including a coordinate position on an image of the infrared LED 1 shot by the CCD camera 10 from a reference position predetermined by the exhibit; a control operation unit 12 for receiving a signal from the infrared light receiver 9 and retrieving from the storage 11 the contents relating to the exhibit such as images, explanatory sentences, sounds and so on; a coordinate position recognizing unit 13 for recognizing in real time a coordinate position of the infrared LED 1 on an image shot by the CCD camera 10; a position-information operation unit 14 for calculating, based on the recognized coordinate data, a visible position of the exhibit by a viewer wearing the HMD 2 of the portable type terminal device 5 and a position of the exhibit; and a content superimposing unit 15 for superimposing the contents relating to the exhibit on a real image of the exhibit on the video display 7 of the HMD 2 by adjusting, based on the position information, the images or explanatory sentences relating to the exhibit which have been retrieved from the storage 11 by the control operation unit 12 to the position of the real image of the exhibit on the video display 7. Incidentally, the storage 11 may be placed outside by adding a communication function.
Next, a method for realizing a mixed reality in a museum, for example, by using a system of the above-stated system will be described with reference to FIGS. 3 to 5 and a flow chart in FIG. 6.
As shown in FIG. 3 and FIG. 4, an identification signal 6 of an exhibit A is transmitted from one of four infrared LEDs 1 located at four corners of the exhibit A and on a same vertical plane W as the exhibit A hanged in a vertical posture on this vertical plane W which is a wall in an exhibition room R. Here, a position outside and close to the upper left corner of the exhibit A where one infrared LED 1 is located is defined as first reference point, and positions where other three infrared LEDs 1 are located are defined as second, third, and fourth reference points clockwise.
When an identification signal 6 transmitted from one of the infrared LEDs 1 is received by the infrared light receiver 9 of the HMD 2 (step S21), an output signal from this infrared light receiver 9 is output to the control operation unit 12 in the computer 4.
Then, this control operation unit 12 operates to retrieve from the storage 11 the contents such as images, explanatory sentences or voices all relating to the exhibit and the reference coordinate data (step S22), and the retrieved reference coordinate data is sent to the position-information operation unit 14 while the contents are sent to the content superimposing unit 15 (steps S23, S24).
The above mentioned reference coordinate data includes, as shown in FIG. 4, distances between the first and fourth reference points (or the second and third reference points) and between the first and second reference points on an image of the infrared LED 1 shot by the CCD camera 10 of the HMD 2 from a reference position BP determined in accordance with the exhibit A, and, as shown in FIG. 5, a coordinate position for superimposing an image SE relating to the exhibit A (a coordinate position in the upper left corner of the image SE when the upper left corner of an image LE shot by the CCD camera 10 from the reference position BP is defined as original point 0).
Then, in the coordinate position recognizing unit 13 of the computer 4, four infrared LEDs 1 are shot by the CCD camera 10 of the HMD 2 from a real visible position RP which is selected arbitrarily by the viewer H wearing the portable type terminal device 5 (step S25), and a coordinate position is obtained on the image LE of the four infrared LEDs thus shot by the CCD camera (the upper left corner of the image is determined as original point) (step S26). From these coordinate points, each distance on the image between the first and the fourth reference points (or the second and the third reference points) and between the first and the second reference points are obtained.
Subsequently, in the position-information operation unit 14 of the computer 4, the obtained distances are compared with those distances which have been obtained in advance by the shooting from the reference position BP, respectively between the first and the fourth reference points (or the second and the third reference points) and between the first and the second reference points, whereby the position information on the viewer H relative to the exhibit A, that is to say a distance and an angle, is calculated.
More specifically, the distance measured in advance between the first and the fourth reference points (or the second and the third reference points) is normalized as “1” to calculate the ratio scale (percentage) of the distance α measured between the first and the fourth reference points (or the second and the third reference points) based on the shooting from the real visible position RP of the viewer H, to the normalized distance. In the same manner, the distance measured in advance between the first and the second reference points is normalized as “1” to calculate the ratio scale (percentage) of the distance α×β measured between the first and the second reference points based on the shooting from the real visible position RP of the viewer H, to the normalized distance (step S27). Incidentally, when it is impossible to shoot the fourth reference point due to existence of other viewers, the distance α may be calculated from the length between the second and the third reference points.
Then, in the content superimposing unit 15 of the computer 4, a coordinate position is determined for superimposing the image SE relating to the exhibit on the image LE which has been shot from the real visible position RP of the viewer H, and scale alignment of the image SE is performed (step S28). Thereafter, the image SE is superimposed on the exhibit A (step S29).
Particularly, in a first place, difference between a coordinate position at the first reference point shot from the real visible position RP of the viewer H and a coordinate position of the first reference point shot from the reference position BP is calculated, and a coordinate position (a coordinate position on the image shot from the reference position BP) at the upper left corner of the image SE for superimposition is translated by the amount of the difference.
Next, difference between the coordinate position at the upper left corner of the image SE, which has been translated, and the coordinate position at the first reference point shot from the real visible position RP of the viewer H is calculated. Then, a Y component (Y represents a vertical direction) of the difference is multiplied by a times, and a X component (X represents a horizontal direction) of the difference is multiplied by α×β times, thereby to determine the coordinate position of the image SE at its upper left corner with respect to the first reference point shot from the real visible position RP of the viewer H.
Finally, images registered in advance in the storage 11 are multiplied by a times in both X and Y directions, and further multiplied by β times in the X direction. Thereafter, the image SE is adjusted so that its coordinate position at the upper left corner corresponds to the coordinate position determined as mentioned above. Thus, the image SE is superimposed in real time on the image LE.
Here, if the viewer H approaches to the exhibit A and the infrared LED 1 shot by the CCD camera 10 is caused to run off the edge of the image LE, the superimposed image before such running off the edge is kept and put into stand-by.
Further, in the embodiment mentioned above, the image SE for superimposition is displayed at a position which comes within a range of the exhibit A, but it is not necessarily restricted to it. It is possible to display the image SE at a position which is not overlapping the exhibit A.
In other words, a display position changing unit (display position changing means) is provided for changing a display position of an image so that a “reference position” for determining a coordinate position at the upper left corner of the image, that is, a “reference position of display” may be changed to a second reference position by this display position changing unit. In this consequence, the display position of the image for superimposition may deviate from within the range of an exhibit to an outside space on the right.
In this case, too, it is possible to use the above mentioned α and β to express the image SE for superimposition as if the image SE follows the same size variation as that of the exhibit A which varies in size according to a visual position.
Further, in the embodiment described above, the exhibit hanged on the vertical plane has a planar form, but it is possible to realize the same mixed reality as the above described one by setting a virtual vertical plane for a three dimensional exhibit placed in a free space and locating an infrared LED 1 at each four corners of the exhibit. Further, by applying three dimensional graphics to the image SE for superimposition, it becomes possible to superimpose a three dimensional image on a three dimensional exhibit. Furthermore, regardless of presence/absence of an image for superimposition, various guidance, explanations and the like may be transmitted from a sound transmitter, allowing a viewer to have a deeper immersion sense.
In addition, although the system of the present invention is aimed to realize a mixed reality in association with each exhibit in cultural facilities such as a museum, it is contemplated that the system may be applied in different fields such as for inspections of various industrial machines and apparatuses, and medical diagnoses, as well as in the fields of education, entertainment and so forth. That is to say, a target object of the system may be just anything as far as reference information thereon may be provided.

Claims (3)

What is claimed is:
1. A mixed reality realizing system, comprising identification signal transmitters located around a target object to transmit identification signals of the target object, a plurality of position indicators for indicating a position of the target object, and a portable type terminal device comprising a head mount display, a controller and a computer,
the head mount display of said portable type terminal device comprising:
a video screen;
an identification signal receiver for receiving the identification signals of the target object transmitted from the identification signal transmitters; and
a camera for shooting the plurality of position indicators,
the computer comprising:
a storage for storing, in advance, data including coordinate positions of the plurality of position indicators on an image shot by the camera from a reference position previously determined based on the target object and contents relating to the target object;
a control operation unit for receiving the signals from the identification signal receiver and retrieving the contents relating to the target object from the storage;
a coordinate position recognizing unit for recognizing, in real time, coordinate positions of the plurality of position indicators on the image shot by the camera;
a position-information operation unit for calculating a visible position and a real position of the target object based upon the plurality of coordinate positions obtained from the position indicators; and
a content superimposing unit for superimposing the contents relating to the target object on the target object on the video display of the head mount display based on the information on the plurality of positions obtained from the position-information operation unit, the contents having been registered in advance in the storage,
wherein the position-information operation unit calculates a difference between each coordinate position of the position indicator on the image registered in advance in the storage of the computer, and a coordinate position recognized in real time by the coordinate position recognizing unit, thereby to superimpose in real time the contents relating to the target object on the target object based on the calculation result.
2. The mixed reality realizing system as defined in claim 1, wherein an optical see-through type system is used as the head mount display, the optical see-through type system including a half mirror for permeating a reflected light from the target object toward a visible position of the target object, and outputting no background image shot by the camera to the video display.
3. The mixed reality realizing system as defined in claim 1, wherein a video see-through type system is used as the head mount display, the video see-through type system including a both-sided mirror located on a reflected light course from the target object toward the visible position of the target object to allow the reflected light from the target object to refract in other directions than the one toward the visible position, and outputting a background image shot by the camera to the video display.
US10/189,127 2000-03-31 2002-07-02 Mixed reality realizing system Expired - Lifetime US6795041B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2001097782A JP3957468B2 (en) 2000-03-31 2001-03-30 Mixed reality realization system
US10/189,127 US6795041B2 (en) 2000-03-31 2002-07-02 Mixed reality realizing system

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2000096055 2000-03-31
JP2001097782A JP3957468B2 (en) 2000-03-31 2001-03-30 Mixed reality realization system
US10/189,127 US6795041B2 (en) 2000-03-31 2002-07-02 Mixed reality realizing system

Publications (2)

Publication Number Publication Date
US20040004583A1 US20040004583A1 (en) 2004-01-08
US6795041B2 true US6795041B2 (en) 2004-09-21

Family

ID=32303121

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/189,127 Expired - Lifetime US6795041B2 (en) 2000-03-31 2002-07-02 Mixed reality realizing system

Country Status (2)

Country Link
US (1) US6795041B2 (en)
JP (1) JP3957468B2 (en)

Cited By (66)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030062675A1 (en) * 2001-09-28 2003-04-03 Canon Kabushiki Kaisha Image experiencing system and information processing method
US20040078263A1 (en) * 2002-10-16 2004-04-22 Altieri Frances Barbaro System and method for integrating business-related content into an electronic game
US20050288078A1 (en) * 2004-05-28 2005-12-29 Cheok Adrian D Game
US20050285878A1 (en) * 2004-05-28 2005-12-29 Siddharth Singh Mobile platform
US20050289590A1 (en) * 2004-05-28 2005-12-29 Cheok Adrian D Marketing platform
US20080094417A1 (en) * 2005-08-29 2008-04-24 Evryx Technologies, Inc. Interactivity with a Mixed Reality
US7474318B2 (en) 2004-05-28 2009-01-06 National University Of Singapore Interactive system and method
US20090243969A1 (en) * 2008-03-31 2009-10-01 Brother Kogyo Kabushiki Kaisha Display processor and display processing system
US20110158478A1 (en) * 2008-09-11 2011-06-30 Brother Kogyo Kabushiki Kaisha Head mounted display
US20110170747A1 (en) * 2000-11-06 2011-07-14 Cohen Ronald H Interactivity Via Mobile Image Recognition
US20130124518A1 (en) * 2011-11-14 2013-05-16 Sony Corporation Information registration device, information registration method, information registration system, information presentation device, informaton presentation method, informaton presentaton system, and program
US20140225916A1 (en) * 2013-02-14 2014-08-14 Research In Motion Limited Augmented reality system with encoding beacons
US20150205100A1 (en) * 2014-01-21 2015-07-23 Osterhout Group, Inc. Eye imaging in head worn computing
USD743963S1 (en) 2014-12-22 2015-11-24 Osterhout Group, Inc. Air mouse
USD751552S1 (en) 2014-12-31 2016-03-15 Osterhout Group, Inc. Computer glasses
USD753114S1 (en) 2015-01-05 2016-04-05 Osterhout Group, Inc. Air mouse
US9377625B2 (en) 2014-01-21 2016-06-28 Osterhout Group, Inc. Optical configurations for head worn computing
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9423612B2 (en) 2014-03-28 2016-08-23 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US9423842B2 (en) 2014-09-18 2016-08-23 Osterhout Group, Inc. Thermal management for head-worn computer
US9436006B2 (en) 2014-01-21 2016-09-06 Osterhout Group, Inc. See-through computer display systems
US9448409B2 (en) 2014-11-26 2016-09-20 Osterhout Group, Inc. See-through computer display systems
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
US9523856B2 (en) 2014-01-21 2016-12-20 Osterhout Group, Inc. See-through computer display systems
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9532715B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US9547465B2 (en) 2014-02-14 2017-01-17 Osterhout Group, Inc. Object shadowing in head worn computing
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
US9720234B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US9740280B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. Eye imaging in head worn computing
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US9784973B2 (en) 2014-02-11 2017-10-10 Osterhout Group, Inc. Micro doppler presentations in head worn computing
US9811153B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US9939646B2 (en) 2014-01-24 2018-04-10 Osterhout Group, Inc. Stray light suppression for head worn computing
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US10062182B2 (en) 2015-02-17 2018-08-28 Osterhout Group, Inc. See-through computer display systems
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US10558050B2 (en) 2014-01-24 2020-02-11 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US11104272B2 (en) 2014-03-28 2021-08-31 Mentor Acquisition One, Llc System for assisted operator safety using an HMD
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US11269182B2 (en) 2014-07-15 2022-03-08 Mentor Acquisition One, Llc Content presentation in head worn computing
US11269041B2 (en) 2017-07-26 2022-03-08 Signify Holding B.V. System for communicating a presence of a device via a light source
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4356983B2 (en) * 2004-02-10 2009-11-04 キヤノン株式会社 Image processing method and image processing apparatus
US8547401B2 (en) * 2004-08-19 2013-10-01 Sony Computer Entertainment Inc. Portable augmented reality device and method
ITBO20040749A1 (en) * 2004-12-02 2005-03-02 Bieffebi Spa MACHINE FOR THE ASSEMBLY OF A FLEXOGRAPHIC CLICHE REGISTER WITH A VIRTUAL INFORMATION SYSTEM
JP4739002B2 (en) * 2005-06-30 2011-08-03 キヤノン株式会社 Image processing method and image processing apparatus
KR101397214B1 (en) * 2007-06-05 2014-05-20 삼성전자주식회사 System and method for generating virtual object using augmented reality
JP5311440B2 (en) * 2007-12-05 2013-10-09 独立行政法人産業技術総合研究所 Information presentation device
JP2011192048A (en) * 2010-03-15 2011-09-29 Nec Corp Speech content output system, speech content output device, and speech content output method
KR101080040B1 (en) 2010-03-25 2011-11-04 동서대학교산학협력단 Method for display spatial augmented reality-based interactive
JP5499854B2 (en) 2010-04-08 2014-05-21 ソニー株式会社 Optical position adjustment method for head mounted display
JP5434848B2 (en) 2010-08-18 2014-03-05 ソニー株式会社 Display device
JP6020009B2 (en) * 2012-09-28 2016-11-02 ブラザー工業株式会社 Head mounted display, method and program for operating the same
KR102013790B1 (en) * 2013-01-18 2019-08-26 한국전자통신연구원 Apparatus for control an interactive content based on l style multiple display and method for performing the same
JP5786910B2 (en) * 2013-08-29 2015-09-30 ソニー株式会社 Optical position adjustment method for head mounted display
DE102014115471B4 (en) * 2014-10-23 2018-08-23 Krauss-Maffei Wegmann Gmbh & Co. Kg Head-mounted display device and method
RU2644125C1 (en) * 2016-11-30 2018-02-07 Эдуард Борисович Попов Personal video guide
WO2019000429A1 (en) * 2017-06-30 2019-01-03 Guangdong Virtual Reality Technology Co., Ltd. Methods and systems for operating an apparatus through augmented reality
JP7056155B2 (en) * 2018-01-10 2022-04-19 富士フイルムビジネスイノベーション株式会社 Information transmission equipment, information transmission systems and programs

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08160879A (en) 1994-12-09 1996-06-21 Shimadzu Corp Head mounted display device
JPH09114543A (en) 1995-10-02 1997-05-02 Xybernaut Corp Handfree computer system
US20020196343A1 (en) * 2001-05-31 2002-12-26 Nassir Navab Input devices using infrared trackers
US6587783B2 (en) * 2000-10-05 2003-07-01 Siemens Corporate Research, Inc. Method and system for computer assisted localization, site navigation, and data navigation

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08160879A (en) 1994-12-09 1996-06-21 Shimadzu Corp Head mounted display device
JPH09114543A (en) 1995-10-02 1997-05-02 Xybernaut Corp Handfree computer system
US5844824A (en) * 1995-10-02 1998-12-01 Xybernaut Corporation Hands-free, portable computer and system
US6587783B2 (en) * 2000-10-05 2003-07-01 Siemens Corporate Research, Inc. Method and system for computer assisted localization, site navigation, and data navigation
US20020196343A1 (en) * 2001-05-31 2002-12-26 Nassir Navab Input devices using infrared trackers

Cited By (160)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110170747A1 (en) * 2000-11-06 2011-07-14 Cohen Ronald H Interactivity Via Mobile Image Recognition
US9087270B2 (en) 2000-11-06 2015-07-21 Nant Holdings Ip, Llc Interactivity via mobile image recognition
US9076077B2 (en) 2000-11-06 2015-07-07 Nant Holdings Ip, Llc Interactivity via mobile image recognition
US8817045B2 (en) 2000-11-06 2014-08-26 Nant Holdings Ip, Llc Interactivity via mobile image recognition
US20030062675A1 (en) * 2001-09-28 2003-04-03 Canon Kabushiki Kaisha Image experiencing system and information processing method
US8228325B2 (en) * 2002-10-16 2012-07-24 Frances Barbaro Altieri Interactive virtual thematic environment
US8458028B2 (en) 2002-10-16 2013-06-04 Barbaro Technologies System and method for integrating business-related content into an electronic game
US20040078263A1 (en) * 2002-10-16 2004-04-22 Altieri Frances Barbaro System and method for integrating business-related content into an electronic game
US20080284777A1 (en) * 2002-10-16 2008-11-20 Barbaro Technologies Interactive virtual thematic environment
US7474318B2 (en) 2004-05-28 2009-01-06 National University Of Singapore Interactive system and method
US20050285878A1 (en) * 2004-05-28 2005-12-29 Siddharth Singh Mobile platform
US20050288078A1 (en) * 2004-05-28 2005-12-29 Cheok Adrian D Game
US20050289590A1 (en) * 2004-05-28 2005-12-29 Cheok Adrian D Marketing platform
US10463961B2 (en) 2005-08-29 2019-11-05 Nant Holdings Ip, Llc Interactivity with a mixed reality
US10617951B2 (en) 2005-08-29 2020-04-14 Nant Holdings Ip, Llc Interactivity with a mixed reality
US20080094417A1 (en) * 2005-08-29 2008-04-24 Evryx Technologies, Inc. Interactivity with a Mixed Reality
US9600935B2 (en) 2005-08-29 2017-03-21 Nant Holdings Ip, Llc Interactivity with a mixed reality
US7564469B2 (en) * 2005-08-29 2009-07-21 Evryx Technologies, Inc. Interactivity with a mixed reality
US20090243969A1 (en) * 2008-03-31 2009-10-01 Brother Kogyo Kabushiki Kaisha Display processor and display processing system
US20110158478A1 (en) * 2008-09-11 2011-06-30 Brother Kogyo Kabushiki Kaisha Head mounted display
US8494212B2 (en) * 2008-09-11 2013-07-23 Brother Kogyo Kabushiki Kaisha Head mounted display
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US9020918B2 (en) * 2011-11-14 2015-04-28 Sony Corporation Information registration device, information registration method, information registration system, information presentation device, informaton presentation method, informaton presentaton system, and program
US20130124518A1 (en) * 2011-11-14 2013-05-16 Sony Corporation Information registration device, information registration method, information registration system, information presentation device, informaton presentation method, informaton presentaton system, and program
US20140225916A1 (en) * 2013-02-14 2014-08-14 Research In Motion Limited Augmented reality system with encoding beacons
US9483875B2 (en) * 2013-02-14 2016-11-01 Blackberry Limited Augmented reality system with encoding beacons
US11169623B2 (en) 2014-01-17 2021-11-09 Mentor Acquisition One, Llc External user interface for head worn computing
US11231817B2 (en) 2014-01-17 2022-01-25 Mentor Acquisition One, Llc External user interface for head worn computing
US11507208B2 (en) 2014-01-17 2022-11-22 Mentor Acquisition One, Llc External user interface for head worn computing
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US11782529B2 (en) 2014-01-17 2023-10-10 Mentor Acquisition One, Llc External user interface for head worn computing
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US11103132B2 (en) 2014-01-21 2021-08-31 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9523856B2 (en) 2014-01-21 2016-12-20 Osterhout Group, Inc. See-through computer display systems
US9529199B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9529192B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. Eye imaging in head worn computing
US9532715B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US9532714B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US9538915B2 (en) 2014-01-21 2017-01-10 Osterhout Group, Inc. Eye imaging in head worn computing
US10866420B2 (en) 2014-01-21 2020-12-15 Mentor Acquisition One, Llc See-through computer display systems
US11947126B2 (en) 2014-01-21 2024-04-02 Mentor Acquisition One, Llc See-through computer display systems
US9594246B2 (en) 2014-01-21 2017-03-14 Osterhout Group, Inc. See-through computer display systems
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems
US9615742B2 (en) 2014-01-21 2017-04-11 Osterhout Group, Inc. Eye imaging in head worn computing
US9651783B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651788B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651789B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-Through computer display systems
US10890760B2 (en) 2014-01-21 2021-01-12 Mentor Acquisition One, Llc See-through computer display systems
US9658458B2 (en) 2014-01-21 2017-05-23 Osterhout Group, Inc. See-through computer display systems
US9658457B2 (en) 2014-01-21 2017-05-23 Osterhout Group, Inc. See-through computer display systems
US11796805B2 (en) 2014-01-21 2023-10-24 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11796799B2 (en) 2014-01-21 2023-10-24 Mentor Acquisition One, Llc See-through computer display systems
US20150205100A1 (en) * 2014-01-21 2015-07-23 Osterhout Group, Inc. Eye imaging in head worn computing
US9684171B2 (en) 2014-01-21 2017-06-20 Osterhout Group, Inc. See-through computer display systems
US9684165B2 (en) 2014-01-21 2017-06-20 Osterhout Group, Inc. Eye imaging in head worn computing
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
US9720227B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9720234B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US9720235B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US11650416B2 (en) 2014-01-21 2023-05-16 Mentor Acquisition One, Llc See-through computer display systems
US9740012B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. See-through computer display systems
US9740280B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. Eye imaging in head worn computing
US9746676B2 (en) 2014-01-21 2017-08-29 Osterhout Group, Inc. See-through computer display systems
US11622426B2 (en) 2014-01-21 2023-04-04 Mentor Acquisition One, Llc See-through computer display systems
US10698223B2 (en) 2014-01-21 2020-06-30 Mentor Acquisition One, Llc See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US9772492B2 (en) 2014-01-21 2017-09-26 Osterhout Group, Inc. Eye imaging in head worn computing
US11002961B2 (en) 2014-01-21 2021-05-11 Mentor Acquisition One, Llc See-through computer display systems
US9811153B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9811152B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9811159B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US11619820B2 (en) 2014-01-21 2023-04-04 Mentor Acquisition One, Llc See-through computer display systems
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9829703B2 (en) 2014-01-21 2017-11-28 Osterhout Group, Inc. Eye imaging in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US9377625B2 (en) 2014-01-21 2016-06-28 Osterhout Group, Inc. Optical configurations for head worn computing
US11353957B2 (en) 2014-01-21 2022-06-07 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US10579140B2 (en) 2014-01-21 2020-03-03 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9885868B2 (en) 2014-01-21 2018-02-06 Osterhout Group, Inc. Eye imaging in head worn computing
US10321821B2 (en) 2014-01-21 2019-06-18 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9927612B2 (en) 2014-01-21 2018-03-27 Osterhout Group, Inc. See-through computer display systems
US9933622B2 (en) 2014-01-21 2018-04-03 Osterhout Group, Inc. See-through computer display systems
US9436006B2 (en) 2014-01-21 2016-09-06 Osterhout Group, Inc. See-through computer display systems
US11054902B2 (en) 2014-01-21 2021-07-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US9958674B2 (en) 2014-01-21 2018-05-01 Osterhout Group, Inc. Eye imaging in head worn computing
US9298007B2 (en) * 2014-01-21 2016-03-29 Osterhout Group, Inc. Eye imaging in head worn computing
US10001644B2 (en) 2014-01-21 2018-06-19 Osterhout Group, Inc. See-through computer display systems
US11126003B2 (en) 2014-01-21 2021-09-21 Mentor Acquisition One, Llc See-through computer display systems
US10139632B2 (en) 2014-01-21 2018-11-27 Osterhout Group, Inc. See-through computer display systems
US11099380B2 (en) 2014-01-21 2021-08-24 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
US9939646B2 (en) 2014-01-24 2018-04-10 Osterhout Group, Inc. Stray light suppression for head worn computing
US10558050B2 (en) 2014-01-24 2020-02-11 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US11822090B2 (en) 2014-01-24 2023-11-21 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9841602B2 (en) 2014-02-11 2017-12-12 Osterhout Group, Inc. Location indicating avatar in head worn computing
US9843093B2 (en) 2014-02-11 2017-12-12 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9784973B2 (en) 2014-02-11 2017-10-10 Osterhout Group, Inc. Micro doppler presentations in head worn computing
US9928019B2 (en) 2014-02-14 2018-03-27 Osterhout Group, Inc. Object shadowing in head worn computing
US9547465B2 (en) 2014-02-14 2017-01-17 Osterhout Group, Inc. Object shadowing in head worn computing
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9423612B2 (en) 2014-03-28 2016-08-23 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US11104272B2 (en) 2014-03-28 2021-08-31 Mentor Acquisition One, Llc System for assisted operator safety using an HMD
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US11880041B2 (en) 2014-04-25 2024-01-23 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US10634922B2 (en) 2014-04-25 2020-04-28 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US11474360B2 (en) 2014-04-25 2022-10-18 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US11727223B2 (en) 2014-04-25 2023-08-15 Mentor Acquisition One, Llc Language translation with head-worn computing
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US11960089B2 (en) 2014-06-05 2024-04-16 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11402639B2 (en) 2014-06-05 2022-08-02 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US10877270B2 (en) 2014-06-05 2020-12-29 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US11360318B2 (en) 2014-06-09 2022-06-14 Mentor Acquisition One, Llc Content presentation in head worn computing
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US10976559B2 (en) 2014-06-09 2021-04-13 Mentor Acquisition One, Llc Content presentation in head worn computing
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US11663794B2 (en) 2014-06-09 2023-05-30 Mentor Acquisition One, Llc Content presentation in head worn computing
US11790617B2 (en) 2014-06-09 2023-10-17 Mentor Acquisition One, Llc Content presentation in head worn computing
US11887265B2 (en) 2014-06-09 2024-01-30 Mentor Acquisition One, Llc Content presentation in head worn computing
US11022810B2 (en) 2014-06-09 2021-06-01 Mentor Acquisition One, Llc Content presentation in head worn computing
US11327323B2 (en) 2014-06-09 2022-05-10 Mentor Acquisition One, Llc Content presentation in head worn computing
US10139635B2 (en) 2014-06-09 2018-11-27 Osterhout Group, Inc. Content presentation in head worn computing
US9720241B2 (en) 2014-06-09 2017-08-01 Osterhout Group, Inc. Content presentation in head worn computing
US11789267B2 (en) 2014-06-17 2023-10-17 Mentor Acquisition One, Llc External user interface for head worn computing
US11054645B2 (en) 2014-06-17 2021-07-06 Mentor Acquisition One, Llc External user interface for head worn computing
US11294180B2 (en) 2014-06-17 2022-04-05 Mentor Acquisition One, Llc External user interface for head worn computing
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US10698212B2 (en) 2014-06-17 2020-06-30 Mentor Acquisition One, Llc External user interface for head worn computing
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US11269182B2 (en) 2014-07-15 2022-03-08 Mentor Acquisition One, Llc Content presentation in head worn computing
US11786105B2 (en) 2014-07-15 2023-10-17 Mentor Acquisition One, Llc Content presentation in head worn computing
US11360314B2 (en) 2014-08-12 2022-06-14 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US10908422B2 (en) 2014-08-12 2021-02-02 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US11630315B2 (en) 2014-08-12 2023-04-18 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US9423842B2 (en) 2014-09-18 2016-08-23 Osterhout Group, Inc. Thermal management for head-worn computer
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US9448409B2 (en) 2014-11-26 2016-09-20 Osterhout Group, Inc. See-through computer display systems
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
US11262846B2 (en) 2014-12-03 2022-03-01 Mentor Acquisition One, Llc See-through computer display systems
US11809628B2 (en) 2014-12-03 2023-11-07 Mentor Acquisition One, Llc See-through computer display systems
USD743963S1 (en) 2014-12-22 2015-11-24 Osterhout Group, Inc. Air mouse
USD792400S1 (en) 2014-12-31 2017-07-18 Osterhout Group, Inc. Computer glasses
USD751552S1 (en) 2014-12-31 2016-03-15 Osterhout Group, Inc. Computer glasses
USD753114S1 (en) 2015-01-05 2016-04-05 Osterhout Group, Inc. Air mouse
USD794637S1 (en) 2015-01-05 2017-08-15 Osterhout Group, Inc. Air mouse
US10062182B2 (en) 2015-02-17 2018-08-28 Osterhout Group, Inc. See-through computer display systems
US11269041B2 (en) 2017-07-26 2022-03-08 Signify Holding B.V. System for communicating a presence of a device via a light source

Also Published As

Publication number Publication date
JP2001344617A (en) 2001-12-14
JP3957468B2 (en) 2007-08-15
US20040004583A1 (en) 2004-01-08

Similar Documents

Publication Publication Date Title
US6795041B2 (en) Mixed reality realizing system
US11237392B2 (en) System for virtual display and method of use
US7817104B2 (en) Augmented reality apparatus and method
KR101637990B1 (en) Spatially correlated rendering of three-dimensional content on display components having arbitrary positions
US9355599B2 (en) Augmented information display
US7071898B2 (en) Method for using a wireless motorized camera mount for tracking in augmented reality
US11822708B2 (en) Methods, apparatus, systems, computer programs for enabling consumption of virtual content for mediated reality
US20170153866A1 (en) Audiovisual Surround Augmented Reality (ASAR)
US20100013738A1 (en) Image capture and display configuration
US10868977B2 (en) Information processing apparatus, information processing method, and program capable of adaptively displaying a video corresponding to sensed three-dimensional information
WO2015048906A1 (en) Augmented reality system and method for positioning and mapping
AU2014281726A1 (en) Virtual object orientation and visualization
JPH09505138A (en) Electro-optical vision system using position and orientation
US11244659B2 (en) Rendering mediated reality content
US11443487B2 (en) Methods, apparatus, systems, computer programs for enabling consumption of virtual content for mediated reality
US20200042792A1 (en) Audio processing
KR20090071851A (en) Development of annotaion system based on spatial augmented reality
KR101665363B1 (en) Interactive contents system having virtual Reality, augmented reality and hologram
US9989762B2 (en) Optically composited augmented reality pedestal viewer
US20220155592A1 (en) System for virtual display and method of use
US9406343B1 (en) Method of tracking for animation insertions to video recordings
US10499036B2 (en) Image sensing module
Lo et al. A directive location-based landscape tour service using augmented reality glass
KR20190142586A (en) Real-time Real Motion Interface by Object Redefinition

Legal Events

Date Code Title Description
AS Assignment

Owner name: HITACHI ZOSEN CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OGAWA, KENJI;SAEKI, TOSHIRO;TAMURA, HARUYO;REEL/FRAME:013393/0543

Effective date: 20021002

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

REMI Maintenance fee reminder mailed
FPAY Fee payment

Year of fee payment: 8

SULP Surcharge for late payment

Year of fee payment: 7

AS Assignment

Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034541/0477

Effective date: 20141014

FPAY Fee payment

Year of fee payment: 12

AS Assignment

Owner name: MICROSOFT CORPORATION, WASHINGTON

Free format text: MERGER;ASSIGNOR:HITACHI ZOSEN CORP;REEL/FRAME:044933/0900

Effective date: 20130328