WO2014144918A2 - Enhanced optical and perceptual digital eyewear - Google Patents

Enhanced optical and perceptual digital eyewear Download PDF

Info

Publication number
WO2014144918A2
WO2014144918A2 PCT/US2014/029520 US2014029520W WO2014144918A2 WO 2014144918 A2 WO2014144918 A2 WO 2014144918A2 US 2014029520 W US2014029520 W US 2014029520W WO 2014144918 A2 WO2014144918 A2 WO 2014144918A2
Authority
WO
WIPO (PCT)
Prior art keywords
optics device
wearable optics
utilized
wearable
lens
Prior art date
Application number
PCT/US2014/029520
Other languages
French (fr)
Other versions
WO2014144918A3 (en
Inventor
Scott W. Lewis
Original Assignee
Percept Technologies, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/841,141 external-priority patent/US8696113B2/en
Priority claimed from US13/841,550 external-priority patent/US11428937B2/en
Application filed by Percept Technologies, Inc. filed Critical Percept Technologies, Inc.
Publication of WO2014144918A2 publication Critical patent/WO2014144918A2/en
Publication of WO2014144918A3 publication Critical patent/WO2014144918A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/163Wearable computers, e.g. on a belt
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0127Head-up displays characterised by optical features comprising devices increasing the depth of field
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0138Head-up displays characterised by optical features comprising image capture systems, e.g. camera
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • G02B2027/0187Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye

Definitions

  • Patent Application No. 13/841 ,141 filed on March 15, 2013, entitled “ENHANCED OPTICAL AND PERCEPTUAL DIGITAL EYEWEAR”, which is a continuation-in-part and claims the benefit of priority to U.S. Patent Application No. 13/739,929, filed on January 1 1 , 2013, entitled “DIGITAL EYEWEAR", which is a continuation
  • This application also claims benefit under 35 USC 1 19(e) of U.S. Provisional Patent Application No. 61 /798,770, filed on March 15, 2013, entitled “ENHANCED DIGITAL EYEWEAR,” which is also incorporated herein by reference in its entirety.
  • the present invention relates generally to wearable optics and more particularly to wearable optics that includes additional functionality.
  • Wearable optics is utilized for a variety of purposes. Wearable optics is used for improving one's vision for reading glasses and to protect one's vision.
  • a wearable optics device and method of use comprises a wearable optics device, wherein parameters personalized to a user can be provided based upon user information.
  • a wearable optics device comprises a lens, wherein parameters personalized to a user can be provided based upon user information.
  • Figure 1 is a diagram that illustrates Media focals.
  • Figure 2 comprises an information bar on media focal wearable optics.
  • Figure 3 is a block diagram of wearable optics that is utilized in a music environment such as an MP3 player.
  • Figure 4 is a block diagram that illustrates wearable optics that is utilized as a cell phone.
  • Figure 5A is a block diagram that illustrates the cellular phone circuitry of Figure 4.
  • Figure 5B illustrates perceptual optimization utilizing optical and perceptual parameters.
  • Figure 5C illustrates the enhanced digital eyewear architecture.
  • Figure 6 illustrates a social networking application utilized with the wearable optics device.
  • Figure 7 illustrates a messaging application utilized with wearable optics device in accordance with an embodiment.
  • Figure 8 illustrates the wearable optics device utilized by an athletic sports spectator in accordance with an embodiment.
  • Figure 9 illustrates the wearable optics device utilized by an athletic sports player in accordance with an embodiment.
  • Figure 10 illustrates an augmented reality information, navigation, and advertising application utilized with the wearable optics device.
  • Figure 1 1 illustrates an augmented reality information patient data application utilized with the wearable optics device used in conjunction with a remote device.
  • Figure 12 illustrates a shading control application utilized with the wearable optics device.
  • Figure 13 illustrates an augmented reality application utilized with the wearable optics device.
  • Figure 14 illustrates a physical gaming application utilized with the wearable optics device.
  • Figure 15 illustrates a first embodiment of an online/mobile gaming application utilized with the wearable optics device.
  • Figure 16 illustrates a second embodiment of an online/mobile gaming application utilized with the wearable optics device.
  • Figure 17 illustrates shading control utilizing the wearable optics device.
  • Figure 18 illustrates an optical/perceptual operating system with the wearable optics device.
  • Figure 19 describes an embodiment of the digital architecture of the wearable optics device.
  • Figure 20 illustrates the embodiment of a system simulator for use by developers of applications and new lenses or expansion of the wearable optics device.
  • Figure 21 A thru Figure 21 F illustrate the embodiment of inverse shading using the wearable optics device.
  • Figure 22 illustrates an embodiment of eye tracking illumination and enhanced efficiency utilizing the wearable optics device.
  • Figure 23 illustrates an embodiment of real-time augmented reality overlay utilizing the wearable optics device.
  • the present invention relates generally to wearable optics and more particularly to wearable optics that includes additional functionality.
  • the following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements. Various modifications to the preferred embodiments and the generic principles and features described herein will be readily apparent to those skilled in the art. Thus, the present invention is not intended to be limited to the
  • a system and method in accordance with the present invention is directed to a variety of ways to enhance the use of wearable optics devices.
  • FIG. 1 is a diagram that illustrates Media focals 100.
  • Media focals 100 comprises an information bar 102, receiver 104, digital circuitry 106, frames 108 and lens 1 10.
  • Media focals 100 allow for enhancing the wearable optics for its primary purpose, for example, a digital camera could be placed within the wearable optics to allow for seeing certain of these images.
  • the circuitry 106 for the media focals 100 could be placed within the frame 108 of the wearable optics.
  • the lens 1 10 could have a totally reflective surface, or a partially reflective surface using LCDs or the like. In effect the wearable optics could look like see-through glasses, but through the use of the circuitry 106 within the wearable optics it is actually a media focal. Additionally, the wearable optics could incorporate a camera to project the user onto a second lens to achieve a see-through effect.
  • an information bar 102 is provided across a portion of the wearable optics which is visible to the user. This information bar 102 is used to convey a variety of types of information.
  • Figure 2 comprises an information bar 102' on media focal wearable optics.
  • the information bar 102' can be a stock ticker scrolling across the top portion of the wearable optics, as is shown in Figure 2.
  • the information bar 102' is shown displaying a stock ticker, other kinds of information such as song titles, lyrics and the like could be displayed in the information bar.
  • This information bar is referred to as E-focals. This information might be provided from a digital receiver through an FM station, through a cellular wireless device, or an MP3 player.
  • E-focal Additional functionality of the E-focal will be described with more detail with respect to the cell phone enhancements as well as the music player enhancements.
  • One of the key features of the media focals 100 is the use of the media focals to enhance the primary function of the user, that is, being able to more accurately and clearly see the objects.
  • a zoom feature circuit to allow for the use of the wearable optics as binoculars. This would allow for the user to see objects more closely based on certain activities of the user.
  • the circuitry 106 would be located somewhere in the frame of the glasses to provide this functionality and as circuits became smaller and devices became smaller it would be easier and easier to embed the circuitry that is well known for use for such functions directly within the device.
  • the circuitry 106 in the device could be, for example, eye sensors which could be pressure sensors, capacitive sensors or some other type of sensor for allowing the eyes to direct the activities. Eye movement sensors, for example, could be used to activate and control the binocular glasses. Similarly, a digital camera could be put on the glasses that would allow the same kinds of technology to take pictures by the person directly.
  • the glasses could be used as a normal corrective lens glass utilizing the digital imagery, so that, for example, a user has a certain prescription that they use with their normal prescription glasses to view an object clearly.
  • an optometrist could download the new prescription to the wearable optics such that a digital transformation of the image information is provided which is compatible with the new prescription.
  • a method for sensing and controlling the digital media could be implemented in a variety of ways.
  • an activity of the eye itself would control the activity of the media focal. So, for example, if the idea was to zoom the image, the eye would blink twice. It would also be possible to detect facial and eye movements (squinting, for example), as well as changes in the pupil and iris.
  • the eyeglasses in accordance with the present invention would be possible for the eyeglasses in accordance with the present invention to function within a client/server model or Bluetooth (Wi-Fi) model.
  • Wi-Fi Wi-Fi
  • Utilization of the client/server model and Bluetooth Wi-Fi would make possible, for example, the display of live news or special reports (such as financial reports) from the Internet or similar sources on the eyeglasses. This would also allow for portions of circuitry to be located remotely such that less circuitry in the wearable optics is required.
  • the wearable optics could also include a logo, for example, law enforcement officers could have their glasses emblazoned with "Police”, “Sheriff”, “MP”, etc.; young people could have their eyeglasses emblazoned with words and images that reflected their favorite performers, etc.; sports teams could offer the eyeglasses at discount with team monograms, etc. They could also be purchased by companies, emblazoned with the company logos, and given out as retirement gifts, etc.
  • Figure 3 is a block diagram of wearable optics 300 that is utilized in a music environment such as an MP3 player.
  • Figure 3 comprises wearable optics 300, an information bar 302, MP3 player circuitry 304, storage 306, frames 308, and one or a plurality of lenses 310.
  • Another environment as has been above described is the music environment. What would be desirable would be to provide music glasses in which an MP3 player on an IPod or the like is incorporated in the wearable optics, either in a wired or wireless environment.
  • a plurality of users could be networked via an MP3 player type environment within a hotspot, or the like, which would allow one to have downloads of whatever music is required through the eyeglasses.
  • the system could allow for downloadable music which could be selected via scrolling and the like through voice recognition systems.
  • the eyeglasses could link to a multimedia network, authorize downloading and billing for selected music.
  • access to a plurality of libraries for music selections could be provided.
  • the wearable optics could also be utilized in conjunction with wireless technology to allow a user or a plurality of users to participate simultaneously in single or group karaoke singing.
  • the wearable optics could be used specifically to display the lyrics of a song, melody, notes, name of the song or other associated references.
  • the headphones can be either digital or analog.
  • the user doesn't need to have 10,000 songs, for example. They can come enrolled in an in-song virtual network library upon entering a hotspot. Therefore, the local storage 306 could be limited. In addition, this would provide location identity information for one who is using the network.
  • the songs can be streamed as well as downloaded.
  • the songs could be purchase using the wearable optics.
  • the system could be scalable; depending upon what kind of device was being used. 3. Telecommunications Environment
  • Figure 4 is a block diagram that illustrates wearable optics that is utilized as a cell phone 400.
  • Figure 4 comprises cellular phone circuitry 402, a microphone 104, frames 408 and one or a plurality of lenses 410.
  • the cell phone wearable optics 400 could be implemented utilizing digital telephone technology.
  • Circuitry 402 within the wearable optics could be utilized to allow a telephone number or other visual information such as that provided by multimedia messaging services to be displayed on the lens 410 of the wearable optics as shown in Figure 3.
  • Figure 5 is a block diagram that illustrates the cellular phone circuitry of Figure 4.
  • Figure 5 comprises noise cancelling circuitry 502, voice recognition circuitry 504, caller ID circuitry 506 and speaker recognition circuitry 508 and media processing circuits 509.
  • the telephone number could be activated via the digital circuitry 402 as part of the media focals 100.
  • the circuitry could be made truly digital via a digital signal processor which is coupled to a camera otherwise in the environment.
  • the above system would allow for voice recording through use of a microphone 104 and would allow for voice recognition through use of the voice recognition circuitry 504, which would allow for signal conditioning on the cell phone in a variety of ways.
  • the cell phone environment 402 provides a plurality of areas for
  • one of the major annoyances in cell phone use is that the users have to speak in a loud manner because of background noise and the like.
  • the users have to speak in a loud manner because of background noise and the like.
  • the microphone could also be located in flip down microphones.
  • noise canceling circuitry 502 could be utilized to remove the background noise.
  • the microphone capability would include the advantage of utilizing noise rejection techniques. Buttons located on the wearable optics can be utilized to control features thereon.
  • the microphone 104 could utilize whisper technology such that the speaker will not have to speak as loudly.
  • the wearable optics would in a preferred embodiment include voice recognition circuitry 504 and caller ID circuitry 506.
  • the conventionality for hearing and talking in a preferred embodiment would be located in ear and nose pad portions of glasses.
  • the electronics for the cell phone in a preferred embodiment would be within the frame 308 of the wearable optics.
  • the wearable optics would include a fully integrated information bar 302.
  • a speaker recognition algorithm 508 as shown in Figure 5 would allow only the voice of the user to be recognized and the background noise would be cancelled. Accordingly, the unique characteristics of the speaker are provided via an audible model.
  • a digital client/server or Bluetooth/wifi model could be adapted to link the wearable optics to external communication equipment.
  • Such equipment could include digital cell phones, PDAs or wifi enabled PCs or other devices.
  • Such an embodiment could enable review of voicemail, screen viewed emails, text to speech audio email conversions, multimedia messaging services, and other data sources.
  • Wireless or Bluetooth interconnection could also make possible VOIP glasses to be utilized instead of a cell phone.
  • Other features enabled by a wireless link could link the eyewear to MP3 devices, an iPod, a printer, wireless/wired TV, coupons, and the like.
  • PDA glasses could provide built in a time display, alarm calendar, interfacing with PCs or network sources, a speaker and the like.
  • digital eyewear is a rapidly evolving field with from the early innovation of digital eyewear with eye tracking capabilities thru Lewis ("185 filed Feb 2008), to eyewear with more complex lenses and communication/display capabilities (Lewis '556, filed Nov 2009), to more enhancements and capabilities (Lewis '594, filed April 201 1 ).
  • This enhanced digital eyewear can be used to solve important areas ranging from superior vision enhancement and mobile advertising, to use in dental/medical procedures and physical and Internet navigation.
  • the application and value of the enhanced eyewear is increased even further when combined with augmented reality, social networking, messaging, and communications.
  • FIG. 5B illustrates a perceptual optimization system 550.
  • the perceptual optimization system 550 receives a variety of inputs including real world inputs, digital media inputs, expression parameter measurements, other parameter measurements to provide wearable optics visual display elements.
  • Real world inputs could be for example inputs from one or more microphones or cameras.
  • Digital media inputs could be for example from digital audio, digital video, graphics, images and augmented reality.
  • perceptual parameters could be for example, smell, touch, brainwave, temperature/humidity of the user, environmental conditions near the user.
  • FIG. 5C illustrates a wearable optics device architecture 560 in accordance with an embodiment.
  • the architecture includes a frame 562 which includes a plurality of sensors on various areas thereon.
  • Biometric sensors include a blood pressure sensor 617, temperature sensor 618, EEG sensor 616 and the like.
  • Environmental sensors 615 are also provided. There are microphone sensors 606,607, 61 1 on various areas of the frame. Included on the frame 562 are cameras rear, front and side 606,607, 61 1 to detect objects .Within the lens is a lens display 601 . A display projector 620 is provided thereon to project images on the lens display 601 . The lens display 601 can be a single unit or multiple unit lens. There are infrared sensors 602 as well as a directional illumination unit 603 on the bridge of the architecture 560. There are facial and mouth movement sensors 604 and or cameras located on the lens holder of the architecture 560. There is a speaker and n extendable speaker 610 located on the frame when worn. The speaker 610 could be held in place with a head band. An outer ear
  • a control communication unit 608 is utilized to control the architecture 560.
  • a power unit can be utilized to enable the architecture.
  • the power unit 613 comprises a rechargeable battery.
  • the battery can be charged via a connector, such as but not limited to an USB connector to a charging device laptop, tablet or desktop PC for example.
  • the device could be solar powered either by solar cells being placed on the device or the solar cells could be placed on articles of clothing (i.e. hat, shirt or pants for example) to facilitate the charging thereof.
  • the architecture 560 includes a directional illumination unit 603, smell sensors 605 and an extendable user microphone 619.
  • the sensors may comprise any or any combination of gyroscopes, accelerometers, torque sensors, weight sensors, pressure sensors, magnetometers, temperature sensors, light sensor, cameras and microphones, GPS, wireless detection, altitude sensors, blood pressure , heart rate sensors, biometric sensors, radio frequency identification (RFID), near field communication (NFC), mobile communication, Wi-Fi, strain gauges, fingerprint sensors, smell sensors gas sensors, chemical sensors, color sensors, sound sensors, acoustic sensors, ultraviolet sensors, electric field sensors, magnetic field sensors, gravity sensors, wind speed sensors, wind direction sensors, compass sensors, geo- locator sensor, polarized light sensors, infrared emitter sensors.
  • RFID radio frequency identification
  • NFC near field communication
  • Wi-Fi strain gauges
  • fingerprint sensors smell sensors gas sensors, chemical sensors, color sensors, sound sensors, acoustic sensors, ultraviolet sensors, electric field sensors, magnetic field sensors, gravity sensors, wind speed sensors, wind direction sensors, compass sensors, geo- locator sensor, polarized light sensors, infrared emitter sensors
  • the frame color or appearance can be controlled and tailored to the user through the use of electrochromatic, transparent LCD, LCD, OLED, and similar materials.
  • the outside appearance of the lenses can be controlled and tailored to the user using images, graphics, or video on the surface of the lenses.
  • This architecture can be utilized with a conventional mobile operating system such as Android or IOS or with a new operating system incorporating perceptual parameters for even further capabilities and enhanced perception - eye optical or perceptual operating system (eyePOS).
  • apps a whole new class of custom applications
  • apps can be created using the standard mobile operating systems or eyePOS and an eyePOS simulator to address manifold valuable applications that can improve human learning, entertainment, and health on one side to new navigation systems (physically linked and search linked) and enhanced perception.
  • a method and system in accordance with an embodiment comprises utilizing dynamic eye tracking with a wearable optics device; wherein parameters are personalized to a user.
  • the method and system which includes providing an enhancement utilizing objective and subjective quality standards based upon perceptual parameters.
  • the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
  • the wearable optics device controls any or any combination of mimics, amplifies, or expands a user perceptual physiology utilizing perceptual parameters.
  • the wearable optics device can include one or more inserts into eyeglasses.
  • the eyeglasses comprise quad state eyeglasses.
  • Shading control can be utilized on the wearable optics device.
  • the shading control can be provided by one or more projectors within the wearable optics device.
  • An occlusion effect can be projected on a lens of the wearable optics device.
  • the shading can be provided on a lens of the wearable optics device wherein the surrounding area is occluded or reversed.
  • the shading is provided by a polarized filter.
  • the shading control can be provided by the lenses within the wearable optics device.
  • the shading can be controlled using optical position measurement. Materials that can electrically control any or any combination of chromatic, refractive, diffractive, transparent, reflective properties of the wearable optics device are utilized with the dynamic eye tracking.
  • the lens can be any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic , electroluminescence, eletroreflective, dynamic
  • the wearable optics device comprises an electrochromatic material.
  • one or more elements are utilized within the wearable optics device to provide image information into the eye.
  • the one or more elements include any or any combination of a lens projector, retinal projection.
  • the retinal projection or projector plus prism provide the occlusion.
  • the wearable optics device includes shading control for the eyewear.
  • portions of an image viewed by the wearable optics device can be shaded to control brightness.
  • the lenses of the wearable optics device can be controlled polarizing, transparent OLED, or projection and prism lenses.
  • the parameters my include any or any combination of prescriptions for improving the vision of a user, a zoom feature, a microscope feature, magnifying feature, retinal projection feature.
  • the wearable optics device can be utilized in a simulator. In an embodiment, a focal of the wearable optics device is utilized.
  • the parameters can include any or any combination of a zoom feature, a microscope feature, magnifying feature, illumination feature; a retinal projection feature.
  • a 360 degree view can be provided. The 360 degree view can be any or any combination of a left or right panning, up and down panning, three dimensional rotations.
  • an illumination feature is directed to a specific area based upon eye position measurement.
  • a wearable optics device camera feature can filter certain light waves for controlled viewing or visual effects.
  • the filtering feature can include controlling noise reduction, polarization, and creative effects.
  • the wearable optics device feature can include controlling a stability control for facial or object focus.
  • An embodiment may include detecting head movement.
  • An acoustic wave mechanism may be utilized within the wearable optics device.
  • a brain wave mechanism may be utilized within the wearable optics device.
  • a magnetic wave mechanism may be utilized within the wearable optics device.
  • the wearable optics device can be utilized in a variety environments including but not limited to athletic, gaming, gambling, educational, military, firefighting, medical dental, and the like.
  • FIG. 6 illustrates a social networking application 700 utilized with the wearable optics device.
  • the networks of Facebook, Linked In, Twitter, Salesforce.com, and other networks, as well as the Internet are connected to the wearable optics device.
  • Individuals that are "Friends" for example can be identified by a highlight by the wearable optics device.
  • Information about individuals can be gathered by using eyes utilized by the wearable optics device architecture.
  • the individual can be selected.
  • the individual can be identified in a variety of ways for example using facial recognition, target individual information, GPS, RFID, NFC, optical information, voice recognition, and mobile location.
  • FIG. 7 illustrates a messaging application 800 utilized with wearable optics device in accordance with an embodiment.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device can utilize R2R, NFC, Wi-Fi, Internet to communicate. It is possible to talk using a microphone, sensors near the face, jaw, and nose can be utilized to provide control of the messaging application. In addition lip motion, and lip reading can be utilized to transmit voice in a silent and
  • An individual can be targeted by using selected eye
  • Figure 8 illustrates the wearable optics device utilized by an athletic sports spectator in accordance with an embodiment 900.
  • Networks such as Twitter, Facebook, Internet are connected to the spectator. For example the spectator can see who has the ball and its course during a play. Who has the ball as well as the ball's location is highlighted. Video information can be overlayed from scores of other games. Information about the location of the football during the game (line of scrimmage, first down line). Video highlights of the game could be provided as well as augmented reality media.
  • Figure 9 illustrates the wearable optics device utilized by an athletic sports player in accordance with an embodiment 1000. Networks such as Twitter, Facebook, Coach/trainer communication, and other player communications are connected to the player. For example the spectator can see that a curve ball is hit at 102 mph. The trajectory of the ball is highlighted.
  • Figure 10 illustrates an augmented reality information, navigation, and advertising application 1 100 utilized with the wearable optics device.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
  • the wearable optics device is utilized in a vehicle.
  • the wearable optics device includes speaker microphone and rear camera on the headset and also on the rear of a vehicle for example.
  • Augmented reality real time information is provided.
  • the Augmented Reality Real time information provided is that the vehicle is traveling at 62 mph.
  • Augmented Reality Mirror Live Video from Rear Camera of the car.
  • Augmented Reality Mirror Live Video from Rear Camera of the car.
  • "Detour 1 Mile” is shown on as an emergency Augmented Reality sign from State/Federal Sources which could also provide additional information.
  • Augmented Reality real-time advertisement "Stage Road 1 Mile", will also be seen as an Augmented Reality Sign while the voice message “Next turn Stage Rd. 1 mile” is transmitted to the driver together comprising an enhanced Augmented Reality GPS and navigation system.
  • the geometry of the displayed sign can be adjusted in real-time taking into account the location, position, or velocity of the augmented reality object and the location, position, or velocity of the user, and well as that of the user's head position and angle, and eye position measurements. So an augmented reality object moving from a distance will increase in apparent size and will be see from different angles and hence the object's apparent geometries will be changed depending on the user's position, angle, and velocity.
  • augmented reality sound eminating from the object the user perceived and apparent sound level, sound direction can be adjusted along with any associated Doppler effect given the position, velocity, and angle or the user and the object. Additional sound effects such as echo, reverberation, and the like can also be added to increase the user's perception of the object in the context of the user and also the object and user's surroundings (in a tunnel, amphitheater or the like.)
  • the sound apparent direction of sound from the augmented reality object can be implemented by control of the sound through the eyewear stereo speakers or ear buds and which can also implement sound perception systems including Dolby surround sound or other similar systems. Shading or light direction effects or algorithms can also be added to increase the realism of the object relative to ambient or augmented reality light sources.
  • Figure 1 1 illustrates an augmented reality information patient data application 1200 utilized with the wearable optics device used in conjunction with a remote device.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
  • Patient records and internet technical information are connected to the eyepiece and microphone of the person who is utilizing the wearable optics device. Utilizes an augmented reality zoom window to identify medical feature. Augmented reality patient data is made available to the person via the eyewear. There may also be a remote device camera utilized on the drill of a dentist for example. The dentist for example can utilize the dynamic eye tracking mechanism to focus on the correct tooth.
  • An overlay of the x-ray of the tooth can be seen utilizing the augmented reality.
  • An augmented reality overlay of dental records and Internet research in tooth treatment is available. Dentist can use a remote drill with augmented reality. Illumination and zoom can also be utilized with an augmented reality window.
  • Figure 12 illustrates a shading control application 1300 utilized with the wearable optics device.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
  • Shading settings can be chosen through push buttons on the eyewear frame, via eye movement, or automatically.
  • the shading can be uniform across the eyewear lens or concentrated in a specific area or areas of the lens.
  • a lamp/flashlight 1302 projects light to eye1310.
  • the camera 1306 and eye sensor 1308 pick up the light.
  • the lens 1304 can be any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic ,
  • Light can be occluded in a specific area 1312 utilizing the wearable optics device.
  • the camera 1306 determines position of light to be occluded (real time).
  • the eye position sensor or camera 1308 determines the position of the eye (real time).
  • the camera1306 /eye position sensor 1308 determines line of sight between light to be occluded and eye 1310 and intersect area on lens 1304 (real time) or area to project occlusion from a projector embodiment.
  • Figure 13 illustrates an augmented reality application 1400 utilized with the wearable optics device 1410.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device 1410 can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
  • an augmented reality keyboard 1404 appears selected by look at the phone/item and then blinking or the like.
  • the augmented reality (AR) keyboard 1404 is utilized that is controlled by the dynamic eye tracking mechanism.
  • An infrared camera 1402 is used to sense the position of any of the user's hand, hand movement, finger position, finger movement on the AR keyboard such as key highlight and key click sound.
  • There is also an augmented reality keyboard which is shown as being on the lens.
  • Figure 14 illustrates a physical gaming application 1500 utilized with the wearable optics device 1510.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device 1510 can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
  • a person wearing the wearable optics device 1510 can analyze game strategy, count cards, determine the score, do analysis (of game statistics), and analyze other player's faces utilizing an augmented reality overlay 1502 and facial recognition.
  • Figure 15 illustrates a first embodiment of an online/mobile gaming
  • the wearable optics device 1610 can utilize mobile, R2R, NFC, Wi- Fi, Internet to communicate.
  • the player and opponent have augmented reality cards in hand. Augmented reality playing cards are utilized. Because of the augmented reality and
  • the AR cards may be lifted by hand movement.
  • an infrared camera 1602 on the glasses to measure and judge hand and finger position and movement.
  • an augmented reality scene or dynamics overlay 1612 which can be seen on the lenses.
  • Figure 16 illustrates a second embodiment of an online/mobile gaming application 1700 utilized with the wearable optics device 1710.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device 1710 can utilize mobile, R2R, NFC, Wi- Fi, Internet to communicate.
  • the scene that the player sees can be an actual real-world video game screen. It could also be utilized as an augmented reality video game screen (e.g. for mobile). Furthermore it could also be utilized as a full 3-D real time Augmented Reality game/battle field which the player sees.
  • the player can use an augmented reality game controller. There is an infrared camera on the glasses to measure and judge hand and finger position and movement on the AR game controller.
  • Augmented reality scenes, AR game controller, AR gun, or AR remote control overlay 1712 are seen on the lenses of the glasses.
  • Figure 17 illustrates shading control mechanism utilizing the wearable optics device.
  • the processing system 1802 can control a display on the lens 1804 to provide shading.
  • Figure 18 illustrates an optical/perceptual operating system 1900 with the wearable optics device. As is seen a plurality of applications 1902-1910 interface with a processing system.
  • the processing system includes a CPU, Memory, computer control and a CPU updating system 1912.
  • the applications include but are not limited to an eye prescription 1902, shading 1904, a glare application, 1906, GPS for navigation of a vehicle 1908 and a medical application 1910.
  • the system would include perceptual
  • measurement/generators 1914 These would include but are not limited to calibration software, augmented reality software, entertainment software, video/audio conferencing software and external communication/databases.
  • the system would also include one or more device drivers. They include but are not limited to display drivers 1916, optical/sensor drivers 1918, operating system drivers 1920, network drivers 1922, external remote object drivers 1924 and
  • Figure 19 describes an embodiment of the digital architecture of a wearable optics device 2000.
  • the wearable optics device eyewear includes mobile/smart phone circuitry / external data communication / and circuitry for transmitting data via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet, along with networking to external networks platforms / data / media sites 2016 .
  • the wearable optics contains a processing system 2002 with memory storage, sensors for eye position and perceptual measurement 2006, circuitry to control the optical display and perceptual generation needs of the device 2004, and interface 2008 to remote devices such as tools, specialty camera, GPS, mobile phone, wearable devices and the like.
  • various types of application software (“apps”) 2018 can be run on the wearable optics device 2000 including shading control applications, focus and user eye adjustment prescriptions, and augmented reality applications for games such as football.
  • An Internet browser 2010 that utilizes eye position or perceptual parameters to drive navigation of the Internet can be used such that eye movements or facial expressions can accelerate the browsing process to the desired information.
  • the wearable optics device 2000 contains a system browser 2012 with file storage that can be on the device or accessed via one of the networks to the device.
  • the device 2000 can be powered by a separate battery (not shown).
  • the battery can be charged via a connector, such as but not limited to an USB
  • the device 200 could be solar powered either by solar cells being placed on the device 2000 or the solar cells could be placed on articles of clothing (i.e. hat, shirt or pants for example) to facilitate the charging thereof.
  • Figure 20 illustrates the embodiment of a system simulator 2100 for use by developers of applications and new lenses or expansion of the wearable optics device.
  • a simulator for the operating system 2102 there is a simulator for the operating system 2102, a lens simulator 2104, a display 21 14, and an eyewear emulator 21 16.
  • position/perceptual measurements 2106, camera signals, and other sensors and measurements are inputs to the simulator.
  • the developer apps or new lenses can then be tested for various types of wearable options with various types of operating systems including iOS, Andriod, and general purpose or optimized
  • Figure 21 A through Figure 21 F illustrate an embodiment of inverse shading using the wearable optics device.
  • Figure 21 A illustrates the problem of glare caused by ambient light which degrades the visibility of a object such as the screen of a mobile phone or laptop.
  • Figure 21 C describes the iris / pupil contraction due to brightness which degrades retina/cornea and brain view of target object such as a phone screen or the like.
  • Figure 21 E the phone screen appears dark since ambient light is far brighter than screen.
  • Figure 21 B illustrates the selection of the target object as in a phone screen via eye or automatically by preferences, rules, camera image capture and object recognition.
  • Figure 21 D shows eye position measurement and capture of the object's position and image by a camera on the eyewear.
  • Figure 21 F shows the resulting elimination or reduction in glare and increase in visibility of the object wherein a shaded or translucent background follows surrounding area object in real time as seen from the user of the wearable optics device.
  • Figure 22 illustrates an embodiment of eye position illumination and enhanced efficiency utilizing the wearable optics device.
  • the line of sight and focal length can be determined and used to control a directional illumination source such that the illumination source illuminates the area corresponding to the area being focused on by the user of the wearable optics device
  • Figure 23 illustrates an embodiment of real-time augmented reality overlay 2400 utilizing the wearable optics device.
  • information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet.
  • the wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
  • the wearable optics device is utilized in a vehicle.
  • the driver's eyewear uses augmented reality to overlay advertising signs which can be personalized to the user which appear as if they are normal roadside billboards.
  • a real-time updated augmented reality danger sign is posted before a road hazard with the augmented reality sign being generated in real-time using information from the networks connected to the wearable eyewear device.
  • This example also shows real-time translation from English to Spanish of navigational warning and advertising signs using augmented reality with the wearable optics device. Accordingly systems and methods in accordance with embodiments are disclosed that provide these enhanced features for wearable optics devices. To describe these features and embodiments in more detail refer now to the following description in conjunction with the following discussion.
  • a key feature associated with these enhancements is providing a variety of perceptual parameters that can be utilized with the wearable optics devices. Examples of perceptual parameters include but are not limited to optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.

Abstract

Improved wearable optics is disclosed. The wearable optics comprises a frame member and a lens. The wearable optics also includes circuitry within the frame member for enhancing the use of the wearable optics. A system and method in accordance with the present invention is directed to a variety of ways to enhance the use of eyeglasses. They are: (1 ) media focals, that is, utilizing the wearable optics for its intended purpose and enhancing that use by using imaging techniques to improve the vision of the user; (2) telecommunications enhancements that allow the eyeglasses to be integrated with telecommunication devices such as cell phones or the like; and (3) entertainment enhancements that allow the wearable optics to be integrated with devices such as MP3 players, radios, or the like.

Description

ENHANCED OPTICAL AND PERCEPTUAL DIGITAL EYEWEAR
CROSS-REFERENCE TO RELATED APPLICATION
This application claims benefit under 35 USC 1 19(e) of U.S. Provisional Patent Application No. 61 /798,770, filed on March 15, 2013, entitled "ENHANCED DIGITAL EYEWEAR," which is incorporated herein by reference in its entirety.
This application is a continuation-in-part and claims the benefit of priority to U.S. Patent Application No. 13/841 ,550, filed on March 15, 2013, entitled
"ENHANCED OPTICAL AND PERCEPTUAL DIGITAL EYEWEAR", and U.S.
Patent Application No. 13/841 ,141 , filed on March 15, 2013, entitled "ENHANCED OPTICAL AND PERCEPTUAL DIGITAL EYEWEAR", which is a continuation-in-part and claims the benefit of priority to U.S. Patent Application No. 13/739,929, filed on January 1 1 , 2013, entitled "DIGITAL EYEWEAR", which is a continuation
application of U.S. Patent Application No. 13/078,589, filed on April 1 , 201 1 , entitled "DIGITAL EYEWEAR", now U.S. Patent No. 8,353,594, issued on January 15,
2013, which is a continuation application of U.S. Patent Application No. 12/621 ,423, filed on November 18, 2009, entitled "DIGITAL EYEWEAR", now U.S. Patent No. 7,918,556, issued on April 5, 201 1 , which is a continuation application of U.S.
Patent Application No. 12/029,068, filed February 1 1 , 2008, entitled "DIGITAL EYEWEAR", now U.S. Patent No. 7,758,185, issued on July 20, 2010, which is a divisional application of U.S. Patent Application No. 1 1/245,756, filed October 7, 2005, entitled "DIGITAL EYEWEAR", all of which are incorporated herein by reference. This application also claims benefit under 35 USC 1 19(e) of U.S. Provisional Patent Application No. 61 /798,770, filed on March 15, 2013, entitled "ENHANCED DIGITAL EYEWEAR," which is also incorporated herein by reference in its entirety.
FIELD OF THE INVENTION
The present invention relates generally to wearable optics and more particularly to wearable optics that includes additional functionality.
BACKGROUND OF THE INVENTION
Wearable optics is utilized for a variety of purposes. Wearable optics is used for improving one's vision for reading glasses and to protect one's vision.
Oftentimes protective goggles are used to protect eyes within dangerous areas. It is desirable to add additional functionality to glasses. This functionality can include a variety of forms, which are electronic, mechanical, aesthetic, etc. Accordingly, it is always desired to provide additional functionality to wearable optics. What is desired is a system and method which will increase the functionality of glasses beyond their normal use while still maintaining them for their primary uses. The present invention addresses such a need.
SUMMARY OF THE INVENTION
A wearable optics device and method of use is disclosed. In a first aspect, a method comprises a wearable optics device, wherein parameters personalized to a user can be provided based upon user information. In a second aspect, a wearable optics device comprises a lens, wherein parameters personalized to a user can be provided based upon user information.
BRIEF DESCRIPTION OF THE DRAWINGS
Figure 1 is a diagram that illustrates Media focals.
Figure 2 comprises an information bar on media focal wearable optics.
Figure 3 is a block diagram of wearable optics that is utilized in a music environment such as an MP3 player.
Figure 4 is a block diagram that illustrates wearable optics that is utilized as a cell phone.
Figure 5A is a block diagram that illustrates the cellular phone circuitry of Figure 4.
Figure 5B illustrates perceptual optimization utilizing optical and perceptual parameters.
Figure 5C illustrates the enhanced digital eyewear architecture.
Figure 6 illustrates a social networking application utilized with the wearable optics device.
Figure 7 illustrates a messaging application utilized with wearable optics device in accordance with an embodiment.
Figure 8 illustrates the wearable optics device utilized by an athletic sports spectator in accordance with an embodiment.
Figure 9 illustrates the wearable optics device utilized by an athletic sports player in accordance with an embodiment. Figure 10 illustrates an augmented reality information, navigation, and advertising application utilized with the wearable optics device.
Figure 1 1 illustrates an augmented reality information patient data application utilized with the wearable optics device used in conjunction with a remote device.
Figure 12 illustrates a shading control application utilized with the wearable optics device.
Figure 13 illustrates an augmented reality application utilized with the wearable optics device.
Figure 14 illustrates a physical gaming application utilized with the wearable optics device.
Figure 15 illustrates a first embodiment of an online/mobile gaming application utilized with the wearable optics device.
Figure 16 illustrates a second embodiment of an online/mobile gaming application utilized with the wearable optics device.
Figure 17 illustrates shading control utilizing the wearable optics device.
Figure 18 illustrates an optical/perceptual operating system with the wearable optics device.
Figure 19 describes an embodiment of the digital architecture of the wearable optics device.
Figure 20 illustrates the embodiment of a system simulator for use by developers of applications and new lenses or expansion of the wearable optics device.
Figure 21 A thru Figure 21 F illustrate the embodiment of inverse shading using the wearable optics device. Figure 22 illustrates an embodiment of eye tracking illumination and enhanced efficiency utilizing the wearable optics device.
Figure 23 illustrates an embodiment of real-time augmented reality overlay utilizing the wearable optics device.
DETAILED DESCRIPTION
The present invention relates generally to wearable optics and more particularly to wearable optics that includes additional functionality. The following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements. Various modifications to the preferred embodiments and the generic principles and features described herein will be readily apparent to those skilled in the art. Thus, the present invention is not intended to be limited to the
embodiments shown, but is to be accorded the widest scope consistent with the principles and features described herein.
A system and method in accordance with the present invention is directed to a variety of ways to enhance the use of wearable optics devices.
To describe the features of the present invention in more detail refer now to the following description in conjunction with the accompanying figures.
1 . Media focals 100
Figure 1 is a diagram that illustrates Media focals 100. Media focals 100 comprises an information bar 102, receiver 104, digital circuitry 106, frames 108 and lens 1 10. Media focals 100 allow for enhancing the wearable optics for its primary purpose, for example, a digital camera could be placed within the wearable optics to allow for seeing certain of these images. For example, the circuitry 106 for the media focals 100 could be placed within the frame 108 of the wearable optics. The lens 1 10 could have a totally reflective surface, or a partially reflective surface using LCDs or the like. In effect the wearable optics could look like see-through glasses, but through the use of the circuitry 106 within the wearable optics it is actually a media focal. Additionally, the wearable optics could incorporate a camera to project the user onto a second lens to achieve a see-through effect.
In a preferred embodiment, an information bar 102 is provided across a portion of the wearable optics which is visible to the user. This information bar 102 is used to convey a variety of types of information.
Figure 2 comprises an information bar 102' on media focal wearable optics. The information bar 102' can be a stock ticker scrolling across the top portion of the wearable optics, as is shown in Figure 2. Although the information bar 102' is shown displaying a stock ticker, other kinds of information such as song titles, lyrics and the like could be displayed in the information bar. This information bar is referred to as E-focals. This information might be provided from a digital receiver through an FM station, through a cellular wireless device, or an MP3 player.
Additional functionality of the E-focal will be described with more detail with respect to the cell phone enhancements as well as the music player enhancements.
One of the key features of the media focals 100 is the use of the media focals to enhance the primary function of the user, that is, being able to more accurately and clearly see the objects. In such an environment, for example, it is possible to have a zoom feature circuit to allow for the use of the wearable optics as binoculars. This would allow for the user to see objects more closely based on certain activities of the user. For example, there may be eye or pressure sensors on the wearable optics that will activate the binocular circuitry in the glasses which could receive visual data through a camera, CCD receiver of the like.
In the preferred embodiment, the circuitry 106 would be located somewhere in the frame of the glasses to provide this functionality and as circuits became smaller and devices became smaller it would be easier and easier to embed the circuitry that is well known for use for such functions directly within the device. The circuitry 106 in the device could be, for example, eye sensors which could be pressure sensors, capacitive sensors or some other type of sensor for allowing the eyes to direct the activities. Eye movement sensors, for example, could be used to activate and control the binocular glasses. Similarly, a digital camera could be put on the glasses that would allow the same kinds of technology to take pictures by the person directly.
In a similar vein, the glasses could be used as a normal corrective lens glass utilizing the digital imagery, so that, for example, a user has a certain prescription that they use with their normal prescription glasses to view an object clearly. As the user's eyes change, it would be possible that an optometrist could download the new prescription to the wearable optics such that a digital transformation of the image information is provided which is compatible with the new prescription.
Also, in a preferred embodiment a method for sensing and controlling the digital media could be implemented in a variety of ways. For example, an activity of the eye itself would control the activity of the media focal. So, for example, if the idea was to zoom the image, the eye would blink twice. It would also be possible to detect facial and eye movements (squinting, for example), as well as changes in the pupil and iris.
In a further embodiment, it would be possible for the eyeglasses in accordance with the present invention to function within a client/server model or Bluetooth (Wi-Fi) model. Utilization of the client/server model and Bluetooth Wi-Fi would make possible, for example, the display of live news or special reports (such as financial reports) from the Internet or similar sources on the eyeglasses. This would also allow for portions of circuitry to be located remotely such that less circuitry in the wearable optics is required.
The wearable optics could also include a logo, for example, law enforcement officers could have their glasses emblazoned with "Police", "Sheriff", "MP", etc.; young people could have their eyeglasses emblazoned with words and images that reflected their favorite performers, etc.; sports teams could offer the eyeglasses at discount with team monograms, etc. They could also be purchased by companies, emblazoned with the company logos, and given out as retirement gifts, etc.
2. Music Environment
Figure 3 is a block diagram of wearable optics 300 that is utilized in a music environment such as an MP3 player. Figure 3 comprises wearable optics 300, an information bar 302, MP3 player circuitry 304, storage 306, frames 308, and one or a plurality of lenses 310. Another environment as has been above described is the music environment. What would be desirable would be to provide music glasses in which an MP3 player on an IPod or the like is incorporated in the wearable optics, either in a wired or wireless environment. Through the use of this type of system, a plurality of users could be networked via an MP3 player type environment within a hotspot, or the like, which would allow one to have downloads of whatever music is required through the eyeglasses. The system could allow for downloadable music which could be selected via scrolling and the like through voice recognition systems.
By connecting to a client-server network or Bluetooth Wi-Fi installation, for example, the eyeglasses could link to a multimedia network, authorize downloading and billing for selected music. By this means, access to a plurality of libraries for music selections could be provided.
It would also be possible to provide access to streaming audio media. Also, access can be provided to multimedia libraries, etc., via the client/server model.
Information could be received via a digital client/server model enabled to work with iPods or MP3 players. Similarly, bluetooth wireless technology could be utilized to provide access to music and live audio sources.
The wearable optics could also be utilized in conjunction with wireless technology to allow a user or a plurality of users to participate simultaneously in single or group karaoke singing. The wearable optics could be used specifically to display the lyrics of a song, melody, notes, name of the song or other associated references.
It would also be possible to receive and listen to AM or FM radio signals, via an AM/FM radio tuner connected to the wearable optics hardware.
In this type of environment, the headphones can be either digital or analog. The user doesn't need to have 10,000 songs, for example. They can come enrolled in an in-song virtual network library upon entering a hotspot. Therefore, the local storage 306 could be limited. In addition, this would provide location identity information for one who is using the network. The songs can be streamed as well as downloaded. The songs could be purchase using the wearable optics. The system could be scalable; depending upon what kind of device was being used. 3. Telecommunications Environment
Figure 4 is a block diagram that illustrates wearable optics that is utilized as a cell phone 400. Figure 4 comprises cellular phone circuitry 402, a microphone 104, frames 408 and one or a plurality of lenses 410. The cell phone wearable optics 400 could be implemented utilizing digital telephone technology. Circuitry 402 within the wearable optics could be utilized to allow a telephone number or other visual information such as that provided by multimedia messaging services to be displayed on the lens 410 of the wearable optics as shown in Figure 3. Figure 5 is a block diagram that illustrates the cellular phone circuitry of Figure 4. Figure 5 comprises noise cancelling circuitry 502, voice recognition circuitry 504, caller ID circuitry 506 and speaker recognition circuitry 508 and media processing circuits 509. The telephone number could be activated via the digital circuitry 402 as part of the media focals 100. In addition, the circuitry could be made truly digital via a digital signal processor which is coupled to a camera otherwise in the environment. The above system would allow for voice recording through use of a microphone 104 and would allow for voice recognition through use of the voice recognition circuitry 504, which would allow for signal conditioning on the cell phone in a variety of ways.
The cell phone environment 402 provides a plurality of areas for
improvement utilizing existing technologies. Firstly, one of the major annoyances in cell phone use is that the users have to speak in a loud manner because of background noise and the like. There are a variety of reasons for this problem including the placement of the microphone of the cell phone relative to the speaker's mouth, due to the aforementioned background noise, and other issues. By placing the microphone 104 strategically on the wearable optics such as near the noise or mouth the user will not have to speak as loudly. The microphone could also be located in flip down microphones. In addition noise canceling circuitry 502 could be utilized to remove the background noise. The microphone capability would include the advantage of utilizing noise rejection techniques. Buttons located on the wearable optics can be utilized to control features thereon. Finally, the microphone 104 could utilize whisper technology such that the speaker will not have to speak as loudly.
The wearable optics would in a preferred embodiment include voice recognition circuitry 504 and caller ID circuitry 506. The conventionality for hearing and talking in a preferred embodiment would be located in ear and nose pad portions of glasses. Referring back to Figure 3, the electronics for the cell phone in a preferred embodiment would be within the frame 308 of the wearable optics. In addition the wearable optics would include a fully integrated information bar 302. Finally, a speaker recognition algorithm 508 as shown in Figure 5 would allow only the voice of the user to be recognized and the background noise would be cancelled. Accordingly, the unique characteristics of the speaker are provided via an audible model.
This can performed utilizing a variety of methods. For example analyzing the voice of the user and combining the analysis with noise cancellation. In another example the user can talk softly and cancel noise and a directional microphone is used which takes advantage of device location.
Similar to the media focal and MP3 player environments, a digital client/server or Bluetooth/wifi model could be adapted to link the wearable optics to external communication equipment. Such equipment could include digital cell phones, PDAs or wifi enabled PCs or other devices. Such an embodiment could enable review of voicemail, screen viewed emails, text to speech audio email conversions, multimedia messaging services, and other data sources.
Wireless or Bluetooth interconnection could also make possible VOIP glasses to be utilized instead of a cell phone. Other features enabled by a wireless link could link the eyewear to MP3 devices, an iPod, a printer, wireless/wired TV, coupons, and the like. Also "PDA glasses" could provide built in a time display, alarm calendar, interfacing with PCs or network sources, a speaker and the like.
As can be seen from the above description, digital eyewear is a rapidly evolving field with from the early innovation of digital eyewear with eye tracking capabilities thru Lewis ("185 filed Feb 2008), to eyewear with more complex lenses and communication/display capabilities (Lewis '556, filed Nov 2009), to more enhancements and capabilities (Lewis '594, filed April 201 1 ). As technology progresses to make sensors, cameras, processors, and circuitry smaller, more and more capabilities become possible to implement using digital eyewear. This enhanced digital eyewear can be used to solve important areas ranging from superior vision enhancement and mobile advertising, to use in dental/medical procedures and physical and Internet navigation. The application and value of the enhanced eyewear is increased even further when combined with augmented reality, social networking, messaging, and communications.
With the introduction and use of new materials for lens and filter integration new enhancements and capabilities of the digital eyewear can be further realized. These materials include advances in OLED, LED, transparent LED, flexible LED, crystalline, prism, holographic, polarizing, and translucent material and the like to electrorefractive, electrodiffractive, electroreflective, composite refractive materials and the like, with one or more lens material layers and passive or active display or projection based implementations.
With these new capabilities an important new set of optical and perceptual parameters can be measured and used as inputs, controls, or feedback elements that increase even further uses and value of the eyewear and lead to important improvements that can be used for teaching, sports, health, and improved perception.
Accordingly systems and methods in accordance with embodiments are disclosed that provide these enhanced features for wearable optics devices. To describe these features and embodiments in more detail refer now to the following description in conjunction with the following discussion. A key feature associated with these enhancements is providing a variety of perceptual parameters that can be utilized with the wearable optics devices. Examples of perceptual parameters include but are not limited to optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive. Figure 5B illustrates a perceptual optimization system 550. The perceptual optimization system 550 receives a variety of inputs including real world inputs, digital media inputs, expression parameter measurements, other parameter measurements to provide wearable optics visual display elements. Real world inputs could be for example inputs from one or more microphones or cameras. Digital media inputs could be for example from digital audio, digital video, graphics, images and augmented reality.
Other perceptual parameters could be for example, smell, touch, brainwave, temperature/humidity of the user, environmental conditions near the user.
Figure 5C illustrates a wearable optics device architecture 560 in accordance with an embodiment. The architecture includes a frame 562 which includes a plurality of sensors on various areas thereon. Biometric sensors include a blood pressure sensor 617, temperature sensor 618, EEG sensor 616 and the like.
Environmental sensors 615 are also provided. There are microphone sensors 606,607, 61 1 on various areas of the frame. Included on the frame 562 are cameras rear, front and side 606,607, 61 1 to detect objects .Within the lens is a lens display 601 . A display projector 620 is provided thereon to project images on the lens display 601 . The lens display 601 can be a single unit or multiple unit lens. There are infrared sensors 602 as well as a directional illumination unit 603 on the bridge of the architecture 560. There are facial and mouth movement sensors 604 and or cameras located on the lens holder of the architecture 560. There is a speaker and n extendable speaker 610 located on the frame when worn. The speaker 610 could be held in place with a head band. An outer ear
speaker//vibration element 612 is provided thereon. A control communication unit 608 is utilized to control the architecture 560. A power unit can be utilized to enable the architecture. Typically the power unit 613 comprises a rechargeable battery. The battery can be charged via a connector, such as but not limited to an USB connector to a charging device laptop, tablet or desktop PC for example. In addition the device could be solar powered either by solar cells being placed on the device or the solar cells could be placed on articles of clothing (i.e. hat, shirt or pants for example) to facilitate the charging thereof. The architecture 560 includes a directional illumination unit 603, smell sensors 605 and an extendable user microphone 619.
In an embodiment, the sensors may comprise any or any combination of gyroscopes, accelerometers, torque sensors, weight sensors, pressure sensors, magnetometers, temperature sensors, light sensor, cameras and microphones, GPS, wireless detection, altitude sensors, blood pressure , heart rate sensors, biometric sensors, radio frequency identification (RFID), near field communication (NFC), mobile communication, Wi-Fi, strain gauges, fingerprint sensors, smell sensors gas sensors, chemical sensors, color sensors, sound sensors, acoustic sensors, ultraviolet sensors, electric field sensors, magnetic field sensors, gravity sensors, wind speed sensors, wind direction sensors, compass sensors, geo- locator sensor, polarized light sensors, infrared emitter sensors.
The frame color or appearance can be controlled and tailored to the user through the use of electrochromatic, transparent LCD, LCD, OLED, and similar materials. Similarly the outside appearance of the lenses can be controlled and tailored to the user using images, graphics, or video on the surface of the lenses. This architecture can be utilized with a conventional mobile operating system such as Android or IOS or with a new operating system incorporating perceptual parameters for even further capabilities and enhanced perception - eye optical or perceptual operating system (eyePOS). By using this approach and capability set, a whole new class of custom applications ("apps") can be created using the standard mobile operating systems or eyePOS and an eyePOS simulator to address manifold valuable applications that can improve human learning, entertainment, and health on one side to new navigation systems (physically linked and search linked) and enhanced perception. To describe these feature in more detail refer now to the following description.
A method and system in accordance with an embodiment comprises utilizing dynamic eye tracking with a wearable optics device; wherein parameters are personalized to a user. The method and system which includes providing an enhancement utilizing objective and subjective quality standards based upon perceptual parameters. The perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive. The wearable optics device controls any or any combination of mimics, amplifies, or expands a user perceptual physiology utilizing perceptual parameters.
The wearable optics device can include one or more inserts into eyeglasses.
The eyeglasses comprise quad state eyeglasses. Shading control can be utilized on the wearable optics device. The shading control can be provided by one or more projectors within the wearable optics device. An occlusion effect can be projected on a lens of the wearable optics device. The shading can be provided on a lens of the wearable optics device wherein the surrounding area is occluded or reversed. The shading is provided by a polarized filter. The shading control can be provided by the lenses within the wearable optics device. The shading can be controlled using optical position measurement. Materials that can electrically control any or any combination of chromatic, refractive, diffractive, transparent, reflective properties of the wearable optics device are utilized with the dynamic eye tracking. The lens can be any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic , electroluminescence, eletroreflective, dynamic filtering materials.
The wearable optics device comprises an electrochromatic material. In a system and method in accordance with an embodiment one or more elements are utilized within the wearable optics device to provide image information into the eye. The one or more elements include any or any combination of a lens projector, retinal projection. The retinal projection or projector plus prism provide the occlusion.
The wearable optics device includes shading control for the eyewear. In the wearable optics device, portions of an image viewed by the wearable optics device can be shaded to control brightness. The lenses of the wearable optics device can be controlled polarizing, transparent OLED, or projection and prism lenses.
The parameters my include any or any combination of prescriptions for improving the vision of a user, a zoom feature, a microscope feature, magnifying feature, retinal projection feature. The wearable optics device can be utilized in a simulator. In an embodiment, a focal of the wearable optics device is utilized. The parameters can include any or any combination of a zoom feature, a microscope feature, magnifying feature, illumination feature; a retinal projection feature. In an embodiment a 360 degree view can be provided. The 360 degree view can be any or any combination of a left or right panning, up and down panning, three dimensional rotations.
In another embodiment, an illumination feature is directed to a specific area based upon eye position measurement. A wearable optics device camera feature can filter certain light waves for controlled viewing or visual effects. The filtering feature can include controlling noise reduction, polarization, and creative effects. The wearable optics device feature can include controlling a stability control for facial or object focus. An embodiment may include detecting head movement. An acoustic wave mechanism may be utilized within the wearable optics device. A brain wave mechanism may be utilized within the wearable optics device. A magnetic wave mechanism may be utilized within the wearable optics device.
The wearable optics device can be utilized in a variety environments including but not limited to athletic, gaming, gambling, educational, military, firefighting, medical dental, and the like. To describe the features of the present invention in more detail refer now to the following description in conjunction with the accompanying figures.
Social networks can be leveraged advantageously with the wearable optics device in accordance with an embodiment. Figure 6 illustrates a social networking application 700 utilized with the wearable optics device. The networks of Facebook, Linked In, Twitter, Salesforce.com, and other networks, as well as the Internet are connected to the wearable optics device. Individuals that are "Friends" for example, can be identified by a highlight by the wearable optics device. Information about individuals can be gathered by using eyes utilized by the wearable optics device architecture. In an embodiment, the individual can be selected. The individual can be identified in a variety of ways for example using facial recognition, target individual information, GPS, RFID, NFC, optical information, voice recognition, and mobile location.
Figure 7 illustrates a messaging application 800 utilized with wearable optics device in accordance with an embodiment. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device can utilize R2R, NFC, Wi-Fi, Internet to communicate. It is possible to talk using a microphone, sensors near the face, jaw, and nose can be utilized to provide control of the messaging application. In addition lip motion, and lip reading can be utilized to transmit voice in a silent and
confidential manner. An individual can be targeted by using selected eye
movements.
Figure 8 illustrates the wearable optics device utilized by an athletic sports spectator in accordance with an embodiment 900. Networks such as Twitter, Facebook, Internet are connected to the spectator. For example the spectator can see who has the ball and its course during a play. Who has the ball as well as the ball's location is highlighted. Video information can be overlayed from scores of other games. Information about the location of the football during the game (line of scrimmage, first down line). Video highlights of the game could be provided as well as augmented reality media. Figure 9 illustrates the wearable optics device utilized by an athletic sports player in accordance with an embodiment 1000. Networks such as Twitter, Facebook, Coach/trainer communication, and other player communications are connected to the player. For example the spectator can see that a curve ball is hit at 102 mph. The trajectory of the ball is highlighted.
Figure 10 illustrates an augmented reality information, navigation, and advertising application 1 100 utilized with the wearable optics device. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate. In one example the wearable optics device is utilized in a vehicle. In this example the wearable optics device includes speaker microphone and rear camera on the headset and also on the rear of a vehicle for example. Augmented reality real time information is provided. For example, the Augmented Reality Real time information provided is that the vehicle is traveling at 62 mph.
There also may be Augmented Reality Mirror Live Video from Rear Camera of the car. For a sign that reads, "Detour 1 Mile" is shown on as an emergency Augmented Reality sign from State/Federal Sources which could also provide additional information.
In another example, "McDonald's Free Coffee" next exit, seen as an
Augmented Reality real-time advertisement. "Stage Road 1 Mile", will also be seen as an Augmented Reality Sign while the voice message "Next turn Stage Rd. 1 mile" is transmitted to the driver together comprising an enhanced Augmented Reality GPS and navigation system. To make the augmented reality safety road signs or advertisements more realistic for the user, the geometry of the displayed sign can be adjusted in real-time taking into account the location, position, or velocity of the augmented reality object and the location, position, or velocity of the user, and well as that of the user's head position and angle, and eye position measurements. So an augmented reality object moving from a distance will increase in apparent size and will be see from different angles and hence the object's apparent geometries will be changed depending on the user's position, angle, and velocity. Similarly video on the object would have it's geometry change in real time to correspond to what would happen with a real world object. Also in the case of augmented reality sound eminating from the object, the user perceived and apparent sound level, sound direction can be adjusted along with any associated Doppler effect given the position, velocity, and angle or the user and the object. Additional sound effects such as echo, reverberation, and the like can also be added to increase the user's perception of the object in the context of the user and also the object and user's surroundings (in a tunnel, amphitheater or the like.) The sound apparent direction of sound from the augmented reality object can be implemented by control of the sound through the eyewear stereo speakers or ear buds and which can also implement sound perception systems including Dolby surround sound or other similar systems. Shading or light direction effects or algorithms can also be added to increase the realism of the object relative to ambient or augmented reality light sources.
Figure 1 1 illustrates an augmented reality information patient data application 1200 utilized with the wearable optics device used in conjunction with a remote device. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
Patient records and internet technical information are connected to the eyepiece and microphone of the person who is utilizing the wearable optics device. Utilizes an augmented reality zoom window to identify medical feature. Augmented reality patient data is made available to the person via the eyewear. There may also be a remote device camera utilized on the drill of a dentist for example. The dentist for example can utilize the dynamic eye tracking mechanism to focus on the correct tooth.
An overlay of the x-ray of the tooth can be seen utilizing the augmented reality. An augmented reality overlay of dental records and Internet research in tooth treatment is available. Dentist can use a remote drill with augmented reality. Illumination and zoom can also be utilized with an augmented reality window.
Figure 12 illustrates a shading control application 1300 utilized with the wearable optics device. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate. Shading settings can be chosen through push buttons on the eyewear frame, via eye movement, or automatically. The shading can be uniform across the eyewear lens or concentrated in a specific area or areas of the lens.
In an embodiment a lamp/flashlight 1302 projects light to eye1310. The camera 1306 and eye sensor 1308 pick up the light. The lens 1304 can be any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic ,
electroluminescence, eletroreflective, dynamic filtering materials.
Light can be occluded in a specific area 1312 utilizing the wearable optics device. The camera 1306 determines position of light to be occluded (real time). The eye position sensor or camera 1308 determines the position of the eye (real time). The camera1306 /eye position sensor 1308 determines line of sight between light to be occluded and eye 1310 and intersect area on lens 1304 (real time) or area to project occlusion from a projector embodiment.
Figure 13 illustrates an augmented reality application 1400 utilized with the wearable optics device 1410. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device 1410 can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
In an embodiment, an augmented reality keyboard 1404 appears selected by look at the phone/item and then blinking or the like. The augmented reality (AR) keyboard 1404 is utilized that is controlled by the dynamic eye tracking mechanism. An infrared camera 1402 is used to sense the position of any of the user's hand, hand movement, finger position, finger movement on the AR keyboard such as key highlight and key click sound. There is an augmented reality display 1406 which is an enlargement of the small phone display, on the lens. There is also an augmented reality keyboard which is shown as being on the lens.
Figure 14 illustrates a physical gaming application 1500 utilized with the wearable optics device 1510. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device 1510 can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate.
In an embodiment, a person wearing the wearable optics device 1510 can analyze game strategy, count cards, determine the score, do analysis (of game statistics), and analyze other player's faces utilizing an augmented reality overlay 1502 and facial recognition.
Figure 15 illustrates a first embodiment of an online/mobile gaming
application 1600 utilized with the wearable optics device 1610. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device 1610 can utilize mobile, R2R, NFC, Wi- Fi, Internet to communicate.
The player and opponent have augmented reality cards in hand. Augmented reality playing cards are utilized. Because of the augmented reality and
communication link the players need not be present in the same location. The AR cards may be lifted by hand movement. There is a physics adjusted augmented reality card movement and dynamics. In an embodiment there can be a virtual game board, a background, and a playing field.
There is an infrared camera 1602 on the glasses to measure and judge hand and finger position and movement. There is an augmented reality scene or dynamics overlay 1612 which can be seen on the lenses.
Figure 16 illustrates a second embodiment of an online/mobile gaming application 1700 utilized with the wearable optics device 1710. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device 1710 can utilize mobile, R2R, NFC, Wi- Fi, Internet to communicate.
The scene that the player sees can be an actual real-world video game screen. It could also be utilized as an augmented reality video game screen (e.g. for mobile). Furthermore it could also be utilized as a full 3-D real time Augmented Reality game/battle field which the player sees. The player can use an augmented reality game controller. There is an infrared camera on the glasses to measure and judge hand and finger position and movement on the AR game controller.
Augmented reality scenes, AR game controller, AR gun, or AR remote control overlay 1712 are seen on the lenses of the glasses.
Figure 17 illustrates shading control mechanism utilizing the wearable optics device. In this embodiment there are one or more cameras 1806 that measures position of the eye and sends the information to a processing system. Thereafter, the processing system 1802 can control a display on the lens 1804 to provide shading.
Figure 18 illustrates an optical/perceptual operating system 1900 with the wearable optics device. As is seen a plurality of applications 1902-1910 interface with a processing system. The processing system includes a CPU, Memory, computer control and a CPU updating system 1912.
The applications include but are not limited to an eye prescription 1902, shading 1904, a glare application, 1906, GPS for navigation of a vehicle 1908 and a medical application 1910. The system would include perceptual
measurement/generators 1914. These would include but are not limited to calibration software, augmented reality software, entertainment software, video/audio conferencing software and external communication/databases. The system would also include one or more device drivers. They include but are not limited to display drivers 1916, optical/sensor drivers 1918, operating system drivers 1920, network drivers 1922, external remote object drivers 1924 and
gaming/entertainment drivers 1926.
Figure 19 describes an embodiment of the digital architecture of a wearable optics device 2000. In this embodiment, the wearable optics device eyewear includes mobile/smart phone circuitry / external data communication / and circuitry for transmitting data via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet, along with networking to external networks platforms / data / media sites 2016 . The wearable optics contains a processing system 2002 with memory storage, sensors for eye position and perceptual measurement 2006, circuitry to control the optical display and perceptual generation needs of the device 2004, and interface 2008 to remote devices such as tools, specialty camera, GPS, mobile phone, wearable devices and the like.
In this embodiment various types of application software ("apps") 2018 can be run on the wearable optics device 2000 including shading control applications, focus and user eye adjustment prescriptions, and augmented reality applications for games such as football. An Internet browser 2010 that utilizes eye position or perceptual parameters to drive navigation of the Internet can be used such that eye movements or facial expressions can accelerate the browsing process to the desired information. The wearable optics device 2000 contains a system browser 2012 with file storage that can be on the device or accessed via one of the networks to the device. The device 2000 can be powered by a separate battery (not shown). The battery can be charged via a connector, such as but not limited to an USB
connector to a charging device laptop, tablet or desktop PC for example. In addition the device 200 could be solar powered either by solar cells being placed on the device 2000 or the solar cells could be placed on articles of clothing (i.e. hat, shirt or pants for example) to facilitate the charging thereof.
Figure 20 illustrates the embodiment of a system simulator 2100 for use by developers of applications and new lenses or expansion of the wearable optics device. In this embodiment, there is a simulator for the operating system 2102, a lens simulator 2104, a display 21 14, and an eyewear emulator 21 16. Eye
position/perceptual measurements 2106, camera signals, and other sensors and measurements are inputs to the simulator. The developer apps or new lenses can then be tested for various types of wearable options with various types of operating systems including iOS, Andriod, and general purpose or optimized
optical/perceptual operating systems.
Figure 21 A through Figure 21 F illustrate an embodiment of inverse shading using the wearable optics device. Figure 21 A illustrates the problem of glare caused by ambient light which degrades the visibility of a object such as the screen of a mobile phone or laptop. Figure 21 C describes the iris / pupil contraction due to brightness which degrades retina/cornea and brain view of target object such as a phone screen or the like. In Figure 21 E the phone screen appears dark since ambient light is far brighter than screen.
Figure 21 B illustrates the selection of the target object as in a phone screen via eye or automatically by preferences, rules, camera image capture and object recognition. Figure 21 D shows eye position measurement and capture of the object's position and image by a camera on the eyewear. Figure 21 F shows the resulting elimination or reduction in glare and increase in visibility of the object wherein a shaded or translucent background follows surrounding area object in real time as seen from the user of the wearable optics device.
Figure 22 illustrates an embodiment of eye position illumination and enhanced efficiency utilizing the wearable optics device. Using the eye position sensor and camera the line of sight and focal length can be determined and used to control a directional illumination source such that the illumination source illuminates the area corresponding to the area being focused on by the user of the wearable optics device
Figure 23 illustrates an embodiment of real-time augmented reality overlay 2400 utilizing the wearable optics device. In this embodiment, information is transmitted via mobile, text, R2R, Internet, Wi-Fi, Facebook message, Twitter's tweet. The wearable optics device can utilize mobile, R2R, NFC, Wi-Fi, Internet to communicate. In one example the wearable optics device is utilized in a vehicle. In this example the driver's eyewear uses augmented reality to overlay advertising signs which can be personalized to the user which appear as if they are normal roadside billboards. In this example a real-time updated augmented reality danger sign is posted before a road hazard with the augmented reality sign being generated in real-time using information from the networks connected to the wearable eyewear device. This example also shows real-time translation from English to Spanish of navigational warning and advertising signs using augmented reality with the wearable optics device. Accordingly systems and methods in accordance with embodiments are disclosed that provide these enhanced features for wearable optics devices. To describe these features and embodiments in more detail refer now to the following description in conjunction with the following discussion. A key feature associated with these enhancements is providing a variety of perceptual parameters that can be utilized with the wearable optics devices. Examples of perceptual parameters include but are not limited to optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
Although the present invention has been described in accordance with the embodiments shown, one of ordinary skill in the art will readily recognize that there could be variations to the embodiments and those variations would be within the spirit and scope of the present invention. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.

Claims

What is claimed is: 1 . A method comprising:
providing a wearable optics device, wherein parameters personalized to a user can be provided based upon user information.
2. A method comprising:
providing a wearable optics device which includes providing an enhancement utilizing objective and subjective quality standards based upon perceptual parameters.
3. The method of claim 2, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
4. The method of claim 2 wherein the wearable optics device controls any or any combination of mimics, amplifies, or expands a user perceptual physiology utilizing perceptual parameters.
5. The method of claim 4 wherein the optical expressions are images.
6. The method of claim 1 , wherein materials that electrically control any or any combination of chromatic, refractive, diffractive, transparent, reflective properties of the wearable optics device are utilized.
7. The method of claim 1 , wherein the wearable optics device comprises any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic,
electroluminescence, eletroreflective, dynamic filtering materials.
8. The method of claim 1 , wherein one or more elements are utilized within the wearable optics device to provide image information into the eye.
9. The method of claim 1 , wherein the one or more elements include any or any combination of a lens projector, retinal projection.
10. The method of claim 9, wherein the retinal projection provides the occlusion.
1 1 . The method of claim 1 , wherein the wearable optics device comprises an electro refractive or diffractive material.
12. The method of claim 1 , wherein the parameters include any or any
combination of prescriptions for improving the vision of a user, a zoom feature, a microscope feature, magnifying feature, retinal projection feature.
13. The method of claim 1 , which includes detecting head movement.
14. The method of claim 1 , wherein a brain wave mechanism is utilized within the wearable optics device.
15. The method of claim 1 , wherein the parameters can be transmitted to one or more third parties.
16. The method of claim 1 , wherein the parameters are athletic information related to the user.
17. The method of claim 16, wherein perceptual parameters are utilized.
18. The method of claim 16, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
19. The method of claim 16, wherein the athletic information comprises training information for the user.
20. The method of claim 1 , wherein the method utilizes one or more camera components and one or more communication components.
21 . The method of claim 20, wherein the camera components utilizes any and any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic ,
electroluminescence, eletroreflective, electrorefractive, dynamic filtering, refractive materials.
22. The method of claim 21 , wherein the camera components utilizes filtering of certain light waves for controlled viewing or visual effects.
23. The method of claim 1 , wherein the parameters are medical features related to the user.
24. The method of claim 23, wherein perceptual parameters are utilized.
25. The method of claim 24, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
26. The method of claim 1 , wherein the parameters are dental features of a patient.
27. The method of claim 26, wherein perceptual parameters are utilized.
28. The method of claim 27, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
29. The method of claim 27, wherein the perceptual parameters are utilized to detect pain or other dental conditions.
30. The method of claim 26, wherein the dental features of the patient can be received by a dentist using the wearable optics to perform dental procedures on the patient utilizing augmented reality.
31 . The method of claim 1 , in which any or any combination of light, acoustic waves, ultrasound x-rays or infrared are utilized.
32. The method of claim 26, wherein the parameters include a zoom feature to aid the dentist in the performing the dental procedures.
33. The method of claim 32, wherein the dental procedures include any or any combination of drilling a tooth, tooth extraction, teeth cleaning and root canal.
34. The method of claim 1 , wherein the method is implemented utilizing one or more distributed networks.
35. The method of claim 1 , which includes wireless inputs to the wearable optics device.
36. The method of claim 1 , wherein advertising is provided.
37. The method of claim 36, wherein perceptual parameters are utilized.
38. The method of claim 37, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
39. The method of claim 36, wherein the advertising is advertising a product or service that is proximate to the user of the wearable optics device.
40. The method of claim 39, wherein an advertising target is determined by any of or any combination of eye movement, GPS, wireless location measurement, accelerometer, and gyroscope.
41 . The method of claim 40, wherein the wearable optics device can view advertising to provide viewing information to a vendor.
42. The method of claim 36, wherein the advertising can appear to be on a billboard that is visible to a user of the wearable optics device.
43. The method of 1 , wherein the wearable optics device interfaces with a user application with the tablet, personal computer, portable device, mobile phone wearable sensors, wearable devices or the like.
44. The method of claim 43, wherein perceptual parameters are utilized.
45. The method of claim 44, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
46. The method of 1 , wherein an application runs on a processor of the wearable optics device.
47. The method of claim 1 , wherein the wearable optics device is utilized with a social networking application.
48. The method of claim 47, wherein perceptual parameters are utilized.
49. The method of claim 48, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
50. The method of claim 49, wherein the perceptual parameters are utilized to identify preferences including any or any combination of friends, company position, company background, relationship status, related parties.
51 . The method of claim 49, wherein the perceptual parameters are utilized for navigation inside and outside the social networking application.
52. The method of claim 47, wherein the wearable optics device can view any or any combination of friends, linked-in related parties, followers by highlighting within a field of view.
53. The method of claim 52, wherein perceptual parameters are utilized.
54. The method of claim 53, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
55. The method of claim 1 , wherein the wearable optics device is utilized for infrared or motion detection.
The method of claim 1 , wherein DVR functions are utilized.
57. The method of claim 56, wherein the DVR functions include any or any combination of play, pause, record and replay.
58. The method of claim 1 , wherein the wearable optics device is utilized for dancing and fitness.
59. A wearable optics device comprising:
a lens; wherein parameters personalized to a user can be provided based upon user information.
60. The wearable optics device of claim 59, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive, facial gestures
61 . The wearable optics device of claim 60, wherein the optical expressions are images.
62. The wearable optics device of claim 61 , wherein the wearable optics device controls any or any combination of mimics, amplifies, or expands a user perceptual physiology utilizing perceptual parameters.
63. The wearable optics device of claim 59, wherein lens comprises controlled refractive lenses.
64. The wearable optics device of claim 59, wherein one or more elements are utilized within the wearable optics device to provide image information into the eye.
65. The wearable optics device of claim 64, wherein the one or more elements include any or any combination of a lens projector, retinal projection.
66. The wearable optics device of claim 59, wherein the wearable optics device comprises an electro refractive material.
67. The wearable optics device of claim 59, wherein the parameters include any of or any combination of prescriptions for improving the vision of a user, a zoom feature, a microscope feature, a retinal projection feature.
68. The wearable optics device of claim 67, wherein perceptual parameters are utilized.
69. The wearable optics device of claim 68, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
70. The wearable optics device of claim 59, wherein a brain wave mechanism is utilized within the wearable optics device.
71 . The wearable optics device of claim 59, wherein the parameters can be transmitted to one or more third parties.
72. The wearable optics device of claim 59, wherein the parameters are athletic information related to the user.
73. The wearable optics device of claim 72, wherein perceptual parameters are utilized.
74. The wearable optics device of claim 73, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
75. The wearable optics device of claim 59, wherein the one or more camera components and one or more communication components are utilized.
76. The wearable optics device of claim 75, wherein the camera component utilizes any and any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic , electroluminescence, electroreflective, dynamic filtering, refractive materials.
77. The wearable optics device of claim 59 includes a camera component, communication components to provide an augmented reality.
78. The wearable optics device of claim 77, wherein the parameters include a zoom feature on the camera component to aid the user in the performance in an athletic event.
79. The wearable optics device of claim 59, wherein the parameters are medical features related to the user.
80. The wearable optics device of claim 59, wherein the parameters are dental features of a patient.
81 . The wearable optics device of claim 80, wherein perceptual parameters are utilized.
82. The wearable optics device of claim 81 , wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
83. The wearable optics device of claim 82, wherein the perceptual parameters are utilized to detect pain or other dental conditions.
84. The wearable optics device of claim 80, wherein the dental features of the patient can be received by a dentist using the wearable optics to perform dental procedures on the patient utilizing augmented reality.
85. The wearable optics device of claim 59, in which any or any combination of light, ultrasound, x-rays, acoustic or infrared are utilized.
86. The wearable optics device of claim 80, wherein the parameters include a zoom feature to aid the dentist in the performing the dental procedures.
87. The wearable optics device of claim 81 , wherein the dental procedures include any or any combination of drilling a tooth, tooth extraction, teeth cleaning and root canal.
88. The wearable optics device of claim 59, wherein one or more distributed networks are utilized by the wearable optics device.
89. The wearable optics device of claim 80, wherein the results of the dental procedures can be transmitted to one or more third parties.
90. The wearable optics device of claim 79, wherein the medical features comprises anatomical features of a patient.
91 . The wearable optics device of claim 90, wherein the anatomical features of the patient can be received by a medical doctor using the wearable optics to perform medical procedures on the patient utilizing augmented reality.
92. The wearable optics device of claim 59, wherein advertising is provided.
93. The wearable optics device of claim 59, wherein advertising is provided based upon user parameters.
94. The wearable optics device of claim 93, wherein a product or service is advertised that is proximate to the user of the wearable optics device.
95. The wearable optics device of claim 93, wherein the wearable optics device can view advertising to provide viewing information to a vendor.
96. The wearable optics device of claim 93, wherein the advertising can appear to be on a billboard that is visible to a user of the wearable optics device.
97. The wearable optics device of 59, wherein the wearable optics device interfaces with a user application with the tablet, portable device cell phone or the like.
98. The wearable optics device of 59, wherein the application runs on the processor of the wearable optics.
99. The wearable optics device of claim 72, wherein the athletic information can be related to any of track, skiing, baseball, football, tennis, basketball, and golf.
100. The wearable optics device of claim 99, wherein in football an intended target is highlighted.
101 . The wearable optics device of claim 100, wherein a particular characteristic of a player is highlighted.
102. The wearable optics device of claim 100, wherein augmented reality is provided remotely.
103. The wearable optics device of claim 100, wherein an Identification of geometry of the field such as first down, goal line is provided.
104. The wearable optics device of claim 103, wherein each player is identified.
105. The wearable optics device of claim 100, wherein possession of the football is identified.
106. The wearable optics device of claim 99, wherein a skier utilizing the wearable optics device has access to skiing information; which includes any of or any combination of speed and location of other skiers, altitude, and topographic features of the terrain.
107. The wearable optics device of claim 99, wherein in baseball an intended target is highlighted.
108. The wearable optics device of claim 107, wherein a particular characteristic of a player is highlighted.
109. The wearable optics device of claim 107, wherein augmented reality is provided remotely.
1 10. The wearable optics device of claim 107, wherein an Identification of geometry of the field such as field size, strike zone and location of stands is provided.
1 1 1 . The wearable optics device of claim 107, wherein each player is identified.
1 12. The wearable optics device of claim 59, wherein the user is military personnel.
1 13. The wearable optics device of claim 1 12, wherein the information is related to use of a firearm.
1 14. The wearable optics device of claim 1 13, wherein the information includes any or any combination of range finder information, wind measurements, and wind adjustments.
1 15. The wearable optics device of claim 1 14, wherein the information is retrieved from a database.
1 16. The wearable optics device of claim 1 15, wherein the information comprises any of or any combination of maps, architectural drawings and wiring schematics.
1 17. The wearable optics device of claim 1 12, wherein the military personnel utilizes augmented reality.
1 18. The wearable optics device of claim 59, wherein social media features are utilized.
1 19. The wearable optics device of claim 1 18, wherein the social media features include real time communication, including any or any combination of texting, instant messaging, video conferencing, twitter.
120. The wearable optics device of claim 59, wherein virtual reality is provided.
121 . The wearable optics device of claim 59, wherein navigation is utilized.
122. The wearable optics device of claim 59, wherein the wearable optics device is utilized in a gaming environment.
123. The wearable optics device of claim 122, wherein the gaming environment is utilized with augmented reality.
124. The wearable optics device of claim 123, wherein the gaming environment is online or mobile.
125. The wearable optics device of claim 124, wherein the wearable optics device determines statistics by referring to card analysis.
126. The wearable optics device of claim 124, wherein the wearable optics analyzes a person's facial expressions.
127. The wearable optics device of claim 124, wherein the wearable optics device is utilized in gambling; wherein gambling includes any or any combination of combination of slots, lotteries, sports betting, poker, and roulette.
128. The wearable optics device of claim 127, wherein perceptual parameters are utilized.
129. The wearable optics device of claim 128, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
130. The wearable optics device of claim 128, wherein the perceptual parameters are utilized to assist in determining strategy based upon viewing other participants.
131 . The wearable optics device of claim 59, wherein the wearable optics device is utilized in online or mobile environments.
132. The wearable optics device of claim 131 , wherein media is incorporated to augment the experience.
133. The wearable optics device of claim 59, wherein the wearable optics device is utilized for illumination.
134. The wearable optics device of claim 133, wherein the illumination is controlled by any or any combination of automatically, manually, voice control.
135. The wearable optics device of claim 59, wherein the wearable optics device is utilized for mobile payments.
136. The wearable optics device of claim 135, wherein the mobile payments are controlled by any or any combination of automatically, manually, via voice control.
137. The wearable optics device of claim 59, wherein the wearable optics device is utilized in conjunction with demographic information.
138. The wearable optics device of claim 137, wherein the demographic information is utilized with any one or combination of online media, web mobile, physical, radio advertising, television.
139. The wearable optics device of claim 137, wherein the demographic information is gathered based upon the channel being watched or the position that is viewed on a display by the user.
140. The wearable optics device of claim 59, wherein the wearable optics device includes stereoscopic lens..
141 . The wearable optics device of claim 59, wherein passwords are utilized in conjunction with user names to enable the wearable optics device.
142. The wearable optics device of claim 59, wherein the wearable optics device creates vibrations around glasses.
143. The wearable optics device of claim 59, wherein the wearable optics device creates a vibration area around a frame worn by the user's face.
144. The wearable optics device of claim 59, wherein the wearable optics device is utilized with any of a visor, helmet, facemask, safety goggles.
145. The wearable optics device of claim 59, wherein the wearable optics are utilized in conjunction with social networks or external databases to identify characteristics of those or that which is viewed by the user.
146. The wearable optics device of claim 59, wherein the wearable optics device is utilized in conjunction with movies.
147. The wearable optics device of claim 146, wherein the movies include any of online mobile movies or physical movies.
148. The wearable optics device of claim 147, wherein the movies include any of a 3D or IMAX movie.
149. The wearable optics device of claim 59, wherein the wearable optics device is utilized for the purchase of songs music or games, movies and video games.
150. The wearable optics device of claim 149, wherein the games comprise word games.
151 . The wearable optics device of claim 149, wherein the games comprise any of chess, checkers solitaire and tic tac toe.
152. The wearable optics device of claim 149, wherein the games comprise games based upon physics.
153. The wearable optics device of claim 59, wherein the wearable optics device is utilized with compensation algorithms in conjunction with a curved lens.
154. The wearable optics device of claim 153, wherein the curved lens used to create a spatially or visually mapped perspective for the user.
155. The wearable optics device of claim 153, wherein the curved lens is used to create a panoramic viewing area utilizing multiple lenses.
156. The wearable optics device of claim 153, wherein the curved lens is used to create a 360 degree viewing area.
157. The wearable optics device of claim 59, wherein browser information is executed.
158. The wearable optics device of claim 59 includes a power management mechanism.
159. The wearable optics device of claim 59 includes a charging mechanism.
160. The wearable optics device of claim 159, wherein the charging mechanism comprises a direct electrical connection.
161 . The wearable optics device of claim 159, wherein the charging mechanism comprises a solar charging system.
162. The wearable optics device of claim 159, wherein the charging mechanism comprises a magnetic induction system.
163. The wearable optics device of claim 159, wherein the charging mechanism comprises a connection to a smartphone.
164. The wearable optics device of claim 163, wherein the connection to the smartphone comprises any and any combination of a connection to a USB port, a direct electrical connection or a connection via magnetic induction.
165. The wearable optics device of claim 163, wherein the connection is a connection to a solar induction pad.
166. The wearable optics device of claim 159, wherein the charging mechanism comprises a connection to a laptop.
167. The wearable optics device of claim 166, wherein the connection to the laptop comprises any and any combination of a connection to a USB port, a direct electrical connection or a connection via magnetic induction.
168. The wearable optics device of claim 167, wherein the connection is a connection to a solar induction pad.
169. The wearable optics device of claim 159, wherein clothes are utilized in conjunction with the charging mechanism to charge the wearable optics device.
170. The wearable optics device of claim 59, wherein the wearable optics device is utilized for dancing and fitness.
171 . A method comprising:
providing a wearable optics device which includes providing an enhancement utilizing objective and subjective quality standards based upon perceptual parameters.
172. The method of claim 171 , wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
173. The method of claim 171 , wherein the wearable optics device controls any or any combination of mimics, amplifies, or expands a user perceptual physiology utilizing perceptual parameters
174. The method of claim 173, wherein the optical expressions are images.
175. A method comprising:
providing a wearable optics device; wherein the wearable optics device comprises one or more inserts into eyeglasses.
176. A method comprising:
providing a wearable optics device; wherein the eyeglasses comprises quad state eyeglasses.
177. A method comprising:
providing a wearable optics device; wherein shading control is utilized on the wearable optics device.
178. The method of claim 177, wherein the shading control is provided by one or more projectors within the wearable optics device.
179. The method of claim 178, wherein an occlusion effect is projected on a lens of the wearable optics device.
180. The method of claim 178, wherein shading is provided on a lens of the wearable optics device wherein the surrounding area is occluded or reversed.
181 . The method of claim 178, wherein shading is provided by a polarized filter.
182. The method of claim 177, wherein the shading control is provided by the lenses within the wearable optics device.
183. The method of claim 182, wherein shading is provided on a lens of the wearable optics device wherein the surrounding area is occluded or reversed.
184. The method of claim 182, wherein shading is provided by a polarized filter.
185. The method of claim 177 in which the shading is controlled using perceptual parameters.
186. The method of claim 185, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
187. The method of claim 177, wherein portions of an image viewed by the wearable optics device are shaded to control brightness.
188. A method comprising:
providing a wearable optics device wherein the wearable optics device comprises an electrochromatic material.
189. A method comprising:
providing a wearable optics device wherein the wearable optics device comprises polychromatic material.
190. A method comprising:
providing a wearable optics device wherein the wearable optics device is utilized in a simulator.
191 . A method comprising:
providing a wearable optics device wherein a focal of the wearable optics device is utilized.
192. A method comprising:
providing a wearable optics device wherein the parameters include any or any combination of a zoom feature, a microscope feature, magnifying feature, illumination feature; a retinal projection feature.
193. A method comprising:
providing a wearable optics device wherein a 360 degree view is provided.
194. The method of claim 193, wherein the 360 degree view can be any or any combination of a left or right panning, up and down panning, three dimensional rotations.
195. A method comprising:
providing a wearable optics device wherein an illumination feature is directed to a specific area.
196. The method of claim 195, wherein the illumination is one of any combination of a laser, visible light, infrared light, ultraviolet light, polarized light.
197. A method comprising:
providing a wearable optics device wherein the wearable optics device comprises lenses to filter certain light waves for controlled viewing or visual effects.
198. The method of claim 197 comprises controlling noise reduction, polarization, creative effects.
199. The method of claim 197 comprises controlling a stability control for facial or object focus.
200. The method of claim 197, wherein perceptual parameters are utilized.
201 . The method of claim 200, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
202. A method comprising: providing a wearable optics device wherein an acoustic wave mechanism is utilized within the wearable optics device.
203. A method comprising:
providing a wearable optics device wherein a brain wave mechanism is utilized within the wearable optics device.
204. A method comprising:
providing a wearable optics device wherein a magnetic wave mechanism is utilized within the wearable optics device.
205. A method comprising:
providing a wearable optics device wherein the wearable optics device is utilized in conjunction with an athletic event by a participant.
206. A method comprising:
providing a wearable optics device wherein the wearable optics device is utilized in conjunction with an athletic event by a spectator.
207. The method of claim 206, wherein the athletic information comprises training information for the user.
208. A method comprising:
providing a wearable optics device wherein the method utilizes one or more camera components and one or more communication components to provide an augmented reality.
209. The method of claim 208, wherein the parameters include a zoom feature on the one or more camera components to aid the user in the performance in an athletic event.
210. A method comprising: providing a wearable optics device wherein sensory inputs are provided utilizing sensors.
21 1 . The method of claim 210, wherein perceptual parameters are utilized.
212. The method of claim 21 1 , wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
213. The method of claim 210, wherein the sensors comprise any or any combination of gyroscopes, accelerometers, torque sensors, weight sensors, pressure sensors, magnetometers, temperature sensors, light sensor, cameras and microphones, GPS, wireless detection, altitude sensors, blood pressure , heart rate sensors, biometric sensors, radio frequency identification (RFID), near field communication (NFC),, mobile communication, Wi-Fi, strain gauges, fingerprint sensors, smell sensors gas sensors, chemical sensors, color sensors, sound sensors, acoustic sensors, ultraviolet sensors, electric field sensors, magnetic field sensors, gravity sensors, wind speed sensors, wind direction sensors, compass sensors, geo-locator sensor, polarized light sensors, infrared emitter sensors.
214. The method of claim 210, wherein the sensors are coupled wirelessly.
215. The method of claim 214, wherein the sensors are coupled via a USB wireless connection.
216. The method of claim 214, wherein the sensors are coupled via a Bluetooth wireless connection.
217. The method of claim 213, wherein the sensor comprises an earpiece connected either by a wire, to a frame, or connected wirelessly
218. The method of claim 217, wherein the earpiece measures the ear muscles.
219. The method of claim 218, wherein the ear muscle measurement is utilized to adjust acoustic volume or acoustic focus.
220. The method of claim 218, wherein the sensor comprises a microphone either connected by a wire, connected to frame, or connected wirelessly
221 . The method of claim 220, wherein the microphone is used in combination with any or any combination of a camera, infrared camera, imaging sensors.
222. The method of claim 210, wherein the sensor comprises an earpiece containing a speaker that includes a microphone for noise cancellation.
223. A method comprising:
providing a wearable optics device wherein an input/output (I/O) capability is integrated with the wearable optic device.
224. The method of claim 223, wherein remote devices are utilized with the I/O capability.
225. A method comprising
providing a wearable optics device; wherein different media can be overlaid.
226. A method comprising:
providing a wearable optics device wherein different media can be blended.
227. A method comprising:
providing a wearable optics device wherein the parameters include a zoom feature to aid the medical doctor in performing the medical procedures.
228. The method of claim 227, wherein the medical procedures include surgical and/or diagnostic procedures.
229. The method of claim 228, wherein the results of the medical procedures can be transmitted to one or more third parties.
230. A method comprising:
providing a wearable optics device wherein the wearable optics device comprises a plurality of lenses on eyeglasses.
231 . The method of claim 230, wherein the plurality of lenses comprises controlled refractive lenses.
232. The method of claim 230, wherein the plurality of lenses comprising polychromatic material.
233. The method of claim 60, wherein the plurality of lenses comprising any or any combination of refractive, diffractive material.
234. The method of claim 230, wherein the plurality of lenses comprising any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED, transparent matrix, semi-transparent matrix, prism based, holographic ,
electroluminescence, eletroreflective, dynamic filtering materials..
235. The method of claim 230, wherein the plurality of lenses comprise a single integrated system.
236. The method of claim 230, wherein shading control is utilized on the plurality of lenses.
237. A method comprising: providing a wearable optics device wherein the wearable optics device can view encoded images.
238. A method comprising:
providing a wearable optics device wherein rapid motion encoding is utilized.
239. The method of claim 171 , wherein the wearable optics device is utilized for dancing and fitness.
240. A wearable optics device comprising:
a lens and an enhancement to the lens utilizing objective and subjective quality standards based upon perceptual parameters.
241 . The wearable optics device of claim 240, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive, facial gestures
242. The wearable optics device of claim 241 , wherein the optical expressions are images.
243. The wearable optics device of claim 241 , wherein the wearable optics device controls any or any combination of mimics, amplifies, or expands a user perceptual physiology utilizing perceptual parameters.
244. A wearable optics device comprising:
eyeglasses including one or more inserts into the eyeglasses.
245. The wearable optics device of claim 244, wherein the eyeglasses comprises quad state eyeglasses.
246. A wearable optics device comprising: a lens, wherein shading control is utilized on the lens.
247. The wearable optics device of claim 246 in which the shading is controlled using perceptual parameters.
248. The wearable optics device of claim 247, wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
249. The wearable optics device of claim 247, wherein the lens comprises any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED transparent matrix, semi-transparent matrix, prism based, holographic ,
electroluminescence, electroreflective, dynamic filtering, refractive materials.
250. The wearable optics device of claim 249, wherein materials that electrically control any or any combination of chromatic, refractive, diffractive, transparent, reflective properties of the wearable optics device are utilized.
251 . The wearable optics device of claim 246, wherein portions of an image viewed by the lens are shaded to minimize brightness.
252. A wearable optics device comprising:
a lens; wherein the lens comprises polychromatic material.
253 A wearable optics device comprising:
a lens; wherein the lens comprises an electrochromatic material.
254 A wearable optics device comprising:
a lens; wherein shading, inverse shading, or occlusion is provided by the any or any combination of lens projector, retinal projection.
255. A wearable optics device comprising:
a lens; wherein the parameters include any or any combination of a zoom feature, a microscope feature, magnifying feature, illumination feature; a retinal projection feature.
256. The wearable optics device of claim 255, wherein an illumination feature is directed to a specific area.
257. The wearable optics device of Claim 255, wherein the illumination is one of any combination of a laser, visible light, infrared light, ultraviolet light, polarized light.
258. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in a simulator.
259. A wearable optics device comprising:
a lens; wherein the focal point of the user of the wearable optics device utilized.
260. The wearable optics device of claim 259, wherein the parameters include any or any combination of a zoom feature, a microscope feature, magnifying feature, illumination feature; a projector feature, a retinal projection feature.
261 . A wearable optics device comprising:
a lens; wherein a 360 degree view is provided.
262. The wearable optics device of claim 261 , wherein the 360 degree view can be any or any combination of a left or right panning, up and down panning ,three dimensional rotation.
263. A wearable optics device comprising:
a lens; wherein filtering for certain light waves is utilized for controlled viewing or visual effects.
264. The wearable optics device of claim 93, wherein the filtering comprises controlling the illumination noise reduction, polarization, and creative effects.
265. A wearable optics device comprising:
a lens; and which includes stability control for facial or object focus.
266. A wearable optics device comprising:
a lens; wherein an acoustic wave mechanism is utilized within the wearable optics device.
267. A wearable optics device comprising:
a lens; wherein a magnetic wave mechanism is utilized within the wearable optics device.
268. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in conjunction with an athletic event by a participant.
269. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in conjunction with an athletic event by a spectator.
270. A wearable optics device comprising:
a lens; wherein sensors are utilized to provide sensory inputs to the device.
271 . The wearable optics device of claim 270, wherein the sensors are coupled wirelessly.
272. The wearable optics device of claim 270, wherein the sensors are coupled via a USB wireless connection.
273. The wearable optics device of claim 270, wherein the sensors are coupled via a Bluetooth wireless connection.
274. The wearable optics device of claim 270, wherein the sensors comprise any or any combination of gyroscopes, accelerometers, pressure sensors, torque sensors, weight sensors, magnetometers, temperature sensors, light sensor, cameras and microphones, GPS, wireless detection, altitude sensors, blood pressure , heart rate sensors, biometric sensors, radio frequency identification (RFID), near field communication (NFC), mobile communication, Wi-Fi, strain gauges, fingerprint sensors, smell sensors gas sensors, chemical sensors, color sensors, sound sensors, acoustic sensors, ultraviolet sensors, electric field sensors, magnetic field sensors, gravity sensors, wind speed sensors, wind direction sensors, compass sensors, geo-locator sensor, polarized light sensors, infrared emitter sensors.
275. The wearable optics device of claim 274, wherein the sensor comprises an earpiece connected either by a wire, to a frame, or connected wirelessly.
276. The wearable optics device of claim 275, wherein the earpiece measures the ear muscles.
277. The wearable optics device of claim 276, wherein the ear muscle
measurement is utilized to adjust acoustic volume or acoustic focus.
278. The wearable optics device of claim 274, wherein the sensor comprises a microphone either connected by a wire, connected to frame, or connected wirelessly.
279. The wearable optics device of claim 277, wherein the microphone is used in combination with any or any combination of a camera, infrared camera, imaging sensors.
280. The wearable optics device of claim 274, wherein the sensor comprises an earpiece containing a speaker that includes a microphone for noise cancellation.
281 . A wearable optics device comprising:
a lens; and includes an input/output (I/O) capability integrated with the wearable optics device.
282. The wearable optics device of claim 281 , wherein remote devices are utilized with the I/O capability.
283. A wearable optics device comprising a lens wherein different media can be overlaid.
284. A wearable optics device comprising:
a lens wherein different media can be blended.
285. The wearable optics device of claim 284 includes wireless inputs.
286. A wearable optics device comprising:
a lens; wherein the parameters include a zoom feature to aid the medical doctor in performing the medical procedures.
287. The wearable optics device of claim 286, wherein the medical procedures include surgical and/or diagnostic procedures.
288. The wearable optics device of claim 286, wherein the results of the medical procedures can be transmitted to one or more third parties.
289. A wearable optics device comprising:
a lens wherein an advertising target is determined by any of or any combination of GPS, wireless location measurement and gyroscope.
290. A wearable optics device comprising:
eyewear, wherein the eyewear comprises a plurality of lenses.
291 . The wearable optics device of claim 290, wherein the plurality of lenses comprises controlled refractive lenses.
292. The wearable optics device of claim 290, wherein the plurality of lenses comprising any one or any combination of polychromatic or electrochromatic material.
293. The wearable optics device of claim 290, wherein the plurality of lenses comprising any combination of refractive or diffractive material.
294. The wearable optics device of claim 290, wherein the plurality of lenses comprising any or any combination of transparent LCD, LED, OLED, flexible LED, flexible OLED materials transparent matrix, semi-transparent matrix, prism based, holographic , electroluminescence, eletroreflective, refractive, dynamic filtering.
295. The wearable optics device of claim 290, wherein the plurality of lenses are a single integrated system.
296. The wearable optics device of claim 290, wherein shading control is utilized on the plurality of lenses.
297. A wearable optics device comprising:
a lens; wherein the wearable optics device can view encoded media.
298. The wearable optics device of claim 268, wherein the athletic information can be related to any of track, skiing, baseball, football, tennis, basketball, and golf.
299. The wearable optics device of claim 298, wherein in football an intended target is highlighted.
300. The wearable optics device of claim 298, wherein a particular characteristic of a player is highlighted.
301 . The wearable optics device of claim 298, wherein augmented reality is provided remotely.
302. The wearable optics device of claim 298, wherein an Identification of geometry of the field such as first down, goal line is provided.
303. The wearable optics device of claim 298, wherein each player is identified.
304. The wearable optics device of claim 298, wherein possession of the football is identified.
305. The wearable optics device of claim 298, wherein a skier utilizing the wearable optics device has access to skiing information; which includes any of or any combination of speed and location of other skiers, altitude, and topographic features of the terrain.
306. The wearable optics device of claim 298, wherein in baseball an intended target is highlighted.
307. The wearable optics device of claim 306, wherein a particular characteristic of a player is highlighted.
308. The wearable optics device of claim 306, wherein augmented reality is provided remotely.
309. The wearable optics device of claim 306, wherein an Identification of geometry of the field such as field size, strike zone and location of stands is provided.
310. The wearable optics device of claim 309, wherein each player is identified.
31 1 . A wearable optics device comprising:
a lens; wherein the wearable optics device can be utilized as binocular glasses.
312. The wearable optics device of claim 31 1 , wherein an object to be enlarged is identified by eye movement.
313. The wearable optics device of claim 31 1 , wherein in zoom mode the object is followed.
314. The wearable optics device of claim 31 1 , wherein in zoom mode the object is stabilized.
315. The wearable optics device of claim 298, wherein the golf information includes any or any combination of club choice, topology, wind measurements, and club angles.
316. A wearable optics device comprising:
a lens; wherein the parameters are related to information for the user.
317. The wearable optics device of claim 316, wherein the user is law
enforcement personnel.
318. The wearable optics device of claim 317, wherein the information is related to use of a firearm.
319. The wearable optics device of claim 318, wherein the information includes any or any combination of range finder information, wind measurements, and wind adjustments.
320. A wearable optics device of claim 316, wherein the user is firefighting personnel.
321 . The wearable optics device of claim 320, wherein the information is related to use of firefighting equipment.
322. The wearable optics device of claim 320, wherein the information includes GPS information regarding the location of a fire.
323. The wearable optics device of claim 320, wherein the information is retrieved from a database.
324. The wearable optics device of claim 323, wherein the information comprises any of or any combination of maps, architectural drawings and wiring schematics.
325. The wearable optics device of claim 320, wherein the firefighting personnel utilizes augmented reality.
326. A wearable optics device of claim 316, wherein the user is security personnel.
327. A wearable optics device comprising:
a lens; wherein advertising is provided utilizing any or any combination of GPS, cellular or other location based system.
328. A wearable optics device comprising:
a lens; wherein advertising is provided utilizing any or any combination of a map database and a building database.
329. A wearable optics device comprising:
a lens; wherein augmented reality is utilized based on location.
330. A wearable optics device comprising:
a lens; wherein augmented reality is utilized based on demographics.
331 . A wearable optics device comprising:
a lens; wherein augmented reality is utilized based on location condition.
332. A wearable optics device comprising:
a lens; wherein audio wave technology is utilized.
333. The wearable optics device of claim 332, wherein the audio wave technology includes audio focus.
334. The wearable optics device of claim 333, wherein the audio wave technology includes noise cancellation.
335. The wearable optics device of claim 334, wherein the noise cancellation is selective.
336. The wearable optics device of claim 335, wherein the selectivity is such that other noises are blocked out.
337. The wearable optics device of claim 335, wherein the selectivity is utilized in conjunction with a real life application.
338. The wearable optics device of claim 335, wherein the selectivity is such that audio mixing control is provided.
339. The wearable optics device of claim 335, wherein the selectivity is such that sourced audio and ambient audio are controlled to produce the effects.
340. The wearable optics device of claim 334, wherein the noise cancellation provides for the cancellation of certain ranges of sounds.
341 . The wearable optics device of claim 334, wherein the noise cancellation provides for the cancellation of certain sources of sounds.
342. The wearable optics device of claim 334, wherein there are classes of noise cancellation.
343. The wearable optics device of claim 334, wherein non-object sound is cancelled or reduced.
344. The wearable optics device of claim 334, wherein a person is selected and the sounds form nonselected individuals are cancelled.
345. The wearable optics device of claim 334, wherein a speaker and microphone are connected to devices on the outer ear or the inner ear on the frames.
346. The wearable optics device of claim 345, wherein a head band is utilized to hold the speaker and microphone.
347. The wearable optics device of claim 332, wherein the audio wave technology includes voice recognition.
348. The wearable optics device of claim 332, wherein the audio wave technology includes utilizing a vibration detection mechanism.
349. The wearable optics device of claim 348, wherein the vibration detection mechanism is utilized for warning.
350. The wearable optics device of claim 348, wherein the vibration detection mechanism includes any and any combination of bone conduction of the jaw, skull, outer ear, and inner ear.
351 . The wearable optics device of claim 349, wherein vibration is utilized to enhance the use of the media.
352. The wearable optics device of claim 349, wherein vibration can provide a high quality perception of sounds.
353. The wearable optics device of claim 349, wherein vibration can be utilized with augmented reality.
354. A wearable optics device comprising:
a lens; wherein a physical display appears as a blurry screen to anyone not using the wearable optics device.
355. A wearable optics device comprising:
a lens; wherein virtual reading features are provided.
356. The wearable optics device of claim 355, wherein zoom control is provided.
357. A wearable optics device comprising:
a lens; wherein 360 degree viewing of media for information is provided.
358. A wearable optics device comprising:
a lens and an augmented reality keyboard display.
359. The wearable optics device of claim 358, wherein the augmented reality keyboard display can be utilized for any or any combination of a laptop, personal computer, mobile phone, tablet computer.
360. A wearable optics device comprising:
a lens and an augmented reality mouse or pointing device.
361 . A wearable optics device comprising:
a lens; wherein an augmented reality display can be utilized in conjunction with a point of sale device.
362. The wearable optics device of claim 361 , wherein the augmented reality display can be utilized in conjunction with RFID or NFC communication.
363. The wearable optics device of claim 362, wherein the augmented reality display can be utilized in conjunction with online or mobile purchasing.
364. The wearable optics device of claim 362, wherein the augmented reality display can be utilized in conjunction with RFID or NFC communication.
365. A wearable optics device comprising:
a lens; wherein augmented reality GPS is utilized.
366. A wearable optics device comprising:
a lens; wherein video conferencing is utilized.
367. A wearable optics device comprising:
a lens; wherein audio conferencing is utilized.
368. The wearable optics device of claim 367, wherein the audio conferencing utilizes any or any combination of a caller ID name, telephone number or picture.
369. The wearable optics device of claim 367, wherein multiple callers are represented by any or any combination any or any combination of a caller ID name, telephone number or picture
370. The wearable optics device of claim 369, wherein an identified caller is more audible than other of the multiple callers.
371 . A wearable optics device comprising:
a lens; wherein tools are utilized.
372. The wearable optics device of claim 371 , wherein the tools can be utilized with remote sensors.
373. The wearable optics device of claim 372, wherein at least one of the remote sensors is a camera.
374. The wearable optics device of claim 371 , wherein the tools are utilized with augmented reality mechanisms.
375. A wearable optics device comprising:
a lens; wherein lip reading technology is utilized.
376. A wearable optics device comprising:
a lens; wherein language translation is utilized.
377. The wearable optics device of claim 376, wherein an overlay translation is utilized in an advertising media.
378. The wearable optics device of claim 376, wherein an overlay translation is translation overlay for street signs.
379. The wearable optics device of claim 376, wherein an overlay translation is an augmented reality overlay.
380. The wearable optics device of claim 376, wherein an overlay translation is provided in the wearable optics device using line of sight or focal length of the lens of the user to the object to provide an augmented reality overlay.
381 . A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized to determine what the user is watching at any particular point.
382. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized to determine what the user is viewing.
383. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized to determine
engagement with the media.
384. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized for identity and security information.
385. The wearable optics device of claim 384, wherein perceptual parameters are utilized.
386. The wearable optics device of claim 385, wherein the perceptual parameters include any and any combination of voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
387. The wearable optics device of claim 384, wherein the identity and security information is controlled by any or any combination of automatically, manually, eye scan, brain wave, facial recognition, fingerprint scan, voice scan.
388. A wearable optics device comprising: a lens; wherein the wearable optics device is utilized as a teleprompter.
389. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized for infrared or motion detection.
390. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in conjunction with an augmented reality remote control for media selection.
391 . A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in conjunction with remote sensors that acts as a periscope for the user.
392. The wearable optics device of claim 391 , wherein the periscope is utilized in an extended camera and microphone arrangement.
393. A wearable optics device comprising:
a lens; wherein the lens is in a transparent mode.
394. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized with a form filling pointing device.
395. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized for glare management.
396. The wearable optics device of claim 395, wherein the wearable optics device utilizes dimming reduction in brightness which can be utilized to improve any combination headlights, sun, spotlights or explosions.
397. A wearable optics device comprising: a lens; wherein at least one fingerprint utilized to enable the wearable optics device.
398. A wearable optics device comprising:
a lens wherein the wearable optics device is utilized with multiple windows surrounding the user.
399. The wearable optics device of claim 398, wherein the multiple windows are controlled.
400. The wearable optics device of claim 398, wherein owl vision is utilized to provide a 360 view.
401 . A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized to identify who is actually speaking.
402. A wearable optics device comprising:
a lens; wherein the wearable optics device includes conference calling features.
403. A wearable optics device comprising:
a curved lens, wherein the wearable optics device is utilized with
compensation algorithms in conjunction with the curved lens.
404. The wearable optics device of claim 403, wherein the curved lens is used to create a spatially or visually mapped perspective for the user
405. The wearable optics device of claim 403, wherein the curved lens is used to create a panoramic viewing area utilizing multiple lenses.
406. The wearable optics device of claim 403, wherein the curved lens is used to create a 360 degree viewing area.
407. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in conjunction with eye tests, eye diagnosis and eye exam.
408. The wearable optics device of claim 407, wherein the eye test comprises controlled cornea or eye lens measurement to eye muscles or muscles on the side of the face or its equivalent.
409. The wearable optics device of claim 407, wherein the wearable optics device is utilized to control light.
410. The wearable optics device of claim 407, wherein the wearable optics device is utilized for macular detection.
41 1 . A wearable optics device comprising:
a lens; wherein perceptual parameters are utilized.
412. The wearable optics device of claim 41 1 , wherein the perceptual parameters include any and any combination of optical expression, voice, brain wave, environmental, audio, video, navigational, augmented reality, algorithmic, spatial, cognitive, interpretive.
413. A wearable optics device comprising:
a lens; wherein the wearable optics device utilizes sound detection to improve audibility.
414. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in conjunction with an ear insert for acoustic processing.
415. A wearable optics device comprising:
a lens; wherein the wearable optics device comprises a contact lens with shading control.
416. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized to determine symptoms of sleep deprivation.
417. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized to determine an alcohol level of the user.
418. A wearable optics device comprising:
a lens; wherein the wearable optics device is utilized in conjunction with a tesla meter.
419. A wearable optics device comprising:
a lens; wherein the wearable optic device turns a page of a virtual book.
420. A wearable optics device comprising:
a lens; wherein the wearable optics device is solar powered.
421 . A wearable optics device comprising:
a lens which includes a bar code sensor.
422. The wearable optics device of claim 421 , wherein the bar code includes product information based on the scanning.
423. A wearable optics device comprising:
a lens; wherein perceptual measurements are utilized to identify information.
424. The wearable optics device of claim 423, wherein muscle measurement is utilized for perceptual enhancement.
425. The wearable optics device of claim 424, wherein central arch, nose, eyebrow muscle group is utilized for focusing and zoom function.
426. A wearable optics device comprising:
a lens; wherein perceptual parameters are utilized to convey transmit and communicate perceptions and emotions of the user.
427. A wearable optics device comprising:
a lens; wherein perceptual parameters are utilized in video games.
428. A wearable optics device comprising:
a lens; wherein perceptual parameters are utilized in a social network environment.
429. A wearable optics device comprising:
a lens; wherein perceptual parameters are utilized to provide a facial image.
430. A wearable optics device comprising:
a lens; where the eyebrow, forehead sensor is utilized to convey surprise or doubt.
431 . A wearable optics device comprising:
a lens; wherein cheek position of the user is utilized to indicate happiness.
432. A wearable optics device comprising:
a lens; wherein muscle movement is utilized as a proxy for speech.
433. The wearable optics device of claim 432, the nose bridge is utilized for perceptual enhancement.
434. The wearable optics device of claim 432, wherein the nose can be a speech vehicle.
435. The wearable optics device of claim 432, wherein a sensor on the face senses moisture content.
436. The wearable optics device of claim 435, wherein the perceptual search utilizes gesture movement to determine the focus of the search.
437. A wearable optics device comprising:
a lens; wherein an eye of a user looks natural.
438. The wearable optics device of claim 437, a natural eye look provides an ability to show a representation of a clear view of the eye in real time.
439. The wearable optics device of claim 437, wherein a camera within the wearable optics device is utilized to provide the natural eye look.
440. A wearable optics device comprising:
a lens; wherein a focal length of the lens is determined to improve performance of the device.
441 . A wearable optics device comprising;
a lens and a 10 click navigation mechanism using the fingers.
442. The wearable optics device of claim 441 , wherein the navigation is simultaneous with the selection,
443. A wearable optics device comprising:
a lens, wherein recognition of information is augmented.
444. A wearable optics device comprising:
a lens; wherein inverse shading is provided when viewing a display.
445. The wearable optics device of claim 444, wherein eye position is used to determine the inverse shading.
446. The wearable optics device of claim 444, wherein the inverse shading is provided by a polarized filter.
447. A wearable optics device comprising:
a lens; wherein algorithms for shading and inverse shading are provided within the wearable optics device.
448. A wearable optics device comprising:
a lens; wherein algorithms are provided to transform non-visual spectrum to visible information.
449. The wearable optics device of claim 444, wherein the display is utilized with any of a smartphone, personal computer, laptop, desktop or portable device.
450. The wearable optics device of claim 449, wherein the background of the display is shaded out to allow one to see an object
451 . The wearable optics device of claim 450, wherein an object is unshaded but the area outside of the object is shaded.
452. The wearable optics device of claim 447, wherein object recognition is utilized.
453. The wearable optics device of claim 447, wherein if ambient light is above some predetermined value then before shading occurs.
454. The wearable optics device of claim 447, wherein each lens has its own inverse shading.
455. The wearable optics device of claim 447, wherein translucency shading is provided for reading material.
456. A wearable optics device comprising:
a lens; wherein rapid motion encoding is used.
457. A wearable optics device comprising:
a lens; wherein DVR functions are used.
458. The wearable optics device of claim 457, wherein the DVR functions include any or any combination of play, pause, record and replay.
459. A wearable optics device comprising:
a lens;
an enhancement to the lens utilizing objective and subjective quality standards based upon perceptual parameters; and
a frame containing the lens; wherein the frame can be tailored to a user preference.
460. The wearable optics device of claim 459, wherein a color of the frame can be tailored to the user.
461 . The wearable optics device of claim 459, wherein images on the frame can be tailored to the user.
462. A wearable optics device comprising:
a lens and an enhancement to the lens utilizing objective and subjective quality standards based upon perceptual parameters; wherein a geometry of an augmented reality displayed sign can be adjusted in real-time.
463. The wearable optics device of claim 462, wherein the adjustment is based on any of into account the location, position, or velocity of the augmented reality displayed object and the location, position, or velocity of a user, the head position, angle, eye position measurements of a user.
PCT/US2014/029520 2013-03-15 2014-03-14 Enhanced optical and perceptual digital eyewear WO2014144918A2 (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US201361798770P 2013-03-15 2013-03-15
US13/841,141 2013-03-15
US13/841,141 US8696113B2 (en) 2005-10-07 2013-03-15 Enhanced optical and perceptual digital eyewear
US13/841,550 US11428937B2 (en) 2005-10-07 2013-03-15 Enhanced optical and perceptual digital eyewear
US61/798,770 2013-03-15
US13/841,550 2013-03-15

Publications (2)

Publication Number Publication Date
WO2014144918A2 true WO2014144918A2 (en) 2014-09-18
WO2014144918A3 WO2014144918A3 (en) 2015-01-22

Family

ID=51538393

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2014/029520 WO2014144918A2 (en) 2013-03-15 2014-03-14 Enhanced optical and perceptual digital eyewear

Country Status (1)

Country Link
WO (1) WO2014144918A2 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105574946A (en) * 2015-12-21 2016-05-11 天津中兴智联科技有限公司 Hand-held ETC reader and employed RFID and AR fusion method thereof
US9649052B2 (en) 2014-09-05 2017-05-16 Vision Service Plan Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual
GB2546883A (en) * 2015-12-21 2017-08-02 Skullcandy Inc Electrical systems and related methods for providing smart mobile electronic device features to a user of a wearable device
US9910298B1 (en) 2017-04-17 2018-03-06 Vision Service Plan Systems and methods for a computerized temple for use with eyewear
CN108540764A (en) * 2017-03-02 2018-09-14 北京戈德思科科技有限公司 A kind of split type wearable augmented reality equipment
US10215568B2 (en) 2015-01-30 2019-02-26 Vision Service Plan Systems and methods for tracking motion, performance, and other data for an individual such as a winter sports athlete
US10338880B2 (en) 2015-06-03 2019-07-02 Skullcandy, Inc. Audio devices and related methods for acquiring audio device use information
EP3210095B1 (en) * 2014-10-21 2019-12-11 Signify Holding B.V. System, method and computer program for hands-free configuration of a luminous distribution
US10617342B2 (en) 2014-09-05 2020-04-14 Vision Service Plan Systems, apparatus, and methods for using a wearable device to monitor operator alertness
US10722128B2 (en) 2018-08-01 2020-07-28 Vision Service Plan Heart rate detection system and method
US11918375B2 (en) 2014-09-05 2024-03-05 Beijing Zitiao Network Technology Co., Ltd. Wearable environmental pollution monitor computer apparatus, systems, and related methods

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060061544A1 (en) * 2004-09-20 2006-03-23 Samsung Electronics Co., Ltd. Apparatus and method for inputting keys using biological signals in head mounted display information terminal
US7629877B2 (en) * 2006-12-19 2009-12-08 Matvey Lvovskiy Multifunctional collimator indicator
US7677723B2 (en) * 2003-10-09 2010-03-16 Ipventure, Inc. Eyeglasses with a heart rate monitor
US20120021806A1 (en) * 2010-07-23 2012-01-26 Maltz Gregory A Unitized, Vision-Controlled, Wireless Eyeglass Transceiver
US20120056993A1 (en) * 2010-09-08 2012-03-08 Salman Luqman Dental Field Visualization System with Improved Ergonomics
US20120242678A1 (en) * 2010-02-28 2012-09-27 Osterhout Group, Inc. See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7677723B2 (en) * 2003-10-09 2010-03-16 Ipventure, Inc. Eyeglasses with a heart rate monitor
US20060061544A1 (en) * 2004-09-20 2006-03-23 Samsung Electronics Co., Ltd. Apparatus and method for inputting keys using biological signals in head mounted display information terminal
US7629877B2 (en) * 2006-12-19 2009-12-08 Matvey Lvovskiy Multifunctional collimator indicator
US20120242678A1 (en) * 2010-02-28 2012-09-27 Osterhout Group, Inc. See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment
US20120021806A1 (en) * 2010-07-23 2012-01-26 Maltz Gregory A Unitized, Vision-Controlled, Wireless Eyeglass Transceiver
US20120056993A1 (en) * 2010-09-08 2012-03-08 Salman Luqman Dental Field Visualization System with Improved Ergonomics

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10542915B2 (en) 2014-09-05 2020-01-28 Vision Service Plan Systems, apparatus, and methods for using a wearable device to confirm the identity of an individual
US10694981B2 (en) 2014-09-05 2020-06-30 Vision Service Plan Wearable physiology monitor computer apparatus, systems, and related methods
US10188323B2 (en) 2014-09-05 2019-01-29 Vision Service Plan Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual
US9795324B2 (en) 2014-09-05 2017-10-24 Vision Service Plan System for monitoring individuals as they age in place
US10617342B2 (en) 2014-09-05 2020-04-14 Vision Service Plan Systems, apparatus, and methods for using a wearable device to monitor operator alertness
US9649052B2 (en) 2014-09-05 2017-05-16 Vision Service Plan Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual
US10448867B2 (en) 2014-09-05 2019-10-22 Vision Service Plan Wearable gait monitoring apparatus, systems, and related methods
US10307085B2 (en) 2014-09-05 2019-06-04 Vision Service Plan Wearable physiology monitor computer apparatus, systems, and related methods
US11918375B2 (en) 2014-09-05 2024-03-05 Beijing Zitiao Network Technology Co., Ltd. Wearable environmental pollution monitor computer apparatus, systems, and related methods
EP3210095B1 (en) * 2014-10-21 2019-12-11 Signify Holding B.V. System, method and computer program for hands-free configuration of a luminous distribution
US10533855B2 (en) 2015-01-30 2020-01-14 Vision Service Plan Systems and methods for tracking motion, performance, and other data for an individual such as a winter sports athlete
US10215568B2 (en) 2015-01-30 2019-02-26 Vision Service Plan Systems and methods for tracking motion, performance, and other data for an individual such as a winter sports athlete
US10338880B2 (en) 2015-06-03 2019-07-02 Skullcandy, Inc. Audio devices and related methods for acquiring audio device use information
US10171971B2 (en) 2015-12-21 2019-01-01 Skullcandy, Inc. Electrical systems and related methods for providing smart mobile electronic device features to a user of a wearable device
CN105574946A (en) * 2015-12-21 2016-05-11 天津中兴智联科技有限公司 Hand-held ETC reader and employed RFID and AR fusion method thereof
CN105574946B (en) * 2015-12-21 2017-11-28 天津中兴智联科技有限公司 A kind of fusion method of hand-held ETC read write lines and its RFID and AR of use
GB2546883A (en) * 2015-12-21 2017-08-02 Skullcandy Inc Electrical systems and related methods for providing smart mobile electronic device features to a user of a wearable device
CN108540764A (en) * 2017-03-02 2018-09-14 北京戈德思科科技有限公司 A kind of split type wearable augmented reality equipment
US9910298B1 (en) 2017-04-17 2018-03-06 Vision Service Plan Systems and methods for a computerized temple for use with eyewear
US10722128B2 (en) 2018-08-01 2020-07-28 Vision Service Plan Heart rate detection system and method

Also Published As

Publication number Publication date
WO2014144918A3 (en) 2015-01-22

Similar Documents

Publication Publication Date Title
US11630311B1 (en) Enhanced optical and perceptual digital eyewear
US20230266590A1 (en) Enhanced Optical and Perceptual Digital Eyewear
US10795183B1 (en) Enhanced optical and perceptual digital eyewear
CN105142498B (en) Enhanced optical and perceptual digital eyewear
US8733928B1 (en) Enhanced optical and perceptual digital eyewear
WO2014144918A2 (en) Enhanced optical and perceptual digital eyewear
KR102182605B1 (en) Systems and methods for gaze-based media selection and editing
US9519640B2 (en) Intelligent translations in personal see through display
US8963956B2 (en) Location based skins for mixed reality displays
JP2013521576A (en) Local advertising content on interactive head-mounted eyepieces
US20230229010A1 (en) Head-mountable device for posture detection
US11954249B1 (en) Head-mounted systems with sensor for eye monitoring
US20240103608A1 (en) Devices, Methods, and Graphical User Interfaces for Providing Computer-Generated Experiences
US20240103617A1 (en) User interfaces for gaze tracking enrollment
US20240118746A1 (en) User interfaces for gaze tracking enrollment
WO2023244515A1 (en) Head-mountable device with guidance features

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14764731

Country of ref document: EP

Kind code of ref document: A2

122 Ep: pct application non-entry in european phase

Ref document number: 14764731

Country of ref document: EP

Kind code of ref document: A2