US20160071423A1 - Systems and method for monitoring an individual's compliance with a weight loss plan - Google Patents
Systems and method for monitoring an individual's compliance with a weight loss plan Download PDFInfo
- Publication number
- US20160071423A1 US20160071423A1 US14/610,589 US201514610589A US2016071423A1 US 20160071423 A1 US20160071423 A1 US 20160071423A1 US 201514610589 A US201514610589 A US 201514610589A US 2016071423 A1 US2016071423 A1 US 2016071423A1
- Authority
- US
- United States
- Prior art keywords
- wearer
- food
- processor
- weight loss
- recommendations
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/112—Gait analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/10—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
- A61B3/11—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for measuring interpupillary distance or diameter of pupils
- A61B3/112—Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for measuring interpupillary distance or diameter of pupils for measuring diameter of pupils
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0015—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
- A61B5/0022—Monitoring a patient using a global network, e.g. telephone networks, internet
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0077—Devices for viewing the surface of the body, e.g. camera, magnifying lens
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/1032—Determining colour for diagnostic purposes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1103—Detecting eye twinkling
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1113—Local tracking of patients, e.g. in a hospital or private home
- A61B5/1114—Tracking parts of the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1116—Determining posture transitions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1126—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique
- A61B5/1128—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique using image analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/117—Identification of persons
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/117—Identification of persons
- A61B5/1171—Identification of persons based on the shapes or appearances of their bodies or parts thereof
- A61B5/1176—Recognition of faces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/145—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue
- A61B5/1455—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue using optical sensors, e.g. spectral photometrical oximeters
- A61B5/14551—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue using optical sensors, e.g. spectral photometrical oximeters for measuring blood gases
- A61B5/14552—Details of sensors specially adapted therefor
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/165—Evaluating the state of mind, e.g. depression, anxiety
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/318—Heart-related electrical modalities, e.g. electrocardiography [ECG]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/40—Detecting, measuring or recording for evaluating the nervous system
- A61B5/4076—Diagnosing or monitoring particular conditions of the nervous system
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/42—Detecting, measuring or recording for evaluating the gastrointestinal, the endocrine or the exocrine systems
- A61B5/4261—Evaluating exocrine secretion production
- A61B5/4266—Evaluating exocrine secretion production sweat secretion
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/44—Detecting, measuring or recording for evaluating the integumentary system, e.g. skin, hair or nails
- A61B5/441—Skin evaluation, e.g. for skin disorder diagnosis
- A61B5/443—Evaluating skin constituents, e.g. elastin, melanin, water
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/486—Bio-feedback
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4884—Other medical applications inducing physiological or psychological stress, e.g. applications for stress testing
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6802—Sensor mounted on worn items
- A61B5/6803—Head-worn items, e.g. helmets, masks, headphones or goggles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7246—Details of waveform analysis using correlation, e.g. template matching or determination of similarity
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
- A61B5/7278—Artificial waveform generation or derivation, e.g. synthesising signals from measured signals
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B7/00—Instruments for auscultation
- A61B7/02—Stethoscopes
- A61B7/04—Electric stethoscopes
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63B—APPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
- A63B24/00—Electric or electronic controls for exercising apparatus of preceding groups; Controlling or monitoring of exercises, sportive games, training or athletic performances
- A63B24/0062—Monitoring athletic performances, e.g. for determining the work of a user on an exercise apparatus, the completed jogging or cycling distance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/34—User authentication involving the use of external additional devices, e.g. dongles or smart cards
- G06F21/35—User authentication involving the use of external additional devices, e.g. dongles or smart cards communicating wirelessly
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/197—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/23—Recognition of whole body movements, e.g. for sport training
- G06V40/25—Recognition of walking or running movements, e.g. gait recognition
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C9/00—Individual registration on entry or exit
- G07C9/30—Individual registration on entry or exit not involving the use of a pass
- G07C9/32—Individual registration on entry or exit not involving the use of a pass in combination with an identity check
- G07C9/37—Individual registration on entry or exit not involving the use of a pass in combination with an identity check using biometric data, e.g. fingerprints, iris scans or voice recognition
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/04—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
- G08B21/0407—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis
- G08B21/0423—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons based on behaviour analysis detecting deviation from an expected pattern of behaviour or schedule
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/04—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
- G08B21/0438—Sensor means for detecting
- G08B21/0461—Sensor means for detecting integrated or attached to an item closely associated with the person but not worn by the person, e.g. chair, walking stick, bed sensor
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B21/00—Alarms responsive to a single specified undesired or abnormal condition and not otherwise provided for
- G08B21/02—Alarms for ensuring the safety of persons
- G08B21/04—Alarms for ensuring the safety of persons responsive to non-activity, e.g. of elderly persons
- G08B21/0438—Sensor means for detecting
- G08B21/0476—Cameras to detect unsafe condition, e.g. video cameras
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/0092—Nutrition
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H20/00—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
- G16H20/40—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/67—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for remote operation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L63/00—Network architectures or network communication protocols for network security
- H04L63/08—Network architectures or network communication protocols for network security for authentication of entities
- H04L63/0861—Network architectures or network communication protocols for network security for authentication of entities using biometrical features, e.g. fingerprint, retina-scan
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2560/00—Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
- A61B2560/04—Constructional details of apparatus
- A61B2560/0475—Special features of memory means, e.g. removable memory cards
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0219—Inertial sensors, e.g. accelerometers, gyroscopes, tilt switches
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0223—Magnetic field sensors
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0257—Proximity sensors
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2576/00—Medical imaging apparatus involving image processing or analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/0205—Simultaneously evaluating both cardiovascular conditions and different types of body conditions, e.g. heart and respiratory condition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/024—Detecting, measuring or recording pulse rate or heart rate
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/053—Measuring electrical impedance or conductance of a portion of the body
- A61B5/0531—Measuring skin impedance
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/08—Detecting, measuring or recording devices for evaluating the respiratory organs
- A61B5/0816—Measuring devices for examining respiratory frequency
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1118—Determining activity level
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
- A61B5/7282—Event detection, e.g. detecting unique waveforms indicative of a medical condition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61F—FILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
- A61F2/00—Filters implantable into blood vessels; Prostheses, i.e. artificial substitutes or replacements for parts of the body; Appliances for connecting them with the body; Devices providing patency to, or preventing collapsing of, tubular structures of the body, e.g. stents
- A61F2/50—Prostheses not implantable in the body
- A61F2/76—Means for assembling, fitting or testing prostheses, e.g. for measuring or balancing, e.g. alignment means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61F—FILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
- A61F2/00—Filters implantable into blood vessels; Prostheses, i.e. artificial substitutes or replacements for parts of the body; Appliances for connecting them with the body; Devices providing patency to, or preventing collapsing of, tubular structures of the body, e.g. stents
- A61F2/50—Prostheses not implantable in the body
- A61F2/76—Means for assembling, fitting or testing prostheses, e.g. for measuring or balancing, e.g. alignment means
- A61F2002/7695—Means for testing non-implantable prostheses
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C11/00—Non-optical adjuncts; Attachment thereof
- G02C11/10—Electronic devices other than hearing aids
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
Definitions
- a computer-implemented method of monitoring compliance with a weight loss plan by a wearer of computerized eyewear comprising (1) receiving, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear; (2) at least partially in response to receiving the at least one signal from the one or more sensors, determining, by at least one processor, an identity of a food that the wearer is preparing to ingest; (3) determining, by at least one processor, a quantity of the food that the wearer is preparing to ingest; (4) tracking, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest; and (5) comparing, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan.
- determining the identity of the food further comprises capturing an image by the forward facing camera of a packaging associated with the food, at least partially in response to capturing the image, detecting a barcode contained on the packaging, and searching a database of barcodes to determine the nutritional value associated with the food.
- determining a quantity of the food further comprises detecting in the captured image a vessel used by the wearer to measure the quantity of the food that the wearer is preparing to ingest, and detecting at least one marking on the vessel that indicates the quantity of the food placed in the vessel.
- a system for monitoring compliance with a weight loss plan by a wearer of computerized eyewear comprises one or more processors and memory operatively coupled to the one or more processors.
- the one or more processors is configured to (1) receive at least one signal generated by one or more sensors operatively coupled to the computerized eyewear; (2) at least partially in response to receiving the at least one signal from the one or more sensors, determine an identity of a food that the wearer is preparing to ingest; (3) determine a quantity of the food that the wearer is preparing to ingest; (4) track the identity and the quantity of the food that the wearer is preparing to ingest; (5) identify the date and time that the food is ingested by the wearer; and (6) compare the identity and the quantity of the food that the wearer ingests to a predetermined dietary plan.
- a computer-implemented method of monitoring compliance with a weight loss plan by a wearer of computerized eyewear comprising the steps of (1) receiving, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear; (2) at least partially in response to receiving the at least one signal from the one or more sensors, determining, by at least one processor, an identity of a food that the wearer is preparing to ingest; (3) determining, by at least one processor, a quantity of the food that the wearer is preparing to ingest; (4) comparing, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan; (5) at least partially in response to comparing the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined dietary plan, calculating, by at least one processor, one or more recommendations to assist the wearer in complying with the predetermined dietary plan; and (6) notifying, by at least one processor,
- FIG. 1 is a block diagram of a weight loss compliance system in accordance with an embodiment of the present system.
- FIG. 2 is a block diagram of the compliance server of FIG. 1 .
- FIG. 3 is an exemplary wearable health monitoring device for use in the weight loss compliance system of FIG. 1
- FIG. 4 illustrates a flowchart that generally illustrates various steps executed by a Weight Loss Compliance Module according to a particular embodiment.
- a weight loss compliance system may include a suitable wearable device that is configured to monitor one or more of a wearer's (e.g., a user) food consumption, physical activities, sleep patterns, and/or compliance with a medicine regime (e.g., weight loss medicine/supplement).
- a wearer's e.g., a user
- a medicine regime e.g., weight loss medicine/supplement
- the system is configured to gather data, for example, using one or more sensors, about the wearer (e.g., such as the wearer's body temperature, heart rate, level of physical activity, geographic location, distance traveled, diet (e.g., food consumed, calories for each meal, total calories, etc.), compliance with a prescribed medical regimen (e.g., medications recently taken, dose, side effects), sleeping patterns (e.g., hours slept, type of sleep, quality of sleep), etc.).
- a prescribed medical regimen e.g., medications recently taken, dose, side effects
- sleeping patterns e.g., hours slept, type of sleep, quality of sleep, etc.
- the system is configured to, for example: (1) identify a food and quantity of the food that the wearer is preparing to ingest; (2) track the identity and the quantity of food (e.g., the nutritional value, the calories associated with the food, the date and time the food is ingested, etc.); (3) compare the identity and quantity of the food to a predetermined weight loss plan (e.g., a dietary plan); (4) calculate one or more recommendations to assist the wearer in complying with the predetermined weight loss plan; and/or (5) notify the wearer of the one or more recommendations.
- a predetermined weight loss plan e.g., a dietary plan
- the wearable device is adapted to facilitate the transmission of the data captured by the one or more sensors to a remote computing device (e.g., a handheld computing device, a central server, or any other suitable smart device that may, in various embodiments, contain a wireless communications device that can connect to the wearable computing device) that analyzes the information to determine whether the wearer is complying with the predetermined weight loss plan (e.g., a dietary plan, an exercise plan, a medicine plan, a sleep plan, or a combination of one or more of the proceeding).
- a remote computing device e.g., a handheld computing device, a central server, or any other suitable smart device that may, in various embodiments, contain a wireless communications device that can connect to the wearable computing device
- the predetermined weight loss plan e.g., a dietary plan, an exercise plan, a medicine plan, a sleep plan, or a combination of one or more of the proceeding.
- the system may provide the wearer feedback, warnings, and/or recommendations on (1) how to better comply with the weight plan (e.g., eat more green vegetables or fruit, don't eat fast food, you should eat smaller portions, etc.), (2) how a particular food or quantity will impact the weight loss plan (e.g., if you eat this, you must perform a particular physical activity to offset the calories, etc.), and/or (3) how a medicine can impact the wearer's ability to lose weight (e.g., a particular side effect of the medicine is weight gain/loss, etc.).
- the weight plan e.g., eat more green vegetables or fruit, don't eat fast food, you should eat smaller portions, etc.
- how a particular food or quantity will impact the weight loss plan (e.g., if you eat this, you must perform a particular physical activity to offset the calories, etc.)
- how a medicine can impact the wearer's ability to lose weight (e.g., a particular side effect of the
- Suitable wearable devices may include, for example: (1) a pair of eyewear (e.g., goggles or eyeglasses); (2) one or more contact lenses; (3) a wristwatch; (4) an article of clothing (e.g., such as a suitable shirt, pair of pants, undergarment, compression sleeve, etc.); (5) footwear; (6) a hat; (7) a helmet; (8) an adhesive strip or other tag that may be selectively attached to an individual or the individual's clothing; (9) a computing device that is embedded into a portion of an individual's body (e.g., under the individual's skin, or within a medical device, such as a pacemaker); (10) an orthopedic cast; or (11) any other suitable wearable item.
- a pair of eyewear e.g., goggles or eyeglasses
- a wristwatch e.g., a wristwatch
- an article of clothing e.g., such as a suitable shirt, pair of pants, undergarment, compression s
- the wearable device comprises one or more sensors that are configured to sense the wearer's food intake, physical activities, sleep patterns, and medicine intake (e.g., prescription weight loss drugs, over-the-counter weight loss drugs, nutritional supplements, prescription drugs, etc.).
- medicine intake e.g., prescription weight loss drugs, over-the-counter weight loss drugs, nutritional supplements, prescription drugs, etc.
- the one or more sensors may include, for example: (1) one or more forward facing cameras; (2) one or more global positioning sensors; (3), one or more olfactory sensors; (4) one or more pedometers; (5) one or more microphones; (6) one or more accelerometers; (7) one or more blood pressure sensors; (8) one or more pulse oximeters; (9) one or more respiratory rate sensors; (10) one or more near-field communication chips; (11) one or more gyroscopes; (12) one or more geomagnetic sensors; (13) one or more global positioning system chips; and/or (14) any other suitable one or more sensors.
- the system's wearable device includes a pair of computerized eyewear that comprises the one or more sensors for monitoring the wearer's compliance with a weight loss plan.
- the one or more sensors are coupled to (e.g., connected to, embedded in, etc.) the pair of glasses, which may be, for example, a pair of computerized or non-computerized eyeglasses.
- the wearer is a person trying to lose weight, become more physically active, under orders of a doctor to lose weight for medical reasons or any other person trying to live a more active and healthy life.
- the present systems and methods may be, for example, embodied as a computer system, a method, or a computer program product. Accordingly, various embodiments may be entirely hardware, entirely software, or a combination of hardware and software. Furthermore, particular embodiments may take the form of a computer program product stored on a computer-readable storage medium having computer-readable instructions (e.g., software) embodied in the storage medium. Various embodiments may also take the form of Internet-implemented computer software. Any suitable computer-readable storage medium may be utilized including, for example, hard disks, compact disks, DVDs, optical storage devices, and/or magnetic storage devices.
- the computer instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on a user's computer and partly on a remote computer, or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including but not limited to: (1) a local area network (LAN); (2) a wide area network (WAN); (3) a cellular network; or (4) the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the function specified in the flowchart block or blocks.
- the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process (e.g., method) such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
- FIG. 1 is a block diagram of a weight loss compliance system 100 according to particular embodiments.
- the weight loss compliance system 100 includes one or more networks 115 , one or more third party servers 50 , a compliance server 120 that includes a weight loss compliance module 400 , a dietary information database 140 , an exercise information database 142 , a medicine database 144 , a sleep information database 146 , one or more remote computing devices 154 (e.g., such as a smart phone, a tablet computer, a wearable computing device, a laptop computer, a desktop computer, a Bluetooth device, an automated dialing apparatus, etc.), and one or more health monitoring devices 156 , which may, for example, be embodied as one or more of eyewear, headwear, clothing, a watch, a hat, a helmet, a cast, an adhesive bandage, a piece of jewelry (e.g., a ring, earring, necklace, bracelet, etc.), or any other suitable wearable device.
- a piece of jewelry
- the one or more computer networks 115 facilitate communication between the one or more third party servers 50 , the compliance server 120 , the dietary information database 140 , the exercise information database 142 , the medicine database 144 , the sleep information database 146 , the one or more remote computing devices 154 , and/or the one or more health monitoring devices 156 .
- the one or more networks 115 may include any of a variety of types of wired or wireless computer networks such as the Internet, a private intranet, a mesh network, a public switch telephone network (PSTN), or any other type of network (e.g., a network that uses Bluetooth or near field communications to facilitate communication between computing devices).
- the communication link between the one or more remote computing devices 154 and the compliance server 120 may be, for example, implemented via a Local Area Network (LAN) or via the Internet.
- LAN Local Area Network
- FIG. 2 illustrates a diagrammatic representation of the architecture for the compliance server 120 that may be used within the weight loss compliance system 100 .
- the computer architecture shown in FIG. 2 may also represent the computer architecture for any one of the one or more remote computing devices 154 , one or more third party servers 50 , and one or more health monitoring devices 156 , shown in FIG. 1 .
- the compliance server 120 may be suitable for use as a computer within the context of the weight loss compliance system 100 that is configured for monitoring the compliance of the wearer with a weight loss plan (e.g., food intake, physical activities, sleeping habits and medicine regimes, etc.).
- a weight loss plan e.g., food intake, physical activities, sleeping habits and medicine regimes, etc.
- the compliance server 120 may be connected (e.g., networked) to other computing devices in a LAN, an intranet, an extranet, and/or the Internet as shown in FIG. 1 .
- the compliance server 120 may operate in the capacity of a server or a client computing device in a client-server network environment, or as a peer computing device in a peer-to-peer (or distributed) network environment.
- the compliance server 120 may be a desktop personal computing device (PC), a tablet PC, a set-top box (STB), a Personal Digital Assistant (PDA), a cellular telephone, a web appliance, a network router, a switch or bridge, or any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device.
- PC personal computing device
- PDA Personal Digital Assistant
- STB set-top box
- a cellular telephone a web appliance
- network router a network router
- switch or bridge or any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device.
- any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device.
- the term “computing device” shall also be interpreted to include any collection of computing devices that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed here
- An exemplary compliance server 120 includes a processing device 202 , a main memory 204 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.), a static memory 206 (e.g., flash memory, static random access memory (SRAM), etc.), and a data storage device 218 , which communicate with each other via a bus 232 .
- main memory 204 e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.
- DRAM dynamic random access memory
- SDRAM synchronous DRAM
- RDRAM Rambus DRAM
- static memory 206 e.g., flash memory, static random access memory (SRAM), etc.
- SRAM static random access memory
- the processing device 202 represents one or more general-purpose or specific processing devices such as a microprocessor, a central processing unit (CPU), or the like. More particularly, the processing device 202 may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or processor implementing other instruction sets, or processors implementing a combination of instruction sets.
- the processing device 202 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like.
- the processing device 202 may be configured to execute processing logic 226 for performing various operations and steps discussed herein.
- the compliance server 120 may further include a network interface device 208 .
- the compliance server 120 may also include a video display unit 210 (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alpha-numeric input device 212 (e.g., a keyboard), a cursor control device 214 (e.g., a mouse), and a signal generation device 216 (e.g., a speaker).
- a video display unit 210 e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)
- an alpha-numeric input device 212 e.g., a keyboard
- a cursor control device 214 e.g., a mouse
- a signal generation device 216 e.g., a speaker
- the data storage device 218 may include a non-transitory computing device-accessible storage medium 230 (also known as a non-transitory computing device-readable storage medium or a non-transitory computing device-readable medium) on which is stored one or more sets of instructions (e.g., the weight loss compliance module 400 ) embodying any one or more of the methodologies or functions described herein.
- the one or more sets of instructions may also reside, completely or at least partially, within the main memory 204 and/or within the processing device 202 during execution thereof by the compliance server 120 —the main memory 204 and the processing device 202 also constituting computing device-accessible storage media.
- the one or more sets of instructions may further be transmitted or received over a network 115 via a network interface device 208 .
- computing device-accessible storage medium 230 is shown in an exemplary embodiment to be a single medium, the term “computing device-accessible storage medium” should be understood to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions.
- the term “computing device-accessible storage medium” should also be understood to include any medium that is capable of storing, encoding, or carrying a set of instructions for execution by the computing device and that causes the computing device to include any one or more of the methodologies of the present invention.
- the term “computing device-accessible storage medium” should accordingly be understood to include, but not be limited to, solid-state memories, optical and magnetic media, etc.
- eyewear 300 includes: (1) an eyewear frame 310 ; (2) a first temple 312 ; and (3) a second temple 314 , as described below.
- eyewear 300 in various embodiments, includes any suitable eyewear frame 310 configured to support one or more lenses 318 , 320 .
- the eyewear frame 310 has a first end 302 and a second end 304 .
- the eyewear frame 310 may be made of any suitable material such as metal, ceramic, polymers or any combination thereof.
- the eyewear frame 310 is configured to support the first and second lenses 318 , 320 about the full perimeter of the first and second lenses 318 , 320 .
- the eyewear frame 310 may be configured to support the first and second lenses 318 , 320 about only a portion of each respective lens.
- the eyewear frame 310 is configured to support a number of lenses other than two lenses (e.g., a single lens, a plurality of lenses, etc.).
- the lenses 318 , 320 may include prescription lenses, sunglass lenses, or any other suitable type of lens (e.g., reading lenses, non-prescription lenses), which may be formed from glass or polymers.
- the eyewear frame 310 includes a first and second nose pad 322 (not shown in figures), 324 , which may be configured to maintain the eyewear 300 adjacent the front of a wearer's face such that the lenses 318 , 320 are positioned substantially in front of the wearer's eyes while the wearer is wearing the eyewear 300 .
- the nose pads 322 , 324 may comprise a material that is configured to be comfortable when worn by the wearer (e.g., rubber, etc.).
- the nose pads may include any other suitable material (e.g., plastic, metal, etc.).
- the nose pads may be integrally formed with the frame 310 .
- the eyewear frame 310 includes a first and second hinge 326 , 328 that attach the first and second temples 312 , 314 to the frame first and second ends 302 , 304 , respectively.
- the hinges may be formed by any suitable connection (e.g., tongue and groove, ball and socket, spring hinge, etc.).
- the first hinge 326 may be welded to, or integrally formed with, the frame 310 and the first temple 312 and the second hinge 328 may be welded to, or integrally formed with, the frame 310 and the second temple 314 .
- the first temple 312 is rotatably connected to the frame 310 at a right angle to extend the first temple 312 substantially perpendicular, substantially parallel, or anywhere in between the right angle to the frame 310 .
- the first temple 312 has a first and second end 312 a , 312 b .
- the first temple 312 includes an earpiece 313 configured to be supported by a wearer's ear.
- the second temple 314 is rotatably connected to the frame 310 at a right angle to extend the second temple 314 substantially perpendicular, substantially parallel, or anywhere in between the right angle to the frame 310 .
- the second temple 314 has a first and second end 314 a , 314 b .
- the second temple 314 includes an earpiece 315 configured to be supported by a wearer's ear.
- the second temple 314 has one or more sensors 330 connected to the second temple 314 .
- the one or more sensors 330 may be coupled to the frame 310 , the first and second temples 312 , 314 , the first and second lenses 318 , 310 , the nose piece 324 , or any other portion of the eyewear 300 in any suitable way.
- the one or more sensors 330 may be embedded into the eyewear 300 , coupled to the eyewear 300 , and/or operatively coupled to the eyewear 300 .
- the one or more sensors may be formed at any point along the eyewear 300 .
- a fingerprint reader may be disposed adjacent the first temple of the eyewear 300 .
- the one or more sensors may be formed in any shape.
- the one or more sensors may be formed on the inner (back) surface of the frame 310 , the first and second temples 312 , 414 , the first and second lenses 318 , 310 , or any other portion of the eyewear 300 .
- the one or more sensors may be formed on the outer (front) surface of the frame 310 , the first and second temples 312 , 314 , the first and second lenses 318 , 310 , or any other portion of the eyewear 300 .
- the one or more sensors 330 that are coupled to the eyewear (or other wearable device) are adapted to detect the identity of a food, the quantity of a food, the identity of a medicine, the dose of the medicine, one or more physical activities performed by the wearer, the length of the physical activity, etc.
- the one or more sensors coupled to the eyewear or other health monitoring device may include, for example, one or more of the following: a near-field communication chip, a gyroscope, a Bluetooth chip, a GPS unit, a forward facing camera, an olfactory sensor, a heart rate monitor, an electrocardiogram (EKG), an electroencephalogram (EEG), a pedometer, a thermometer, a microphone, an accelerometer, a magnetometer, a blood pressure sensor, a pulse oximeter, a skin conductance response sensor, a blood sensor, or any other suitable sensor.
- the sensors coupled to the eyewear may include one or more electronic communications devices such as a near field communication chip, a Bluetooth chip, a forward facing camera, a microphone and a GPS unit.
- the one or more sensors are coupled to a computing device that is associated with (e.g., embedded within, attached to) the eyewear or other wearable device.
- the eyewear or other wearable device comprises at least one processor, computer memory, suitable wireless communications components (e.g., a Bluetooth chip) and a power supply for powering the wearable device and/or the various sensors.
- the one or more sensors may be coupled to a Bluetooth device that is configured to transmit the one or more signals to a handheld wireless device.
- the system is configured to receive input from a user (e.g., a wearer of the eyewear) via one or more gestures, for example, using at least one of the sensors described above.
- the system may, for example, be configured to: (1) identify a gesture performed by the user; and (2) at least partially in response to identifying the gesture, perform a function associated with the gesture.
- the system may be configured to perform a particular function in response to identifying a particular gesture, where the particular gesture is associated with the particular function.
- the system may be configured to enable the user to provide one or more gestures for performing a particular function.
- the system may, for example: (1) receive a selection of a particular function from the user; (2) receive input of one or more gestures from the user; and (3) associate the particular function with the one or more gestures.
- the one or more gestures may include, for example: (1) one or more hand gestures (e.g., a thumbs up, a wave, two thumbs up, holding up any particular number of fingers, making one or more fists, performing a particular movement with one or more hands, etc.); (2) one or more head movements (e.g., shaking of the user's head, a nod, etc.); (3) one or more eye movements (e.g., looking in a particular direction for a particular period of time, a wink, blinking, blinking in a particular pattern, etc.); (4) one or more facial movements (e.g., a smile, a frown, sticking out of a tongue, etc.); and/or (5) any suitable combination of these or any other suitable gestures.
- hand gestures e.g., a thumbs up, a wave, two thumbs up, holding up any particular number of fingers, making one or more fists, performing a particular movement with one or more hands, etc.
- head movements
- the system is configured to identify the one or more gestures, for example, using a suitable imaging device (e.g., camera) that is part of the system.
- a suitable imaging device e.g., camera
- the imaging device may be directed toward an area in front of the user while the user is wearing the eyewear 300 and configured to identify gestures performed by the user's hands, arms, feet, legs, etc.
- the system may include an imaging device directed toward the user's face and/or eyes while the user is wearing the eyewear 300 that is configured to identify gestures performed by the user's face and/or eyes.
- the system comprises one or more gyroscopes and/or accelerometers configured to determine a position or change in position of the eyewear 300 while the user is wearing the eyewear.
- the one or more gyroscopes and/or accelerometers are configured to identify one or more gestures performed by the user that include one or more gestures that include movement of the user's head.
- the system comprises one or more gyroscopes and/or one or more accelerometers, disposed on any other portion of the user's body, configured to identify any gesture performed by the user using the other portion of the user's body (e.g., arm, hand, leg, foot, etc.).
- the system comprises any other suitable sensor for identifying one or more gestures performed by the user.
- one or more of the sensors may be detachable from the eyewear. For instance, if a wearer does not need a temperature sensor or other particular sensor, the sensor may be removed from the eyewear. In still other embodiments, some of the one or more sensors may be coupled to the handheld wireless device, while other of the one or more sensors may be coupled to the wearable device.
- a system is adapted to monitor a wearer's compliance with a weight loss plan.
- Various aspects of the system's functionality may be executed by certain system modules, including the weight loss compliance module 400 , which is discussed in greater detail below.
- FIG. 4 is a flow chart of operations performed by an exemplary weight loss compliance module 400 , which may, for example, run on the compliance server 120 , or any suitable computing device (such as the one or more health monitoring devices 156 , a handheld computing device coupled to communicate with the one or more health monitoring devices 156 or a suitable mobile computing device).
- the weight loss compliance module 400 may assess a wearer's compliance with a weight loss plan (e.g., a dietary plan, an exercise plan, a sleep plan, and/or a medicine regime) and provide recommendations, feedback and/or supportive messages on how the wearer can improve their compliance with the weight loss plan.
- a weight loss plan e.g., a dietary plan, an exercise plan, a sleep plan, and/or a medicine regime
- the system begins, in various embodiments, at Step 405 where the system receives, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear.
- the at least one signal received may include one or more images taken by a forward facing camera coupled to the health monitoring device 156 .
- the at least one signal received may include one or more signals from a global positioning system unit that are associated with the location of the wearer of the health monitoring device 156 .
- the at least one signal received may include one or more signals from an olfactory sensor that may be used to determine the smell associated with one or more foods being prepared for ingestion by the wearer of the health monitoring device 156 .
- the at least one signal received may include one or more audio signals received by a microphone coupled to the health monitoring device 156 .
- the system receives the signals substantially automatically after the sensor generates the signal.
- the system may receive the signal periodically (e.g., by the second, by the minute, hourly, daily, etc.). For example, the system may receive the signal every thirty seconds throughout the day.
- the system may receive the signal after receiving an indication from the wearer or a third party that the system should receive the signal. For instance, the wearer may speak a voice command to the wearable device requesting that the device track the food being prepared by the wearer.
- the system may receive an indication from the wearer or a third party of when to have the system receive the signal.
- the system may receive an indication from the wearer or a third party to have particular data received from a particular sensor at the same time that the system receives second particular data from a second particular sensor. For example, when the system receives one or more images, the system may at least partially in response to receiving the one or more images, also obtain global position data of the wearer, olfactory data associated with food being prepared by the wearer, etc.
- the system may store data associated with the received signal in memory (e.g., local memory, remote memory, etc.). In various embodiments, the system may store the data substantially automatically after receiving the signal. In other embodiments, the system may store the data after receiving manual input from the wearer or a third party requesting that the system store the data. In various embodiments, the system may store the data for a specified period of time. For instance, the system may store the data for a day, a month, a year, etc., in, for example, the dietary information database 140 , the exercise information database 142 , the medicine database 144 , and/or the sleep information database 146 . In some embodiments, the system may store the data on any suitable server, database, or device. In other embodiments, the system may store the data on the compliance server 120 . In still other embodiments, the system may store data in an account associated with the wearer. In various embodiments, the system may store the data with a timestamp of when the data was received.
- memory e.g., local
- the system determines, by at least one processor, an identity of a food that the wearer is preparing to ingest.
- the at least one signal may be an image that is captured by a forward facing camera.
- the at least one signal may be a signal received from a GPS chip that provides the current location of the wearer.
- the at least one signal may be an audio signal received from a microphone.
- the signal may be a smell captured by the olfactory sensor.
- the system may capture an image of packaging associated with a food that is being prepared by the wearer.
- the system may analyze the image to detect a barcode located on the packaging associated with the food. Once the system detects the barcode, the system may decipher the bar code and search the dietary information database 140 of barcodes to find a matching barcode. Once the system identifies the food associated with the barcode, the system may also obtain from the dietary information database 140 nutritional information (e.g., calories, fat content, protein content, etc.) that is associated with the matching barcode.
- nutritional information e.g., calories, fat content, protein content, etc.
- the system may analyze the captured one or more images to identify a food located in the captured one or more images. For example, the system may capture an image of a hamburger that the wearer is preparing to eat. The system may search the dietary information database 140 of images of food to find a matching image of a hamburger that is similar (e.g., the same, substantially similar, etc.) to the hamburger that the wearer is preparing. Once a match is determined by the system, the system can identify the food and obtain the nutritional information associated with matching image stored in the dietary information database 140 . In various embodiments, the system may capture a series of images that show the wearer preparing the hamburger. Thus, in some embodiments, the system may first capture an image of the hamburger as the wearer barbecues the hamburger.
- the system may identify the hamburger by finding a matching image of a hamburger in the dietary information database 140 .
- the system may capture one or more images of the user removing a hamburger roll from a package of hamburger rolls.
- the system may detect a barcode from the image of the packaging of the hamburger rolls and identify the roll and the nutritional information associated with the roll from the barcode associated with the one or more images by searching the dietary information database 140 for a matching barcode.
- the system may continue to identify various other condiments that the wearer places on the hamburger in a similar fashion.
- the system may capture an image of a food that the wearer is preparing to ingest while the system simultaneously receives a GPS signal from a GPS chip contained in the computerized eyewear worn by the wearer.
- the system may determine the location of the wearer from the GPS signal, determine if a restaurant or similar food establishment is located at the location, and simultaneously analyze one or more captured images to identify the food that the wearer is preparing to ingest by, for example, one of the methods described above.
- the wearer's location and a determination of a particular restaurant or similar food establishment may assist the system in identifying the food that the wearer is preparing to ingest.
- the system may display a list of one or more food items associated with the restaurant or similar food establishment on a display associated with the computerized eyewear (e.g., a display coupled to the computerized eyewear, a display on a handheld computing device, etc.) and allow the wearer to select the particular food item from the displayed list of one or more food items.
- a display associated with the computerized eyewear e.g., a display coupled to the computerized eyewear, a display on a handheld computing device, etc.
- the system may receive an audio signal from one or more microphones coupled to the computerized eyewear.
- the system may convert the received audio signal to text and use the text to search the dietary information database 140 for matching text words to identify the food that the wearer is preparing to ingest.
- the dietary information database 140 may also contain nutritional information associated with each word in the database.
- the wearer may place the computerized eyewear in a state where it is ready to accept voice input. The wearer may, for example, say “hamburger”.
- the system may convert the received voice command to text and search the dietary information database 140 for a matching word.
- Each word in the database may contain associated nutritional information.
- the system finds a match for the word “hamburger”, the system identifies the food item as a “hamburger” and obtains nutritional information associated with the word “hamburger.”
- the system may seek confirmation of the converted word by the wearer before searching the database of food items.
- the system may determine a wearer's location using a signal received from the GPS chip in the computerized eyewear or from a GPS chip on a handheld computing device coupled to the computerized eyewear.
- the system may begin to monitor the wearer's speech in order to detect one or more food items ordered by the wearer.
- the system may convert the speech to text and search the dietary information database 140 based on one or more of: (1) the location of the wearer, (2) the associated name of the restaurant or other similar food establishment, or (3) a detected name of a food item ordered by the wearer.
- the system may automatically detect the ordered food item. In other embodiments, once the system identifies an ordered item, the system may request confirmation that the detected item is accurate prior to making a final identification of the food item that the wearer is preparing to ingest.
- confirmation of the identity of the food item may be made by the wearer orally using the microphone on the computerized eyewear.
- the system may request confirmation by displaying the identified food item on a display associated with the computerized eyewear.
- the computerized eyewear may contain a built in display that the wearer can view while wearing the computerized eyewear.
- the computerized eyewear may be coupled (e.g., wirelessly, wired, etc.) to a handheld computing device that displays the selected food item.
- the system may be configured to allow the user to select a “confirm” or “deny” button displayed on the handheld computing device.
- the system determines, by at least one processor, a quantity of the food that the wearer is preparing to ingest.
- the system may determine the serving size associated with the barcode by searching the dietary information database 140 for a matching barcode and nutritional information, which may also include a serving size for the food item.
- the matching barcode may indicate that the serving size associated with the packaged food is a single serving (e.g., a candy bar, a frozen meal, etc.).
- the matching barcode may indicate that the serving size is a teaspoon, a tablespoon, a cup, etc.
- the system may calculate an estimated volume from the captured one or more image.
- the system may capture and analyze one or more images of the food and calculate an estimate of the quantity of the food that the wearer is preparing to ingest. For example, in some embodiments, the system may analyze the captured one or more images to count the number of teaspoons, tablespoons or cups of a food that the wearer selects.
- the wearer may input the quantity of food using an input device (e.g, a food scale) coupled to the computerized eyewear or a handheld computing device coupled to the computerized eyewear or the wearer may input the quantity using an audio input via the one or more microphones coupled to the computerized eyewear.
- an input device e.g, a food scale
- the system may analyze the captured one or more images to detect a vessel (e.g., bowl, measuring cup, container, etc.) that the wearer is using to measure a quantity of food.
- the system may determine the quantity of the selected food based on the identity of the vessel. For example, a set of measuring cups and/or spoons may be color coded based on the size of each cup or spoon.
- the system may analyze the captured one or more images to detect the particular measuring cup and/or spoon being used.
- the system may analyze the captured one or more images to detect one or more markings located on the vessel that would allow the system to determine an accurate measurement of the food based on the identity of the vessel and/or the one or more markings on the vessel.
- the system may be configured to operatively couple (e.g. via Bluetooth, Wi-Fi, physical wiring, etc.) to a measuring device such as a scale, an electronic measuring cup, etc., and use a received signal from the measuring device to determine the quantity of the selected food.
- a measuring device such as a scale, an electronic measuring cup, etc.
- the system tracks, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest.
- the system may store the identity and quantity of a food that a wearer prepares to ingest in an account associated with the wearer on the system.
- the system may allow the user to input when the wearer plans on eating the prepared food (e.g., for breakfast, lunch, dinner, midmorning snack, late afternoon snack, etc.).
- the system may determine a date and time that the wearer ingests the prepared food.
- the system may determine the date, time, and associated meal (e.g., breakfast, lunch, dinner, etc.) and stored the information in a database (e.g., dietary information database 140 ) or an account on the system associated with the wearer.
- a database e.g., dietary information database 140
- the system may track the number of calories associated with the prepared food, the nutritional information (e.g., amount of fat, protein, carbohydrates, vitamins, minerals, etc.) associated with the prepared food, and also store this information in the database or account associated with the wearer.
- the system compares, by at least one processor, the identity and quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan.
- the weight loss plan may contain one or more of a dietary plan, an exercise plan, a medicine regime, or a sleep plan.
- the weight loss plan may be prepared by the wearer.
- the weight loss plan may be prepared by a third-party (e.g., a doctor, a trainer, a nutritionist, etc.).
- the dietary plan for the wearer may consist of at least one of: (A) one or more daily values for: (1) total calorie intake, (2) total fat intake, (3) total protein intake, (4) total carbohydrate intake; (B) one or more prohibited foods; (C) one or more prohibited vitamins; (D) one or more prohibited minerals; (E) a daily expected weight loss value; (F) a weekly expected weight loss value; (G) a recommended amount of daily sleep; (H) etc.
- the system may compare the amount and quantity of food and its associated nutritional information to the dietary plan values in order to determine if the wearer is complying with the dietary plan.
- the system may compare the actual consumed total daily calorie intake to the planned total daily calorie intake.
- the system may, in addition to the total daily calorie intake also compare one or more of the actual daily protein intake, daily fat intake or daily carbohydrate intake to the planned daily values.
- the exercise plan may consist of at least one of: (A) a weight loss goal; (B) a target daily calories to burn in order to lose a particular amount of weight per week; (C) various suggested exercises to achieve the desired weight loss goal; or (D) one or more recommended daily physical activities (e.g., walking, stretching, etc.).
- the system may be configured to track one or more of the wearer's (1) movements, (2) heart rate, or (3) respiration rate to determine the amount of calories burned by the wearer and the type of exercise performed by the wearer.
- the system may use signals received from the one or more sensors (e.g., gyroscope, accelerometer, GPS unit, heart rate sensor, etc.) to determine the wearer's movements and calories burned.
- the system may determine the wearer's current movements by tracking the distance traveled by the wearer for a particular day.
- the system may determine the wearer's current movements by tracking the orientation of the wearer using the gyroscope.
- the current movements of the wearer may include actions such as running, jumping, bicycling, weight lifting, sitting, standing, etc.
- the system may be configured to determine a total calories burned by the wearer for any particular activity based on one or more of (1) the movements associated with the wearer during the activity, (2) the amount of elapsed time during which the activity is performed, (3) the wearer's heart rate (actual, average, mean, etc.) during the time the activity is performed and/or after completion of the activity, or (4) the wearer's respiration rate during and/or after the activity is performed by the wearer.
- the system determines the wearer's current movements, heart rate, and/or respiration rate substantially automatically after receiving the signals from the one or more sensors.
- the system may determine the wearer's current movements, heart rate, and/or respiration rate periodically (e.g., by the second, by the minute, hourly, daily, etc.). For example, the system may determine the wearer's current movements, heart rate, and/or respiration rate every thirty seconds throughout the day. In other embodiments, the system may determine the wearer's current movements, heart rate, and/or respiration rate after receiving an indication from the wearer that the system should analyze the signals received from the one or more sensors.
- the wearer may speak a voice command to the computerized eyewear requesting that the device analyze the wearer's steps taken, heart rate and/or respiration rate.
- the wearer may indicate to the system using a handheld computing device operatively coupled to the computerized eyewear that the system should capture the wearer's current movements, heart rate, and/or respiration rate.
- the weight loss plan may contain a medicine regime that is designed to support the wearer's weight loss goals.
- the medicine regime may comprise at least one of (A) one or more prescription weight loss medicines; (B) one or more over the counter weight loss medicines; or (C) one or more over the counter supplements (e.g. Chitosan, Chromium Picolinate, etc.).
- the medicine regime may also comprise tracking one or more prescription medicines that the wearer is already taking to determine any weight loss or weight gain side effects that may be caused by the one or prescribed medications taken by the wearer. For example, if the wearer is diabetic, the medicine regime may track the wearer's use of insulin in conjunction with the wearer's food intake, physical activity, and glucose levels to determine any weight loss or weight gain effects on the wearer.
- the wearer or a third-party may manually input a medicine regime into an account associated with the wearer on the system.
- a medicine regime into an account associated with the wearer on the system.
- the wearer may manually enter one or more prescription medicines into the wearer's account on the system or the wearer's account may be linked to a medical records database so that prescribed medicines are automatically linked to the wearer's weight loss plan.
- the one or more sensors on the computerized eyewear may capture one or more images of the one or more medicines/supplements that the wearer is preparing to ingest.
- the system may be configured to analyze the captured image to detect the presence of a label on a medicine/supplement bottle to identify the medicine/supplement and dose being ingested by the wearer.
- the system may perform optical character recognition techniques on the label to identify the type of medicine/supplement and the dose taken by the wearer.
- the system may add the medicine/supplement to the wearer's medicine regime and perform a look-up of the medicine/supplement in a medicine database 144 to determine if the medicine/supplement has any positive or negative side effect on weight loss.
- the system may track the wearer's compliance with the medicine regime.
- the system may capture one or more images throughout the day and analyze the images to determine one or more of (1) a type of medicine/supplement taken by the wearer; (2) the time the medicine/supplement is ingested by the wearer; and (3) the dose of medicine/supplement ingested by the wearer.
- the system may detect one or more pills in the captured one or more images, compare the one or more detected pills found in the captured one or more images to known images of pills stored in the medicine database 144 , identify the one or more pills by matching the one or more pills from the captured one or more images to the known images of pills stored in the medicine database 144 , and detect the time that the captured one or more images was taken. Based on this information, the system is able to determine if the wearer is complying with the medicine regime prescribed by the wearer's physician in addition to monitoring the potential side effects of medicine/supplements that the wearer takes while following the weight loss plan.
- the weight loss plan may include a sleep plan that is designed to support the wearer's weight loss goals.
- the sleep plan for the wearer may consist of at least one of: (1) a recommended amount of sleep; (2) a recommended time for the wearer to go to sleep; (3) a recommended time for the wearer to wake up; and/or (4) a recommended nap time.
- the sleep plan may include a position for the wearer to sleep in, a temperature for the room where the wearer is sleeping, and/or a lighting condition for the room where the wearer is sleeping.
- the sleep plan may include advice for the wearer on avoiding certain types of light.
- the sleep plan for the wearer may include suggestions for relaxation techniques.
- the system may track the amount and quality of sleep the wearer is obtaining to determine if the wearer is complying with the sleep plan.
- the system may do this, for example, using one or more of the sensors (e.g., motion sensors) described herein.
- the system may compare the actual daily sleep of the wearer to the prescribed sleep plan.
- the system may, in addition to tracking the wearer's sleep patterns, track the wearer's hormone levels (e.g., through perspiration), the wearer's glucose levels, and the wearer's activity levels.
- the system calculates, using at least one processor, one or more recommendations to assist the wearer in complying with the predetermined weight loss plan.
- the system may analyze the wearer's calorie intake, compare the current calorie intake to the predetermined weight loss plan, analyze the wearer's current calories burned and calculate one or more activities that would allow the wearer meet the daily goals for the predetermined weight loss plan.
- the system may determine that the wearer is preparing to ingest a 500 calorie hamburger. In some such embodiments, the system may recommend that the wearer run for 45 minutes to offset the additional 500 calories.
- the system may determine that the wearer is preparing to ingest a food item that contains 1500 mg of sodium. If the predetermined weight loss plan contains a restriction on sodium, the system may recommend that the wearer not eat the food item since the food item contains too much sodium. In still other embodiments, the system may be configured to recommend an alternative food item that is substantially similar (e.g., the same, in the same category as the food item, etc.) as the food item that the wearer is preparing to ingest. In still other embodiments, the system may be configured to identify a particular food item, its associated nutritional information and then question the wearer's desire to ingest the particular food item. For example, the wearer may visit his local eating establishment and order “the heart attack” hamburger off the menu.
- the system may recommend to the wearer that he choose a more nutritional food item since “the heart attack” does not comply with the wearer's weight loss plan.
- the system may recommend to the wearer that he choose a more nutritional food item since “the heart attack” does not comply with the wearer's weight loss plan.
- the weight loss plan may also include allergy information associated with the wearer.
- the wearer's predetermined weight loss plan may include information relating to the wearer's allergy to soy.
- the system may send a recommendation to the wearer that states “the heart attack burger contains soy—you're allergic to soy—don't eat it!
- the system may be configured to monitor the wearer's exercise routine to determine whether the wearer is complying with daily exercise requirements. For example, the system may determine by mid-afternoon that that the wearer has only completed 50 percent of the steps required by the wearer's predetermined weight loss plan. In some such embodiments, the system may calculate additional exercise recommendations that the wearer may perform in order to meet the daily recommended steps set forth in the wearer's weight loss plan. In other embodiments, if the system detects that the wearer has been sitting for a large portion of the day, the system may calculate recommended activities that lessens the amount of time that the wearer is in the seated position.
- the system may monitor whether the wearer is exercising at the proper intensity based on at least one of the wearer's movements, the wearer's heart rate, the wearer's respiration rate or the weight loss plan. If the system detects that the wearer is not exercising at a predetermined intensity established in the weight loss plan, the system may recommend that the wearer increase the pace of the wearer's movements.
- the system may monitor the wearer's compliance with the medicine regime and calculate recommendations to help the wearer comply with the medicine regime. For example, in some embodiments, the system may calculate reminders for the wearer to assist the wearer in taking medicine/supplements. In other embodiments, the system may identify a side effect of a medicine that the wearer is taking and recommend that the wearer monitor any weight gain since a common side effect of the medicine is weight gain.
- the system may be configured to monitor the wearer's compliance with the sleep plan to determine whether the wearer is complying with the sleep plan.
- the system may use one or more sensors described herein (e.g., motion sensors) to determine by that the wearer has only slept a total of five hours for a particular day when the wearer's sleep plan called for seven hours per day.
- the system may calculate an additional nap time for the wearer in order to meet the daily recommended hours of sleep set forth in the wearer's weight loss plan.
- the system may calculate recommended activities that lessen the amount of time that the wearer is sleeping during the day.
- the system may recommend that wearer monitor any weight gain and comply with the wearer's dietary, exercise, and medicinal regimes since sleep apnea is a common side effect of gaining weight.
- the system notifies the wearer of one or more recommendations regarding the wearer's compliance with the predetermined weight loss plan.
- the system notifies the wearer and/or a third-party of the one or more recommendations calculated by the system by sending a notification to the wearer's and/or the third-party's handheld computing devices.
- the system may notify the wearer and/or the third-party of the one or more recommendations by email or text message.
- the system notifies the user of the one or more inconsistencies by communicating through a speaker coupled to the computerized eyewear.
- the system may notify the wearer and/or the third-party of a recommendation substantially immediately (e.g., immediately) after the system calculates the recommendation. In yet other embodiments, the system may notify the wearer and/or the third-party of all recommendations at the end of that day. In still other embodiments, the system may notify the wearer of an exercise recommendation when the wearer provides an indication to the system that the wearer is beginning to exercise.
- the third-party may be a relative or friend of the wearer so that the third-party may assist the wearer in meeting the requirements of the predetermined weight loss plan.
- the third-party may be the wearer's nutritionist and/or physician to allow the nutritionist and/or physician to monitor the wearer's compliance with the predetermined weight loss plan.
- the third-party may be an insurance company that may adjust the wearer's life/health insurance rate(s) based on the wearer's compliance with the predetermined weight loss plan.
- the system when executing the Behavior Pattern Analysis Module 400 , may omit particular steps, perform particular steps in an order other than the order presented above, or perform additional steps not discussed directly above.
- the wearer may put on the computerized eyewear in the morning and continue to wear the device throughout the day.
- the computerized eyewear may be operatively coupled (e.g., via a suitable wireless or wired connection) to a smart phone, a laptop, a desktop computer, or any other computing device that can receive signals from the computerized eyewear and either transmit the signals to a central system (e.g., via a wireless or wired telephone network, the Internet, etc.) or analyze the signals and make decisions based on the received signals (e.g., identify a food, a quantity of the food, calculate recommendations, etc.).
- a central system e.g., via a wireless or wired telephone network, the Internet, etc.
- the system (1) identifies a food item and a quantity of the food item that the wearer is preparing to ingest using the forward facing camera, the GPS unit, the microphone, etc., (2) tracks the food item and quantity, (3) compares the food item and quantity to a predetermined weight loss plan, (4) tracks the movements of the wearer using the motion sensor, the accelerometer, the global positioning sensor, the gyroscope, and the front-facing camera, (5) tracks the wearer's compliance with a medicine regime, as described in detail above, and/or (6) tracks the wearer's compliance with a sleep plan, as also described in detail above.
- the wearer may be under doctor's orders to lose weight.
- the device is able to capture one or more images of one or more food items that the wearer is preparing to ingest, identify the one or more food items, and compare the nutritional information for the one or more food items to a predetermined weight loss plan set by the doctor.
- the system may receive a GPS signal that can be used to determine the location of the wearer.
- the system may also determine that the wearer is at a particular food establishment (e.g., a restaurant, a food store, etc.) located at the geographical location.
- the system can identify the food item and the nutritional information associated with the food item by comparing the one or more captured images of the food item to images of food items available at the food establishment. Once the system determines a match and identifies the food item, the system may also identify nutritional information associated with the food item from a food information database accessed by the system.
- the system tracks the nutritional information associated with the food item and compares this information to the weight loss program. For example, the system may add the calories associated with the food item to a “total calories” consumed for the day value. In some embodiments, if the system determines that the food item will cause the wearer to exceed the total calories allocated to the wearer for the day under the weight loss program, the system may calculate one or more recommendations for the wearer.
- the one or more recommendations may include a statement to the wearer that says “if you eat this food item, you must perform one or more of the following exercises for 40 minutes to offset the excess calories due to the food item.”
- Another recommendation may be “Instead of eating the ‘Blue Bomber hamburger’ why not have the Chicken Caesar Salad?”
- the system may continue to remind the wearer that the wearer must perform additional physical activity to offset the excess calories consumed during the day.
Abstract
A computer-implemented method, and related system, for monitoring the wellbeing of an individual by providing eyewear that includes at least one sensor for monitoring the motion of the user. In various embodiments, the system receives data generated by the at least one sensor, uses the data to determine the user's movements using the received data, and compares the user's movements to previously established movement patterns of the user. If the system detects one or more inconsistencies between the user's current movements as compared to the previously established movement patterns of the user, the system may notify the user or a third party of the detected one or more inconsistencies. The system may similarly monitor a user's compliance with a medical regime and notify the user or a third party of the user's compliance with the regime.
Description
- This application claims the benefit of U.S. Provisional Patent Application No. 62/046,406, filed Sep. 5, 2014, and entitled, “Wearable Health Computer Apparatus, Systems, and Related Methods,” which is hereby incorporated herein by reference in its entirety.
- Public health officials concur that Americans are not adhering to current healthful lifestyle recommendations. Only one in five of the US population follows recommendations for fruit and vegetable consumption. Only one in four adheres to the recommendations for exercise, and three of four adhere to recommendations not to smoke. Possibly, the most telling statistic is that only three of every 100 US adults follow all recommendations to consume five fruits and vegetables daily, get regular physical activity, maintain a healthy weight, and not smoke.
- Being able to monitor an individual's compliance with weight loss recommendations is advantageous for the individual, for health care providers and for insurance companies. Accordingly, there is a need for improved systems and methods for monitoring an individual's compliance with a weight loss regime. Various embodiments of the present systems and methods recognize and address the foregoing considerations, and others, of prior art systems and methods.
- A computer-implemented method of monitoring compliance with a weight loss plan by a wearer of computerized eyewear, the method comprising (1) receiving, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear; (2) at least partially in response to receiving the at least one signal from the one or more sensors, determining, by at least one processor, an identity of a food that the wearer is preparing to ingest; (3) determining, by at least one processor, a quantity of the food that the wearer is preparing to ingest; (4) tracking, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest; and (5) comparing, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan. In various embodiments, determining the identity of the food further comprises capturing an image by the forward facing camera of a packaging associated with the food, at least partially in response to capturing the image, detecting a barcode contained on the packaging, and searching a database of barcodes to determine the nutritional value associated with the food. In other embodiments, determining a quantity of the food further comprises detecting in the captured image a vessel used by the wearer to measure the quantity of the food that the wearer is preparing to ingest, and detecting at least one marking on the vessel that indicates the quantity of the food placed in the vessel.
- A system for monitoring compliance with a weight loss plan by a wearer of computerized eyewear comprises one or more processors and memory operatively coupled to the one or more processors. In various embodiments, the one or more processors is configured to (1) receive at least one signal generated by one or more sensors operatively coupled to the computerized eyewear; (2) at least partially in response to receiving the at least one signal from the one or more sensors, determine an identity of a food that the wearer is preparing to ingest; (3) determine a quantity of the food that the wearer is preparing to ingest; (4) track the identity and the quantity of the food that the wearer is preparing to ingest; (5) identify the date and time that the food is ingested by the wearer; and (6) compare the identity and the quantity of the food that the wearer ingests to a predetermined dietary plan.
- A computer-implemented method of monitoring compliance with a weight loss plan by a wearer of computerized eyewear comprising the steps of (1) receiving, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear; (2) at least partially in response to receiving the at least one signal from the one or more sensors, determining, by at least one processor, an identity of a food that the wearer is preparing to ingest; (3) determining, by at least one processor, a quantity of the food that the wearer is preparing to ingest; (4) comparing, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan; (5) at least partially in response to comparing the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined dietary plan, calculating, by at least one processor, one or more recommendations to assist the wearer in complying with the predetermined dietary plan; and (6) notifying, by at least one processor, the wearer of the one or more recommendations.
- Various embodiments of systems and methods for monitoring an individual's compliance with a weight loss plan are described below. In the course of this description, reference will be made to the accompanying drawings, which are not necessarily drawn to scale and wherein:
-
FIG. 1 is a block diagram of a weight loss compliance system in accordance with an embodiment of the present system. -
FIG. 2 is a block diagram of the compliance server ofFIG. 1 . -
FIG. 3 is an exemplary wearable health monitoring device for use in the weight loss compliance system ofFIG. 1 -
FIG. 4 illustrates a flowchart that generally illustrates various steps executed by a Weight Loss Compliance Module according to a particular embodiment. - Various embodiments will now be described more fully hereinafter with reference to the accompanying drawings. It should be understood that the invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like numbers refer to like elements throughout.
- A weight loss compliance system, according to various embodiments, may include a suitable wearable device that is configured to monitor one or more of a wearer's (e.g., a user) food consumption, physical activities, sleep patterns, and/or compliance with a medicine regime (e.g., weight loss medicine/supplement). In particular embodiments, the system is configured to gather data, for example, using one or more sensors, about the wearer (e.g., such as the wearer's body temperature, heart rate, level of physical activity, geographic location, distance traveled, diet (e.g., food consumed, calories for each meal, total calories, etc.), compliance with a prescribed medical regimen (e.g., medications recently taken, dose, side effects), sleeping patterns (e.g., hours slept, type of sleep, quality of sleep), etc.). In various embodiments, the system is configured to, for example: (1) identify a food and quantity of the food that the wearer is preparing to ingest; (2) track the identity and the quantity of food (e.g., the nutritional value, the calories associated with the food, the date and time the food is ingested, etc.); (3) compare the identity and quantity of the food to a predetermined weight loss plan (e.g., a dietary plan); (4) calculate one or more recommendations to assist the wearer in complying with the predetermined weight loss plan; and/or (5) notify the wearer of the one or more recommendations.
- In particular embodiments, the wearable device is adapted to facilitate the transmission of the data captured by the one or more sensors to a remote computing device (e.g., a handheld computing device, a central server, or any other suitable smart device that may, in various embodiments, contain a wireless communications device that can connect to the wearable computing device) that analyzes the information to determine whether the wearer is complying with the predetermined weight loss plan (e.g., a dietary plan, an exercise plan, a medicine plan, a sleep plan, or a combination of one or more of the proceeding). As the system tracks the wearer's food intake (e.g., calories, nutritional values, etc.), physical activities, sleep patterns, and/or medicine intake, the system may provide the wearer feedback, warnings, and/or recommendations on (1) how to better comply with the weight plan (e.g., eat more green vegetables or fruit, don't eat fast food, you should eat smaller portions, etc.), (2) how a particular food or quantity will impact the weight loss plan (e.g., if you eat this, you must perform a particular physical activity to offset the calories, etc.), and/or (3) how a medicine can impact the wearer's ability to lose weight (e.g., a particular side effect of the medicine is weight gain/loss, etc.).
- Suitable wearable devices may include, for example: (1) a pair of eyewear (e.g., goggles or eyeglasses); (2) one or more contact lenses; (3) a wristwatch; (4) an article of clothing (e.g., such as a suitable shirt, pair of pants, undergarment, compression sleeve, etc.); (5) footwear; (6) a hat; (7) a helmet; (8) an adhesive strip or other tag that may be selectively attached to an individual or the individual's clothing; (9) a computing device that is embedded into a portion of an individual's body (e.g., under the individual's skin, or within a medical device, such as a pacemaker); (10) an orthopedic cast; or (11) any other suitable wearable item.
- In various embodiments, the wearable device comprises one or more sensors that are configured to sense the wearer's food intake, physical activities, sleep patterns, and medicine intake (e.g., prescription weight loss drugs, over-the-counter weight loss drugs, nutritional supplements, prescription drugs, etc.). The one or more sensors may include, for example: (1) one or more forward facing cameras; (2) one or more global positioning sensors; (3), one or more olfactory sensors; (4) one or more pedometers; (5) one or more microphones; (6) one or more accelerometers; (7) one or more blood pressure sensors; (8) one or more pulse oximeters; (9) one or more respiratory rate sensors; (10) one or more near-field communication chips; (11) one or more gyroscopes; (12) one or more geomagnetic sensors; (13) one or more global positioning system chips; and/or (14) any other suitable one or more sensors.
- In a particular implementation, the system's wearable device includes a pair of computerized eyewear that comprises the one or more sensors for monitoring the wearer's compliance with a weight loss plan. In particular embodiments, the one or more sensors are coupled to (e.g., connected to, embedded in, etc.) the pair of glasses, which may be, for example, a pair of computerized or non-computerized eyeglasses. In particular embodiments, the wearer is a person trying to lose weight, become more physically active, under orders of a doctor to lose weight for medical reasons or any other person trying to live a more active and healthy life.
- As will be appreciated by one skilled in the relevant field, the present systems and methods may be, for example, embodied as a computer system, a method, or a computer program product. Accordingly, various embodiments may be entirely hardware, entirely software, or a combination of hardware and software. Furthermore, particular embodiments may take the form of a computer program product stored on a computer-readable storage medium having computer-readable instructions (e.g., software) embodied in the storage medium. Various embodiments may also take the form of Internet-implemented computer software. Any suitable computer-readable storage medium may be utilized including, for example, hard disks, compact disks, DVDs, optical storage devices, and/or magnetic storage devices.
- Various embodiments are described below with reference to block diagram and flowchart illustrations of methods, apparatuses, (e.g., systems), and computer program products. It should be understood that each block of the block diagrams and flowchart illustrations, and combinations of blocks in the block diagrams and flowchart illustrations, respectively, can be implemented by a computer executing computer program instructions. These computer program instructions may be loaded onto a general purpose computer, a special purpose computer, or other programmable data processing apparatus that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the functions specified in the flowchart block or blocks.
- The computer instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on a user's computer and partly on a remote computer, or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including but not limited to: (1) a local area network (LAN); (2) a wide area network (WAN); (3) a cellular network; or (4) the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner such that the instructions stored in the computer-readable memory produce an article of manufacture that is configured for implementing the function specified in the flowchart block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process (e.g., method) such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
-
FIG. 1 is a block diagram of a weightloss compliance system 100 according to particular embodiments. As may be understood from this figure, the weightloss compliance system 100 includes one ormore networks 115, one or morethird party servers 50, acompliance server 120 that includes a weightloss compliance module 400, adietary information database 140, anexercise information database 142, amedicine database 144, asleep information database 146, one or more remote computing devices 154 (e.g., such as a smart phone, a tablet computer, a wearable computing device, a laptop computer, a desktop computer, a Bluetooth device, an automated dialing apparatus, etc.), and one or morehealth monitoring devices 156, which may, for example, be embodied as one or more of eyewear, headwear, clothing, a watch, a hat, a helmet, a cast, an adhesive bandage, a piece of jewelry (e.g., a ring, earring, necklace, bracelet, etc.), or any other suitable wearable device. In particular embodiments, the one ormore computer networks 115 facilitate communication between the one or morethird party servers 50, thecompliance server 120, thedietary information database 140, theexercise information database 142, themedicine database 144, thesleep information database 146, the one or moreremote computing devices 154, and/or the one or morehealth monitoring devices 156. - The one or
more networks 115 may include any of a variety of types of wired or wireless computer networks such as the Internet, a private intranet, a mesh network, a public switch telephone network (PSTN), or any other type of network (e.g., a network that uses Bluetooth or near field communications to facilitate communication between computing devices). The communication link between the one or moreremote computing devices 154 and thecompliance server 120 may be, for example, implemented via a Local Area Network (LAN) or via the Internet. -
FIG. 2 illustrates a diagrammatic representation of the architecture for thecompliance server 120 that may be used within the weightloss compliance system 100. It should be understood that the computer architecture shown inFIG. 2 may also represent the computer architecture for any one of the one or moreremote computing devices 154, one or morethird party servers 50, and one or morehealth monitoring devices 156, shown inFIG. 1 . In particular embodiments, thecompliance server 120 may be suitable for use as a computer within the context of the weightloss compliance system 100 that is configured for monitoring the compliance of the wearer with a weight loss plan (e.g., food intake, physical activities, sleeping habits and medicine regimes, etc.). - In particular embodiments, the
compliance server 120 may be connected (e.g., networked) to other computing devices in a LAN, an intranet, an extranet, and/or the Internet as shown inFIG. 1 . As noted above, thecompliance server 120 may operate in the capacity of a server or a client computing device in a client-server network environment, or as a peer computing device in a peer-to-peer (or distributed) network environment. Thecompliance server 120 may be a desktop personal computing device (PC), a tablet PC, a set-top box (STB), a Personal Digital Assistant (PDA), a cellular telephone, a web appliance, a network router, a switch or bridge, or any other computing device capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that computing device. Further, while only a single computing device is illustrated, the term “computing device” shall also be interpreted to include any collection of computing devices that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein. - An
exemplary compliance server 120 includes aprocessing device 202, a main memory 204 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.), a static memory 206 (e.g., flash memory, static random access memory (SRAM), etc.), and adata storage device 218, which communicate with each other via abus 232. - The
processing device 202 represents one or more general-purpose or specific processing devices such as a microprocessor, a central processing unit (CPU), or the like. More particularly, theprocessing device 202 may be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or processor implementing other instruction sets, or processors implementing a combination of instruction sets. Theprocessing device 202 may also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like. Theprocessing device 202 may be configured to executeprocessing logic 226 for performing various operations and steps discussed herein. - The
compliance server 120 may further include anetwork interface device 208. Thecompliance server 120 may also include a video display unit 210 (e.g., a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alpha-numeric input device 212 (e.g., a keyboard), a cursor control device 214 (e.g., a mouse), and a signal generation device 216 (e.g., a speaker). - The
data storage device 218 may include a non-transitory computing device-accessible storage medium 230 (also known as a non-transitory computing device-readable storage medium or a non-transitory computing device-readable medium) on which is stored one or more sets of instructions (e.g., the weight loss compliance module 400) embodying any one or more of the methodologies or functions described herein. The one or more sets of instructions may also reside, completely or at least partially, within themain memory 204 and/or within theprocessing device 202 during execution thereof by thecompliance server 120—themain memory 204 and theprocessing device 202 also constituting computing device-accessible storage media. The one or more sets of instructions may further be transmitted or received over anetwork 115 via anetwork interface device 208. - While the computing device-
accessible storage medium 230 is shown in an exemplary embodiment to be a single medium, the term “computing device-accessible storage medium” should be understood to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “computing device-accessible storage medium” should also be understood to include any medium that is capable of storing, encoding, or carrying a set of instructions for execution by the computing device and that causes the computing device to include any one or more of the methodologies of the present invention. The term “computing device-accessible storage medium” should accordingly be understood to include, but not be limited to, solid-state memories, optical and magnetic media, etc. - As shown in
FIG. 3 ,eyewear 300, according to various embodiments, includes: (1) aneyewear frame 310; (2) afirst temple 312; and (3) asecond temple 314, as described below. - Eyewear Frame
- Referring still to
FIG. 3 ,eyewear 300, in various embodiments, includes anysuitable eyewear frame 310 configured to support one ormore lenses eyewear frame 310 has afirst end 302 and asecond end 304. Theeyewear frame 310 may be made of any suitable material such as metal, ceramic, polymers or any combination thereof. In particular embodiments, theeyewear frame 310 is configured to support the first andsecond lenses second lenses eyewear frame 310 may be configured to support the first andsecond lenses eyewear frame 310 is configured to support a number of lenses other than two lenses (e.g., a single lens, a plurality of lenses, etc.). In particular embodiments, thelenses - The
eyewear frame 310 includes a first and second nose pad 322 (not shown in figures), 324, which may be configured to maintain theeyewear 300 adjacent the front of a wearer's face such that thelenses eyewear 300. In particular embodiments, thenose pads 322, 324 may comprise a material that is configured to be comfortable when worn by the wearer (e.g., rubber, etc.). In other embodiments, the nose pads may include any other suitable material (e.g., plastic, metal, etc.). In still other embodiments, the nose pads may be integrally formed with theframe 310. - The
eyewear frame 310 includes a first andsecond hinge second temples first hinge 326 may be welded to, or integrally formed with, theframe 310 and thefirst temple 312 and thesecond hinge 328 may be welded to, or integrally formed with, theframe 310 and thesecond temple 314. - First and Second Temples
- Still referring to
FIG. 4 , thefirst temple 312, according to various embodiments, is rotatably connected to theframe 310 at a right angle to extend thefirst temple 312 substantially perpendicular, substantially parallel, or anywhere in between the right angle to theframe 310. Thefirst temple 312 has a first andsecond end second end 312 b, thefirst temple 312 includes anearpiece 313 configured to be supported by a wearer's ear. Similarly, thesecond temple 314, according to various embodiments, is rotatably connected to theframe 310 at a right angle to extend thesecond temple 314 substantially perpendicular, substantially parallel, or anywhere in between the right angle to theframe 310. Thesecond temple 314 has a first andsecond end second end 314 b, thesecond temple 314 includes anearpiece 315 configured to be supported by a wearer's ear. - Sensors
- In various embodiments, the
second temple 314 has one ormore sensors 330 connected to thesecond temple 314. In various embodiments, the one ormore sensors 330 may be coupled to theframe 310, the first andsecond temples second lenses nose piece 324, or any other portion of theeyewear 300 in any suitable way. For instance, the one ormore sensors 330 may be embedded into theeyewear 300, coupled to theeyewear 300, and/or operatively coupled to theeyewear 300. In various embodiments, the one or more sensors may be formed at any point along theeyewear 300. For instance, a fingerprint reader may be disposed adjacent the first temple of theeyewear 300. In various embodiments, the one or more sensors may be formed in any shape. In addition, the one or more sensors may be formed on the inner (back) surface of theframe 310, the first andsecond temples 312, 414, the first andsecond lenses eyewear 300. In other embodiments, the one or more sensors may be formed on the outer (front) surface of theframe 310, the first andsecond temples second lenses eyewear 300. - In various embodiments, the one or
more sensors 330 that are coupled to the eyewear (or other wearable device) are adapted to detect the identity of a food, the quantity of a food, the identity of a medicine, the dose of the medicine, one or more physical activities performed by the wearer, the length of the physical activity, etc. In various embodiments, the one or more sensors coupled to the eyewear or other health monitoring device may include, for example, one or more of the following: a near-field communication chip, a gyroscope, a Bluetooth chip, a GPS unit, a forward facing camera, an olfactory sensor, a heart rate monitor, an electrocardiogram (EKG), an electroencephalogram (EEG), a pedometer, a thermometer, a microphone, an accelerometer, a magnetometer, a blood pressure sensor, a pulse oximeter, a skin conductance response sensor, a blood sensor, or any other suitable sensor. In particular embodiments, the sensors coupled to the eyewear may include one or more electronic communications devices such as a near field communication chip, a Bluetooth chip, a forward facing camera, a microphone and a GPS unit. - In various embodiments, the one or more sensors are coupled to a computing device that is associated with (e.g., embedded within, attached to) the eyewear or other wearable device. In particular embodiments, the eyewear or other wearable device comprises at least one processor, computer memory, suitable wireless communications components (e.g., a Bluetooth chip) and a power supply for powering the wearable device and/or the various sensors. In some such embodiments, the one or more sensors may be coupled to a Bluetooth device that is configured to transmit the one or more signals to a handheld wireless device.
- In particular embodiments, the system is configured to receive input from a user (e.g., a wearer of the eyewear) via one or more gestures, for example, using at least one of the sensors described above. In various embodiments, the system may, for example, be configured to: (1) identify a gesture performed by the user; and (2) at least partially in response to identifying the gesture, perform a function associated with the gesture. In particular embodiments, the system may be configured to perform a particular function in response to identifying a particular gesture, where the particular gesture is associated with the particular function. In particular embodiments, the system may be configured to enable the user to provide one or more gestures for performing a particular function. In such embodiments, the system may, for example: (1) receive a selection of a particular function from the user; (2) receive input of one or more gestures from the user; and (3) associate the particular function with the one or more gestures.
- In various embodiments, the one or more gestures may include, for example: (1) one or more hand gestures (e.g., a thumbs up, a wave, two thumbs up, holding up any particular number of fingers, making one or more fists, performing a particular movement with one or more hands, etc.); (2) one or more head movements (e.g., shaking of the user's head, a nod, etc.); (3) one or more eye movements (e.g., looking in a particular direction for a particular period of time, a wink, blinking, blinking in a particular pattern, etc.); (4) one or more facial movements (e.g., a smile, a frown, sticking out of a tongue, etc.); and/or (5) any suitable combination of these or any other suitable gestures.
- In particular embodiments, the system is configured to identify the one or more gestures, for example, using a suitable imaging device (e.g., camera) that is part of the system. In particular embodiments, the imaging device may be directed toward an area in front of the user while the user is wearing the
eyewear 300 and configured to identify gestures performed by the user's hands, arms, feet, legs, etc. In other embodiments, the system may include an imaging device directed toward the user's face and/or eyes while the user is wearing theeyewear 300 that is configured to identify gestures performed by the user's face and/or eyes. - In other embodiments, the system comprises one or more gyroscopes and/or accelerometers configured to determine a position or change in position of the
eyewear 300 while the user is wearing the eyewear. In such embodiments, the one or more gyroscopes and/or accelerometers are configured to identify one or more gestures performed by the user that include one or more gestures that include movement of the user's head. In still other embodiments, the system comprises one or more gyroscopes and/or one or more accelerometers, disposed on any other portion of the user's body, configured to identify any gesture performed by the user using the other portion of the user's body (e.g., arm, hand, leg, foot, etc.). In various embodiments, the system comprises any other suitable sensor for identifying one or more gestures performed by the user. - In particular embodiments, one or more of the sensors may be detachable from the eyewear. For instance, if a wearer does not need a temperature sensor or other particular sensor, the sensor may be removed from the eyewear. In still other embodiments, some of the one or more sensors may be coupled to the handheld wireless device, while other of the one or more sensors may be coupled to the wearable device.
- As noted above, a system, according to various embodiments, is adapted to monitor a wearer's compliance with a weight loss plan. Various aspects of the system's functionality may be executed by certain system modules, including the weight
loss compliance module 400, which is discussed in greater detail below. - Weight Loss Compliance Module
-
FIG. 4 is a flow chart of operations performed by an exemplary weightloss compliance module 400, which may, for example, run on thecompliance server 120, or any suitable computing device (such as the one or morehealth monitoring devices 156, a handheld computing device coupled to communicate with the one or morehealth monitoring devices 156 or a suitable mobile computing device). In particular embodiments, the weightloss compliance module 400 may assess a wearer's compliance with a weight loss plan (e.g., a dietary plan, an exercise plan, a sleep plan, and/or a medicine regime) and provide recommendations, feedback and/or supportive messages on how the wearer can improve their compliance with the weight loss plan. - The system begins, in various embodiments, at
Step 405 where the system receives, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear. In various embodiments, the at least one signal received may include one or more images taken by a forward facing camera coupled to thehealth monitoring device 156. In other embodiments, the at least one signal received may include one or more signals from a global positioning system unit that are associated with the location of the wearer of thehealth monitoring device 156. In still other embodiments, the at least one signal received may include one or more signals from an olfactory sensor that may be used to determine the smell associated with one or more foods being prepared for ingestion by the wearer of thehealth monitoring device 156. In yet other embodiments, the at least one signal received may include one or more audio signals received by a microphone coupled to thehealth monitoring device 156. - In various embodiments, the system receives the signals substantially automatically after the sensor generates the signal. In some embodiments, the system may receive the signal periodically (e.g., by the second, by the minute, hourly, daily, etc.). For example, the system may receive the signal every thirty seconds throughout the day. In other embodiments, the system may receive the signal after receiving an indication from the wearer or a third party that the system should receive the signal. For instance, the wearer may speak a voice command to the wearable device requesting that the device track the food being prepared by the wearer. In various embodiments, the system may receive an indication from the wearer or a third party of when to have the system receive the signal. In particular embodiments, the system may receive an indication from the wearer or a third party to have particular data received from a particular sensor at the same time that the system receives second particular data from a second particular sensor. For example, when the system receives one or more images, the system may at least partially in response to receiving the one or more images, also obtain global position data of the wearer, olfactory data associated with food being prepared by the wearer, etc.
- In particular embodiments, the system may store data associated with the received signal in memory (e.g., local memory, remote memory, etc.). In various embodiments, the system may store the data substantially automatically after receiving the signal. In other embodiments, the system may store the data after receiving manual input from the wearer or a third party requesting that the system store the data. In various embodiments, the system may store the data for a specified period of time. For instance, the system may store the data for a day, a month, a year, etc., in, for example, the
dietary information database 140, theexercise information database 142, themedicine database 144, and/or thesleep information database 146. In some embodiments, the system may store the data on any suitable server, database, or device. In other embodiments, the system may store the data on thecompliance server 120. In still other embodiments, the system may store data in an account associated with the wearer. In various embodiments, the system may store the data with a timestamp of when the data was received. - At
Step 410, the system, at least partially in response to receiving the at least one signal from the one or more sensors, determines, by at least one processor, an identity of a food that the wearer is preparing to ingest. In various embodiments, the at least one signal may be an image that is captured by a forward facing camera. In other embodiments, the at least one signal may be a signal received from a GPS chip that provides the current location of the wearer. In still other embodiments, the at least one signal may be an audio signal received from a microphone. In yet other embodiments, the signal may be a smell captured by the olfactory sensor. - In embodiments, where the at least one signal is one or more captured images, the system may capture an image of packaging associated with a food that is being prepared by the wearer. In some such embodiments, the system may analyze the image to detect a barcode located on the packaging associated with the food. Once the system detects the barcode, the system may decipher the bar code and search the
dietary information database 140 of barcodes to find a matching barcode. Once the system identifies the food associated with the barcode, the system may also obtain from thedietary information database 140 nutritional information (e.g., calories, fat content, protein content, etc.) that is associated with the matching barcode. - In other embodiments, the system may analyze the captured one or more images to identify a food located in the captured one or more images. For example, the system may capture an image of a hamburger that the wearer is preparing to eat. The system may search the
dietary information database 140 of images of food to find a matching image of a hamburger that is similar (e.g., the same, substantially similar, etc.) to the hamburger that the wearer is preparing. Once a match is determined by the system, the system can identify the food and obtain the nutritional information associated with matching image stored in thedietary information database 140. In various embodiments, the system may capture a series of images that show the wearer preparing the hamburger. Thus, in some embodiments, the system may first capture an image of the hamburger as the wearer barbecues the hamburger. From the first image, the system may identify the hamburger by finding a matching image of a hamburger in thedietary information database 140. Next, the system may capture one or more images of the user removing a hamburger roll from a package of hamburger rolls. The system may detect a barcode from the image of the packaging of the hamburger rolls and identify the roll and the nutritional information associated with the roll from the barcode associated with the one or more images by searching thedietary information database 140 for a matching barcode. The system may continue to identify various other condiments that the wearer places on the hamburger in a similar fashion. - In still other embodiments, the system may capture an image of a food that the wearer is preparing to ingest while the system simultaneously receives a GPS signal from a GPS chip contained in the computerized eyewear worn by the wearer. In some such embodiments, the system may determine the location of the wearer from the GPS signal, determine if a restaurant or similar food establishment is located at the location, and simultaneously analyze one or more captured images to identify the food that the wearer is preparing to ingest by, for example, one of the methods described above. In some of these embodiments, the wearer's location and a determination of a particular restaurant or similar food establishment may assist the system in identifying the food that the wearer is preparing to ingest. For example, if the system cannot positively identify the exact food item the wearer is preparing to ingest, the system may display a list of one or more food items associated with the restaurant or similar food establishment on a display associated with the computerized eyewear (e.g., a display coupled to the computerized eyewear, a display on a handheld computing device, etc.) and allow the wearer to select the particular food item from the displayed list of one or more food items.
- In still other embodiments, the system may receive an audio signal from one or more microphones coupled to the computerized eyewear. In some embodiments, the system may convert the received audio signal to text and use the text to search the
dietary information database 140 for matching text words to identify the food that the wearer is preparing to ingest. In some embodiments, thedietary information database 140 may also contain nutritional information associated with each word in the database. For example, the wearer may place the computerized eyewear in a state where it is ready to accept voice input. The wearer may, for example, say “hamburger”. The system may convert the received voice command to text and search thedietary information database 140 for a matching word. Each word in the database may contain associated nutritional information. Thus, if the system finds a match for the word “hamburger”, the system identifies the food item as a “hamburger” and obtains nutritional information associated with the word “hamburger.” In some embodiments, once the system converts the received audio signal into text, the system may seek confirmation of the converted word by the wearer before searching the database of food items. - As noted above, in various embodiments, the system may determine a wearer's location using a signal received from the GPS chip in the computerized eyewear or from a GPS chip on a handheld computing device coupled to the computerized eyewear. In some such embodiments, if the system detects that the user is located in a restaurant or other food establishment, the system may begin to monitor the wearer's speech in order to detect one or more food items ordered by the wearer. As the system detects the wearer's speech, the system may convert the speech to text and search the
dietary information database 140 based on one or more of: (1) the location of the wearer, (2) the associated name of the restaurant or other similar food establishment, or (3) a detected name of a food item ordered by the wearer. In various embodiments, the system may automatically detect the ordered food item. In other embodiments, once the system identifies an ordered item, the system may request confirmation that the detected item is accurate prior to making a final identification of the food item that the wearer is preparing to ingest. - In various embodiments, confirmation of the identity of the food item may be made by the wearer orally using the microphone on the computerized eyewear. In other embodiments, the system may request confirmation by displaying the identified food item on a display associated with the computerized eyewear. For example, the computerized eyewear may contain a built in display that the wearer can view while wearing the computerized eyewear. In other embodiments, the computerized eyewear may be coupled (e.g., wirelessly, wired, etc.) to a handheld computing device that displays the selected food item. In some such embodiments, the system may be configured to allow the user to select a “confirm” or “deny” button displayed on the handheld computing device.
- At
Step 415, the system determines, by at least one processor, a quantity of the food that the wearer is preparing to ingest. In various embodiments where the system detects and reads a barcode located in one or more captured images, the system may determine the serving size associated with the barcode by searching thedietary information database 140 for a matching barcode and nutritional information, which may also include a serving size for the food item. For example, in some instances where the wearer is eating prepackaged food, the matching barcode may indicate that the serving size associated with the packaged food is a single serving (e.g., a candy bar, a frozen meal, etc.). In other embodiments, the matching barcode may indicate that the serving size is a teaspoon, a tablespoon, a cup, etc. - In various embodiments where the serving size is not a single serving, the system may calculate an estimated volume from the captured one or more image. In other embodiments, the system may capture and analyze one or more images of the food and calculate an estimate of the quantity of the food that the wearer is preparing to ingest. For example, in some embodiments, the system may analyze the captured one or more images to count the number of teaspoons, tablespoons or cups of a food that the wearer selects. In other embodiments, the wearer may input the quantity of food using an input device (e.g, a food scale) coupled to the computerized eyewear or a handheld computing device coupled to the computerized eyewear or the wearer may input the quantity using an audio input via the one or more microphones coupled to the computerized eyewear.
- In other embodiments, the system may analyze the captured one or more images to detect a vessel (e.g., bowl, measuring cup, container, etc.) that the wearer is using to measure a quantity of food. In some embodiments, the system may determine the quantity of the selected food based on the identity of the vessel. For example, a set of measuring cups and/or spoons may be color coded based on the size of each cup or spoon. In these embodiments, the system may analyze the captured one or more images to detect the particular measuring cup and/or spoon being used. In other embodiments, the system may analyze the captured one or more images to detect one or more markings located on the vessel that would allow the system to determine an accurate measurement of the food based on the identity of the vessel and/or the one or more markings on the vessel. In yet other embodiments, the system may be configured to operatively couple (e.g. via Bluetooth, Wi-Fi, physical wiring, etc.) to a measuring device such as a scale, an electronic measuring cup, etc., and use a received signal from the measuring device to determine the quantity of the selected food.
- At
Step 420, the system tracks, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest. In various embodiments, the system may store the identity and quantity of a food that a wearer prepares to ingest in an account associated with the wearer on the system. In some embodiments, the system may allow the user to input when the wearer plans on eating the prepared food (e.g., for breakfast, lunch, dinner, midmorning snack, late afternoon snack, etc.). In other embodiments, the system may determine a date and time that the wearer ingests the prepared food. For example, the system may determine the date, time, and associated meal (e.g., breakfast, lunch, dinner, etc.) and stored the information in a database (e.g., dietary information database 140) or an account on the system associated with the wearer. In still other embodiments, the system may track the number of calories associated with the prepared food, the nutritional information (e.g., amount of fat, protein, carbohydrates, vitamins, minerals, etc.) associated with the prepared food, and also store this information in the database or account associated with the wearer. - At
step 425, the system compares, by at least one processor, the identity and quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan. In various embodiments, the weight loss plan may contain one or more of a dietary plan, an exercise plan, a medicine regime, or a sleep plan. In some embodiments, the weight loss plan may be prepared by the wearer. In other embodiments, the weight loss plan may be prepared by a third-party (e.g., a doctor, a trainer, a nutritionist, etc.). - In various embodiments, the dietary plan for the wearer may consist of at least one of: (A) one or more daily values for: (1) total calorie intake, (2) total fat intake, (3) total protein intake, (4) total carbohydrate intake; (B) one or more prohibited foods; (C) one or more prohibited vitamins; (D) one or more prohibited minerals; (E) a daily expected weight loss value; (F) a weekly expected weight loss value; (G) a recommended amount of daily sleep; (H) etc. In various embodiments, the system may compare the amount and quantity of food and its associated nutritional information to the dietary plan values in order to determine if the wearer is complying with the dietary plan. In some embodiments, the system may compare the actual consumed total daily calorie intake to the planned total daily calorie intake. In other embodiments, the system may, in addition to the total daily calorie intake also compare one or more of the actual daily protein intake, daily fat intake or daily carbohydrate intake to the planned daily values.
- In various embodiments, the exercise plan may consist of at least one of: (A) a weight loss goal; (B) a target daily calories to burn in order to lose a particular amount of weight per week; (C) various suggested exercises to achieve the desired weight loss goal; or (D) one or more recommended daily physical activities (e.g., walking, stretching, etc.). In some such embodiments, the system may be configured to track one or more of the wearer's (1) movements, (2) heart rate, or (3) respiration rate to determine the amount of calories burned by the wearer and the type of exercise performed by the wearer.
- In various embodiments, the system may use signals received from the one or more sensors (e.g., gyroscope, accelerometer, GPS unit, heart rate sensor, etc.) to determine the wearer's movements and calories burned. In some embodiments, the system may determine the wearer's current movements by tracking the distance traveled by the wearer for a particular day. In still other embodiments, the system may determine the wearer's current movements by tracking the orientation of the wearer using the gyroscope. In particular embodiments, the current movements of the wearer may include actions such as running, jumping, bicycling, weight lifting, sitting, standing, etc. In various embodiments, the system may be configured to determine a total calories burned by the wearer for any particular activity based on one or more of (1) the movements associated with the wearer during the activity, (2) the amount of elapsed time during which the activity is performed, (3) the wearer's heart rate (actual, average, mean, etc.) during the time the activity is performed and/or after completion of the activity, or (4) the wearer's respiration rate during and/or after the activity is performed by the wearer.
- In various embodiments, the system determines the wearer's current movements, heart rate, and/or respiration rate substantially automatically after receiving the signals from the one or more sensors. In various embodiments, the system may determine the wearer's current movements, heart rate, and/or respiration rate periodically (e.g., by the second, by the minute, hourly, daily, etc.). For example, the system may determine the wearer's current movements, heart rate, and/or respiration rate every thirty seconds throughout the day. In other embodiments, the system may determine the wearer's current movements, heart rate, and/or respiration rate after receiving an indication from the wearer that the system should analyze the signals received from the one or more sensors. For instance, the wearer may speak a voice command to the computerized eyewear requesting that the device analyze the wearer's steps taken, heart rate and/or respiration rate. In other embodiments, the wearer may indicate to the system using a handheld computing device operatively coupled to the computerized eyewear that the system should capture the wearer's current movements, heart rate, and/or respiration rate.
- In various embodiments, the weight loss plan may contain a medicine regime that is designed to support the wearer's weight loss goals. For example, the medicine regime may comprise at least one of (A) one or more prescription weight loss medicines; (B) one or more over the counter weight loss medicines; or (C) one or more over the counter supplements (e.g. Chitosan, Chromium Picolinate, etc.). In various embodiments, the medicine regime may also comprise tracking one or more prescription medicines that the wearer is already taking to determine any weight loss or weight gain side effects that may be caused by the one or prescribed medications taken by the wearer. For example, if the wearer is diabetic, the medicine regime may track the wearer's use of insulin in conjunction with the wearer's food intake, physical activity, and glucose levels to determine any weight loss or weight gain effects on the wearer.
- In some embodiments, the wearer or a third-party (e.g., the wearer's physician, etc.) may manually input a medicine regime into an account associated with the wearer on the system. For example, where a physician prescribes one or more weight loss medicines or supplements as part of the wearer's weight loss plan, the prescribed medicines or supplements may be entered into the wearer's weight loss plan. In other embodiments where the wearer is already taking prescription medications, the wearer may manually enter one or more prescription medicines into the wearer's account on the system or the wearer's account may be linked to a medical records database so that prescribed medicines are automatically linked to the wearer's weight loss plan.
- In still other embodiments, the one or more sensors on the computerized eyewear may capture one or more images of the one or more medicines/supplements that the wearer is preparing to ingest. The system may be configured to analyze the captured image to detect the presence of a label on a medicine/supplement bottle to identify the medicine/supplement and dose being ingested by the wearer. In some embodiments the system may perform optical character recognition techniques on the label to identify the type of medicine/supplement and the dose taken by the wearer. Once the medicine/supplement and/or dose is identified, the system may add the medicine/supplement to the wearer's medicine regime and perform a look-up of the medicine/supplement in a
medicine database 144 to determine if the medicine/supplement has any positive or negative side effect on weight loss. Once the system receives or identifies all medicines/supplements taken by the wearer, the system may track the wearer's compliance with the medicine regime. - In various embodiments, the system may capture one or more images throughout the day and analyze the images to determine one or more of (1) a type of medicine/supplement taken by the wearer; (2) the time the medicine/supplement is ingested by the wearer; and (3) the dose of medicine/supplement ingested by the wearer. For example, in some embodiments, the system may detect one or more pills in the captured one or more images, compare the one or more detected pills found in the captured one or more images to known images of pills stored in the
medicine database 144, identify the one or more pills by matching the one or more pills from the captured one or more images to the known images of pills stored in themedicine database 144, and detect the time that the captured one or more images was taken. Based on this information, the system is able to determine if the wearer is complying with the medicine regime prescribed by the wearer's physician in addition to monitoring the potential side effects of medicine/supplements that the wearer takes while following the weight loss plan. - In various embodiments, the weight loss plan may include a sleep plan that is designed to support the wearer's weight loss goals. In some such embodiments, the sleep plan for the wearer may consist of at least one of: (1) a recommended amount of sleep; (2) a recommended time for the wearer to go to sleep; (3) a recommended time for the wearer to wake up; and/or (4) a recommended nap time. In various embodiments, the sleep plan may include a position for the wearer to sleep in, a temperature for the room where the wearer is sleeping, and/or a lighting condition for the room where the wearer is sleeping. In particular embodiments, the sleep plan may include advice for the wearer on avoiding certain types of light. In other embodiments, the sleep plan for the wearer may include suggestions for relaxation techniques. In various embodiments, the system may track the amount and quality of sleep the wearer is obtaining to determine if the wearer is complying with the sleep plan. The system may do this, for example, using one or more of the sensors (e.g., motion sensors) described herein. In some embodiments, the system may compare the actual daily sleep of the wearer to the prescribed sleep plan. In particular embodiments, the system may, in addition to tracking the wearer's sleep patterns, track the wearer's hormone levels (e.g., through perspiration), the wearer's glucose levels, and the wearer's activity levels.
- At
step 430, at least partially in response to comparing the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan, the system calculates, using at least one processor, one or more recommendations to assist the wearer in complying with the predetermined weight loss plan. In various embodiments, the system may analyze the wearer's calorie intake, compare the current calorie intake to the predetermined weight loss plan, analyze the wearer's current calories burned and calculate one or more activities that would allow the wearer meet the daily goals for the predetermined weight loss plan. In some embodiments, for example, the system may determine that the wearer is preparing to ingest a 500 calorie hamburger. In some such embodiments, the system may recommend that the wearer run for 45 minutes to offset the additional 500 calories. - In other embodiments, the system may determine that the wearer is preparing to ingest a food item that contains 1500 mg of sodium. If the predetermined weight loss plan contains a restriction on sodium, the system may recommend that the wearer not eat the food item since the food item contains too much sodium. In still other embodiments, the system may be configured to recommend an alternative food item that is substantially similar (e.g., the same, in the same category as the food item, etc.) as the food item that the wearer is preparing to ingest. In still other embodiments, the system may be configured to identify a particular food item, its associated nutritional information and then question the wearer's desire to ingest the particular food item. For example, the wearer may visit his local eating establishment and order “the heart attack” hamburger off the menu. Once the system identifies the food item and its associated nutritional information (e.g., calories, sodium content, fat content, etc.), the system may recommend to the wearer that he choose a more nutritional food item since “the heart attack” does not comply with the wearer's weight loss plan.
- In various embodiments, the weight loss plan may also include allergy information associated with the wearer. For example, the wearer's predetermined weight loss plan may include information relating to the wearer's allergy to soy. Thus, once the system identifies that the wearer is ordering “the Heart Attack” hamburger, which based on the associated nutritional information indicates that it contains soy products, the system may send a recommendation to the wearer that states “the heart attack burger contains soy—you're allergic to soy—don't eat it!
- In other embodiments, the system may be configured to monitor the wearer's exercise routine to determine whether the wearer is complying with daily exercise requirements. For example, the system may determine by mid-afternoon that that the wearer has only completed 50 percent of the steps required by the wearer's predetermined weight loss plan. In some such embodiments, the system may calculate additional exercise recommendations that the wearer may perform in order to meet the daily recommended steps set forth in the wearer's weight loss plan. In other embodiments, if the system detects that the wearer has been sitting for a large portion of the day, the system may calculate recommended activities that lessens the amount of time that the wearer is in the seated position.
- In other embodiments, if the wearer's is engaging in a physical activity, the system may monitor whether the wearer is exercising at the proper intensity based on at least one of the wearer's movements, the wearer's heart rate, the wearer's respiration rate or the weight loss plan. If the system detects that the wearer is not exercising at a predetermined intensity established in the weight loss plan, the system may recommend that the wearer increase the pace of the wearer's movements.
- In various embodiments, the system may monitor the wearer's compliance with the medicine regime and calculate recommendations to help the wearer comply with the medicine regime. For example, in some embodiments, the system may calculate reminders for the wearer to assist the wearer in taking medicine/supplements. In other embodiments, the system may identify a side effect of a medicine that the wearer is taking and recommend that the wearer monitor any weight gain since a common side effect of the medicine is weight gain.
- In various embodiments, the system may be configured to monitor the wearer's compliance with the sleep plan to determine whether the wearer is complying with the sleep plan. For example, the system may use one or more sensors described herein (e.g., motion sensors) to determine by that the wearer has only slept a total of five hours for a particular day when the wearer's sleep plan called for seven hours per day. In some such embodiments, the system may calculate an additional nap time for the wearer in order to meet the daily recommended hours of sleep set forth in the wearer's weight loss plan. In other embodiments, if the system detects that the wearer has been sleeping for a large portion of the day, the system may calculate recommended activities that lessen the amount of time that the wearer is sleeping during the day. In still other embodiments, if the system determines that the wearer's sleep cycles are getting interrupted while the wearer is sleeping, for instance by sleep apnea, the system may recommend that wearer monitor any weight gain and comply with the wearer's dietary, exercise, and medicinal regimes since sleep apnea is a common side effect of gaining weight.
- At
Step 435, the system notifies the wearer of one or more recommendations regarding the wearer's compliance with the predetermined weight loss plan. In various embodiments, the system notifies the wearer and/or a third-party of the one or more recommendations calculated by the system by sending a notification to the wearer's and/or the third-party's handheld computing devices. In particular embodiments, the system may notify the wearer and/or the third-party of the one or more recommendations by email or text message. In yet other embodiments, the system notifies the user of the one or more inconsistencies by communicating through a speaker coupled to the computerized eyewear. - In various embodiments, the system may notify the wearer and/or the third-party of a recommendation substantially immediately (e.g., immediately) after the system calculates the recommendation. In yet other embodiments, the system may notify the wearer and/or the third-party of all recommendations at the end of that day. In still other embodiments, the system may notify the wearer of an exercise recommendation when the wearer provides an indication to the system that the wearer is beginning to exercise.
- In various embodiments, the third-party may be a relative or friend of the wearer so that the third-party may assist the wearer in meeting the requirements of the predetermined weight loss plan. In other embodiments, the third-party may be the wearer's nutritionist and/or physician to allow the nutritionist and/or physician to monitor the wearer's compliance with the predetermined weight loss plan. In still other embodiments, the third-party may be an insurance company that may adjust the wearer's life/health insurance rate(s) based on the wearer's compliance with the predetermined weight loss plan.
- In various embodiments, the system, when executing the Behavior
Pattern Analysis Module 400, may omit particular steps, perform particular steps in an order other than the order presented above, or perform additional steps not discussed directly above. - Exemplary User Experience
- Independent Living of Elderly
- In a particular example of a wearer using the health monitoring device 156 (in the form of computerized eyewear) and the
system 100 to monitor the wearer's compliance with a weight loss plan, the wearer may put on the computerized eyewear in the morning and continue to wear the device throughout the day. In various embodiments, the computerized eyewear may be operatively coupled (e.g., via a suitable wireless or wired connection) to a smart phone, a laptop, a desktop computer, or any other computing device that can receive signals from the computerized eyewear and either transmit the signals to a central system (e.g., via a wireless or wired telephone network, the Internet, etc.) or analyze the signals and make decisions based on the received signals (e.g., identify a food, a quantity of the food, calculate recommendations, etc.). During this time, the system (1) identifies a food item and a quantity of the food item that the wearer is preparing to ingest using the forward facing camera, the GPS unit, the microphone, etc., (2) tracks the food item and quantity, (3) compares the food item and quantity to a predetermined weight loss plan, (4) tracks the movements of the wearer using the motion sensor, the accelerometer, the global positioning sensor, the gyroscope, and the front-facing camera, (5) tracks the wearer's compliance with a medicine regime, as described in detail above, and/or (6) tracks the wearer's compliance with a sleep plan, as also described in detail above. - In this example, the wearer may be under doctor's orders to lose weight. Thus, by wearing the computerized eyewear throughout the day, the device is able to capture one or more images of one or more food items that the wearer is preparing to ingest, identify the one or more food items, and compare the nutritional information for the one or more food items to a predetermined weight loss plan set by the doctor. Continuing with this example, if the wearer goes out to eat, the system may receive a GPS signal that can be used to determine the location of the wearer. In addition to determining the location of the wearer, the system may also determine that the wearer is at a particular food establishment (e.g., a restaurant, a food store, etc.) located at the geographical location. Thus, based on the captured one or more images of the food item that the wearer is preparing to eat, the location of the wearer and the name of the food establishment, the system can identify the food item and the nutritional information associated with the food item by comparing the one or more captured images of the food item to images of food items available at the food establishment. Once the system determines a match and identifies the food item, the system may also identify nutritional information associated with the food item from a food information database accessed by the system.
- In various embodiments, the system tracks the nutritional information associated with the food item and compares this information to the weight loss program. For example, the system may add the calories associated with the food item to a “total calories” consumed for the day value. In some embodiments, if the system determines that the food item will cause the wearer to exceed the total calories allocated to the wearer for the day under the weight loss program, the system may calculate one or more recommendations for the wearer. For example, the one or more recommendations may include a statement to the wearer that says “if you eat this food item, you must perform one or more of the following exercises for 40 minutes to offset the excess calories due to the food item.” Another recommendation may be “Instead of eating the ‘Blue Bomber hamburger’ why not have the Chicken Caesar Salad?” In various embodiments, if the system determines that the wearer ingests the food item, the system may continue to remind the wearer that the wearer must perform additional physical activity to offset the excess calories consumed during the day.
- Many modifications and other embodiments of the invention will come to mind to one skilled in the art to which this invention pertains, having the benefit of the teaching presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the invention is not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of the appended claims. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for the purposes of limitation.
Claims (26)
1. A computer-implemented method of monitoring compliance with a weight loss plan by a wearer of computerized eyewear, the method comprising:
a. receiving, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear;
b. at least partially in response to receiving the at least one signal from the one or more sensors, determining, by at least one processor, an identity of a food that the wearer is preparing to ingest based on the received at least one signal generated by the one or more sensors;
c. determining, by at least one processor, a quantity of the food that the wearer is preparing to ingest based on the received at least one signal generated by the one or more sensors;
d. tracking, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest;
e. comparing, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan; and
f. at least partially in response to comparing the identity and the quantity of the food to the predetermined weight loss plan, at least one of:
i. updating, by at least one processor, a database to indicate whether the wearer is compliant with the weight loss plan;
ii. calculating, by at least one processor, one or more recommendations for the wearer associated with the wearer's compliance with the weight loss plan;
iii. notifying, by at least one processor, the wearer of the wearer's compliance with the weight loss plan;
iv. notifying, by at least one processor, a third-party of the wearer's compliance with the weight loss plan;
v. sending, by at least one processor, an alert to the wearer; or vi. sending, by at least one processor, an alert to the wearer regarding the calculated one or more recommendations.
2. The computer-implemented method of claim 1 , wherein the one or more sensors comprises at least one sensor selected from a group consisting of:
a. a forward facing camera;
b. a global positioning system unit;
c. an olfactory sensor; and
d. a microphone.
3. The computer-implemented method of claim 1 , wherein determining the identity of the food further comprises:
a. capturing, by at least one processor, an image by a forward facing camera of a packaging associated with the food;
b. at least partially in response to capturing the image, detecting, by at least one processor, a barcode contained on the packaging; and
c. searching, by at least one processor, a database of barcodes to determine a nutritional value associated with the food.
4. The computer-implemented method of claim 3 , wherein determining a quantity of the food further comprises, at least partially in response to capturing the image, estimating, by at least one processor, the quantity of food selected by the wearer.
5. The computer-implemented method of claim 3 , wherein determining a quantity of the food further comprises:
a. detecting, by at least one processor, in the captured image a vessel used by the wearer to measure the quantity of the food that the wearer is preparing to ingest; and
b. detecting, by at least one processor, at least one marking on the vessel that indicates the quantity of the food placed in the vessel.
6. The computer-implemented method of claim 3 , wherein determining a quantity of the food further comprises:
a. receiving, by at least one processor, a signal from a measuring device that is used by the wearer to measure the quantity of the food; and
b. at least partially in response to receiving the signal from the measuring device, determining, by at least one processor, the quantity of the food.
7. The computer-implemented method of claim 6 , wherein the measuring device is a wireless scale.
8. The computer-implemented method of claim 1 , wherein tracking the identity and the quantity of the food that the wearer is preparing to ingest further comprises:
a. determining, by at least one processor, a date and a time that the wearer ingests the food;
b. calculating, by at least one processor, a number of calories associated with the food;
c. determining, by at least one processor, a nutritional value associated with the food; and
d. storing the date and time that the wearer ingests the food, the number of calories associated with the food and the nutritional value associated with the food in memory operatively coupled to the computerized eyewear.
9. The computer-implemented method of claim 1 , wherein the predetermined weight loss plan further comprises one or more values selected from a group consisting of:
a. a total daily calorie intake value;
b. a total daily fat intake value;
c. a total daily protein intake value;
d. a total daily carbohydrate intake value;
e. one or more prohibited foods;
f. one or more prohibited nutrients;
g. a weekly weight loss value;
h. a total daily calorie burned value;
i. a total daily amount of sleep value; and
j. one or more daily physical activities
10. The computer-implemented method of claim 9 , further comprising:
a. at least partially in response to comparing the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan, calculating, by at least one processor, one or more recommendations to assist the wearer in complying with the predetermined weight loss plan; and
b. notifying, by at least one processor, the wearer of the one or more recommendations.
11. The computer-implemented method of claim 10 , wherein calculating one or more recommendations to assist the wearer in complying with the predetermined weight loss plan further comprises:
a. determining, by at least one processor, one or more physical activity recommendations;
b. accessing, by at least one processor, one or more daily caloric intake recommendations;
c. comparing, by at least one processor, the one or more physical activity recommendations to the one or more daily caloric intake recommendations;
d. at least partially in response to comparing the one or more physical activity recommendations to the one or more daily caloric intake recommendations, calculating, by at least one processor, a recommended combination of at least one physical activity recommendation and at least one daily caloric intake recommendation to assist the wearer in complying with the predetermined weight loss plan; and
e. notifying, by at least one processor, the wearer of the recommended combination of the at least one physical activity recommendation and the at least one daily caloric intake recommendation.
12. The computer-implemented method of claim 2 , wherein the at least one signal generated by one or more sensors is an audio signal received by the microphone, the method further comprising:
a. converting, by at least one processor, the audio signals to text using speech recognition techniques;
b. at least partially based on the converted text, searching, by at least one processor, a database of food items using the text; and
c. matching, by at least one processor, the converted text to at least one food item in the database of food items to identify the food item the wearer is preparing to ingest.
13. The computer-implemented method of claim 2 , further comprising:
a. receiving, by at least one processor, a signal from the global positioning system unit;
b. determining, by at least one processor, the location of the wearer at least partially based on the received signal; and
c. using, by at least one processor, the location of the wearer in determining the identity of the food that the wearer is preparing to ingest.
14. A system for monitoring compliance with a weight loss plan by a wearer of computerized eyewear, comprising:
a. one or more processors;
b. memory operatively coupled to the one or more processors;
wherein the one or more processors is configured to:
i. receive at least one signal generated by one or more sensors operatively coupled to the computerized eyewear;
ii. at least partially in response to receiving the at least one signal from the one or more sensors, determine an identity of a food that the wearer is preparing to ingest;
iii. determine a quantity of the food that the wearer is preparing to ingest;
iv. track the identity and the quantity of the food that the wearer is preparing to ingest;
v. identify the date and time that the food is ingested by the wearer;
vi. compare the identity and the quantity of the food that the wearer ingests to a predetermined weight loss plan; and
vii. at least partially in response to comparing the identity and the quantity of the food to the predetermined weight loss plan, at least one of:
(a) update a database to indicate whether the wearer is compliant with the weight loss plan;
(b) calculate one or more recommendations for the wearer associated with the wearer's compliance with the weight loss plan;
(c) notify the wearer of the wearer's compliance with the weight loss plan;
(d) notify a third-party of the wearer's compliance with the weight loss plan;
(e) send an alert to the wearer; or
(f) send an alert to the wearer regarding the calculated one or more recommendations.
15. The system of claim 14 , wherein the one or more sensors comprises at least one sensor selected from a group consisting of:
a. a forward facing camera;
b. a global positioning system unit;
c. an olfactory sensor; and
d. a microphone.
16. The system of claim 14 , wherein the one or more processors is further configured to:
a. receive a signal from a global positioning system unit;
b. determine the location of the wearer at least partially based on the received signal; and
c. use the location of the wearer to determine the identity of the food that the wearer is preparing to ingest.
17. The system of claim 14 , wherein the one or more processors is further configured to:
a. capture one or more images by a forward-facing camera of medicine ingested by the wearer;
b. detect one or more pills in the one or more images;
c. compare the one or more detected pills found in the one or more images to one or more known images of pills stored in memory;
d. at least partially in response to comparing the one or more detected pills to one or more known images of pills stored in memory, determine data selected from a group consisting of one or more:
i. types of medicine taken by the user;
ii. times the medicine is taken by the user; and
iii. doses of medicine taken by the user;
e. determine, at least partially based on the data, if the medicine ingested by the wearer causes a side effect that can affect the wearer's ability to lose weight; and
f. notify either the wearer or a third party of the side effect.
18. The system of claim 17 , wherein the one or more processors is further configured to:
a. detect, by a processor, a level of one or more medicines in the wearer's bloodstream based on a signal received from a blood testing sensor operatively coupled to the system;
b. compare the level of the one or medicines in the user's bloodstream to a predefined level for each of the one or more medicines stored in a database for the wearer; and
c. at least partially in response to comparing the level of the one or medicines in the wearer's bloodstream, notify one of the wearer or the third party when the level of the one or more medicines is below the predefined level for each of the one or more medicines.
19. The system of claim 18 , wherein the one or more medicines is selected from a group consisting of:
a. a prescription drug;
b. a nutritional supplement;
c. a prescription weight loss drug; and
d. an over-the-counter weight loss drug.
20. A computer-implemented method of monitoring compliance with a weight loss plan by a wearer of computerized eyewear, the method comprising:
a. receiving, by at least one processor, at least one signal generated by one or more sensors operatively coupled to the computerized eyewear;
b. at least partially in response to receiving the at least one signal from the one or more sensors, determining, by at least one processor, an identity of a food that the wearer is preparing to ingest;
c. determining, by at least one processor, a quantity of the food that the wearer is preparing to ingest;
d. comparing, by at least one processor, the identity and the quantity of the food that the wearer is preparing to ingest to a predetermined weight loss plan;
e. at least partially in response to comparing the identity and the quantity of the food that the wearer is preparing to ingest to the predetermined weight loss plan, calculating, by at least one processor, one or more recommendations to assist the wearer in complying with the predetermined weight loss plan; and
f. notifying, by at least one processor, the wearer of the one or more recommendations.
21. The computer-implemented method of claim 20 , wherein the one or more sensors comprises at least one sensor selected from a group consisting of:
a. a forward facing camera;
b. a global positioning system unit;
c. an olfactory sensor; and
d. a microphone.
22. The computer-implemented method of claim 21 , wherein determining the identity of the food further comprises:
a. capturing, by at least one processor, an image by the forward facing camera of a packaging associated with the food;
b. at least partially in response to capturing the image, detecting, by at least one processor, a barcode contained on the packaging; and
c. searching, by at least one processor, a database of barcodes to determine the nutritional value associated with the food.
23. The computer-implemented method of claim 22 , wherein determining a quantity of the food further comprises:
a. detecting, by at least one processor, in the captured image a vessel used by the wearer to measure the quantity of the food that the wearer is preparing to ingest; and
b. detecting, by at least one processor, at least one marking on the vessel that indicates the quantity of the food placed in the vessel.
24. The computer-implemented method of claim 20 , wherein the predetermined weight loss plan further comprises one or more values selected from a group consisting of:
a. a total daily calorie intake value;
b. a total daily fat intake value;
c. a total daily protein intake value;
d. a total daily carbohydrate intake value;
e. one or more prohibited foods;
f. one or more prohibited nutrients;
g. a weekly weight loss value;
h. a total daily amount of sleep value;
i. a total daily calorie burned value; and
j. one or more daily physical activities.
25. The computer-implemented method of claim 24 , wherein calculating one or more recommendations to assist the wearer in complying with the predetermined weight loss plan further comprises:
a. determining, by at least one processor, one or more physical activity recommendations;
b. determining, by at least one processor, one or more daily caloric intake recommendations;
c. at least partially in response to determining the one or more physical activity recommendations and the one or more daily caloric intake recommendations, calculating, by at least one processor, a recommended combination of at least one physical activity recommendation and at least one daily caloric intake recommendation; and
d. notifying, by at least one processor, the wearer of the recommended combination of the at least one physical activity recommendation and the at least one daily caloric intake recommendation to assist the wearer in complying with the predetermined weight loss plan.
26. The computer-implemented method of claim 24 , wherein calculating one or more recommendations to assist the wearer in complying with the predetermined weight loss plan further comprises:
a. determining, by at least one processor, one or more total daily amounts of sleep recommendations;
b. determining, by at least one processor, one or more daily caloric intake recommendations;
c. comparing, by at least one processor, the one or more total daily amounts of sleep recommendations to the one or more daily caloric intake recommendations;
d. at least partially in response to comparing the one or more total daily amounts of sleep recommendations to the one or more daily caloric intake recommendations, calculating, by at least one processor, a recommended combination of at least one total daily amount of sleep recommendation and at least one daily caloric intake recommendation; and
e. notifying, by at least one processor, the wearer of the recommended combination of the at least one total daily amount of sleep recommendation and the at least one daily caloric intake recommendation to assist the wearer in complying with the predetermined weight loss plan.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/610,589 US20160071423A1 (en) | 2014-09-05 | 2015-01-30 | Systems and method for monitoring an individual's compliance with a weight loss plan |
EP15838411.5A EP3148435B1 (en) | 2014-09-05 | 2015-09-04 | System for monitoring health related information for individuals |
CA2953856A CA2953856A1 (en) | 2014-09-05 | 2015-09-04 | System for monitoring health related information for individuals |
PCT/US2015/048612 WO2016037091A1 (en) | 2014-09-05 | 2015-09-04 | System for monitoring health related information for individuals |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462046406P | 2014-09-05 | 2014-09-05 | |
US14/610,589 US20160071423A1 (en) | 2014-09-05 | 2015-01-30 | Systems and method for monitoring an individual's compliance with a weight loss plan |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160071423A1 true US20160071423A1 (en) | 2016-03-10 |
Family
ID=55436366
Family Applications (12)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/550,406 Active 2037-11-20 US10448867B2 (en) | 2014-09-05 | 2014-11-21 | Wearable gait monitoring apparatus, systems, and related methods |
US14/562,454 Active 2034-12-15 US9795324B2 (en) | 2014-09-05 | 2014-12-05 | System for monitoring individuals as they age in place |
US14/578,039 Abandoned US20160066829A1 (en) | 2014-09-05 | 2014-12-19 | Wearable mental state monitor computer apparatus, systems, and related methods |
US14/588,122 Active 2035-02-18 US9649052B2 (en) | 2014-09-05 | 2014-12-31 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US14/610,439 Active 2037-03-18 US10307085B2 (en) | 2014-09-05 | 2015-01-30 | Wearable physiology monitor computer apparatus, systems, and related methods |
US14/610,589 Abandoned US20160071423A1 (en) | 2014-09-05 | 2015-01-30 | Systems and method for monitoring an individual's compliance with a weight loss plan |
US14/610,628 Abandoned US20160070122A1 (en) | 2014-09-05 | 2015-01-30 | Computerized replacement temple for standard eyewear |
US15/594,898 Active 2035-02-02 US10188323B2 (en) | 2014-09-05 | 2017-05-15 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US15/791,196 Abandoned US20180042523A1 (en) | 2014-09-05 | 2017-10-23 | System for monitoring individuals as they age in place |
US16/259,646 Active US10542915B2 (en) | 2014-09-05 | 2019-01-28 | Systems, apparatus, and methods for using a wearable device to confirm the identity of an individual |
US16/429,480 Active US10694981B2 (en) | 2014-09-05 | 2019-06-03 | Wearable physiology monitor computer apparatus, systems, and related methods |
US16/657,982 Abandoned US20200046260A1 (en) | 2014-09-05 | 2019-10-18 | Wearable gait monitoring apparatus, systems, and related methods |
Family Applications Before (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/550,406 Active 2037-11-20 US10448867B2 (en) | 2014-09-05 | 2014-11-21 | Wearable gait monitoring apparatus, systems, and related methods |
US14/562,454 Active 2034-12-15 US9795324B2 (en) | 2014-09-05 | 2014-12-05 | System for monitoring individuals as they age in place |
US14/578,039 Abandoned US20160066829A1 (en) | 2014-09-05 | 2014-12-19 | Wearable mental state monitor computer apparatus, systems, and related methods |
US14/588,122 Active 2035-02-18 US9649052B2 (en) | 2014-09-05 | 2014-12-31 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US14/610,439 Active 2037-03-18 US10307085B2 (en) | 2014-09-05 | 2015-01-30 | Wearable physiology monitor computer apparatus, systems, and related methods |
Family Applications After (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/610,628 Abandoned US20160070122A1 (en) | 2014-09-05 | 2015-01-30 | Computerized replacement temple for standard eyewear |
US15/594,898 Active 2035-02-02 US10188323B2 (en) | 2014-09-05 | 2017-05-15 | Systems, apparatus, and methods for using eyewear, or other wearable item, to confirm the identity of an individual |
US15/791,196 Abandoned US20180042523A1 (en) | 2014-09-05 | 2017-10-23 | System for monitoring individuals as they age in place |
US16/259,646 Active US10542915B2 (en) | 2014-09-05 | 2019-01-28 | Systems, apparatus, and methods for using a wearable device to confirm the identity of an individual |
US16/429,480 Active US10694981B2 (en) | 2014-09-05 | 2019-06-03 | Wearable physiology monitor computer apparatus, systems, and related methods |
US16/657,982 Abandoned US20200046260A1 (en) | 2014-09-05 | 2019-10-18 | Wearable gait monitoring apparatus, systems, and related methods |
Country Status (4)
Country | Link |
---|---|
US (12) | US10448867B2 (en) |
EP (3) | EP3189371A4 (en) |
CA (3) | CA2960425A1 (en) |
WO (3) | WO2016037091A1 (en) |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180374386A1 (en) * | 2017-06-27 | 2018-12-27 | International Business Machines Corporation | Physical activity and dietary based services |
US20190080629A1 (en) * | 2017-09-13 | 2019-03-14 | At&T Intellectual Property I, L.P. | Monitoring food intake |
CN109543546A (en) * | 2018-10-26 | 2019-03-29 | 复旦大学 | The gait age estimation method returned based on the distribution of depth sequence |
US10582897B2 (en) | 2018-07-24 | 2020-03-10 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US20200086133A1 (en) * | 2018-09-18 | 2020-03-19 | Biointellisense, Inc. | Validation, compliance, and/or intervention with ear device |
WO2020053307A1 (en) * | 2018-09-12 | 2020-03-19 | Dnanudge | Product recommendation system and method |
US10650919B2 (en) | 2015-10-01 | 2020-05-12 | Dnanudge Limited | Wearable device |
US10699247B2 (en) | 2017-05-16 | 2020-06-30 | Under Armour, Inc. | Systems and methods for providing health task notifications |
US10699806B1 (en) | 2019-04-15 | 2020-06-30 | Dnanudge Limited | Monitoring system, wearable monitoring device and method |
US10811140B2 (en) | 2019-03-19 | 2020-10-20 | Dnanudge Limited | Secure set-up of genetic related user account |
US20200372826A1 (en) * | 2019-05-10 | 2020-11-26 | Tso-Cheng Chien | Food quantum tracking tools and methods related thereto |
US10861594B2 (en) | 2015-10-01 | 2020-12-08 | Dnanudge Limited | Product recommendation system and method |
US10922397B2 (en) | 2018-07-24 | 2021-02-16 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US20210307642A1 (en) * | 2015-07-27 | 2021-10-07 | Massachusetts Institute Of Technology | Systems, apparatus, and methods related to modeling, monitoring, and/or managing metabolism |
US11183291B2 (en) | 2018-02-12 | 2021-11-23 | Zoe Limited | Generating personalized nutritional recommendations using predicted values of biomarkers |
US11183080B2 (en) | 2018-02-12 | 2021-11-23 | Zoe Limited | Generating predicted values of biomarkers for scoring food |
US11295860B2 (en) | 2018-02-12 | 2022-04-05 | Zoe Limited | Using at home measures to predict clinical state and improving the accuracy of at home measurements/predictions data associated with circadian rhythm and meal timing |
US11348479B2 (en) | 2018-05-23 | 2022-05-31 | Zoe Limited | Accuracy of measuring nutritional responses in a non-clinical setting |
US11380215B2 (en) * | 2018-08-30 | 2022-07-05 | Kyndryl, Inc. | Reward-based ecosystem for tracking nutritional consumption |
US11403382B2 (en) | 2019-07-24 | 2022-08-02 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US11589807B2 (en) * | 2017-05-11 | 2023-02-28 | The Regents Of The University Of California | Biosensor for monitoring eyedrop usage compliance |
US11915151B2 (en) | 2018-08-27 | 2024-02-27 | Zoe Limited | Accuracy of test data outside the clinic |
Families Citing this family (107)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9900669B2 (en) | 2004-11-02 | 2018-02-20 | Pierre Touma | Wireless motion sensor system and method |
US10831982B2 (en) | 2009-10-14 | 2020-11-10 | Iplcontent, Llc | Hands-free presenting device |
US9968297B2 (en) * | 2012-06-14 | 2018-05-15 | Medibotics Llc | EEG glasses (electroencephalographic eyewear) |
US11612786B2 (en) * | 2012-08-31 | 2023-03-28 | Blue Goji Llc | System and method for targeted neurological therapy using brainwave entrainment with passive treatment |
US20140142442A1 (en) * | 2012-11-19 | 2014-05-22 | Judy Sibille SNOW | Audio Feedback for Medical Conditions |
US9400549B2 (en) | 2013-03-08 | 2016-07-26 | Chi Fai Ho | Method and system for a new-era electronic book |
WO2014179507A1 (en) * | 2013-04-30 | 2014-11-06 | White Chester | Body impact bracing apparatus |
US10448867B2 (en) | 2014-09-05 | 2019-10-22 | Vision Service Plan | Wearable gait monitoring apparatus, systems, and related methods |
US10617342B2 (en) | 2014-09-05 | 2020-04-14 | Vision Service Plan | Systems, apparatus, and methods for using a wearable device to monitor operator alertness |
US11918375B2 (en) | 2014-09-05 | 2024-03-05 | Beijing Zitiao Network Technology Co., Ltd. | Wearable environmental pollution monitor computer apparatus, systems, and related methods |
FR3028980B1 (en) * | 2014-11-20 | 2017-01-13 | Oberthur Technologies | METHOD AND DEVICE FOR AUTHENTICATING A USER |
US10215568B2 (en) | 2015-01-30 | 2019-02-26 | Vision Service Plan | Systems and methods for tracking motion, performance, and other data for an individual such as a winter sports athlete |
KR20160101497A (en) * | 2015-02-17 | 2016-08-25 | 삼성전자주식회사 | Wearable device and method for operating thereof |
US9886633B2 (en) * | 2015-02-23 | 2018-02-06 | Vivint, Inc. | Techniques for identifying and indexing distinguishing features in a video feed |
NZ773833A (en) | 2015-03-16 | 2022-07-01 | Magic Leap Inc | Methods and systems for diagnosing and treating health ailments |
CN107408151B (en) * | 2015-03-27 | 2022-02-18 | 英特尔公司 | Motion tracking using electronic devices |
US9930102B1 (en) | 2015-03-27 | 2018-03-27 | Intuit Inc. | Method and system for using emotional state data to tailor the user experience of an interactive software system |
US10169827B1 (en) | 2015-03-27 | 2019-01-01 | Intuit Inc. | Method and system for adapting a user experience provided through an interactive software system to the content being delivered and the predicted emotional impact on the user of that content |
US10387173B1 (en) | 2015-03-27 | 2019-08-20 | Intuit Inc. | Method and system for using emotional state data to tailor the user experience of an interactive software system |
US10275902B2 (en) | 2015-05-11 | 2019-04-30 | Magic Leap, Inc. | Devices, methods and systems for biometric user recognition utilizing neural networks |
US10154129B2 (en) * | 2015-05-15 | 2018-12-11 | Polar Electro Oy | Wearable electronic apparatus |
US20160350138A1 (en) * | 2015-05-31 | 2016-12-01 | Roya Caroline SALAS | Biofeedback system |
US11103139B2 (en) | 2015-06-14 | 2021-08-31 | Facense Ltd. | Detecting fever from video images and a baseline |
US11103140B2 (en) | 2015-06-14 | 2021-08-31 | Facense Ltd. | Monitoring blood sugar level with a comfortable head-mounted device |
US11064892B2 (en) | 2015-06-14 | 2021-07-20 | Facense Ltd. | Detecting a transient ischemic attack using photoplethysmogram signals |
US10799122B2 (en) | 2015-06-14 | 2020-10-13 | Facense Ltd. | Utilizing correlations between PPG signals and iPPG signals to improve detection of physiological responses |
US11154203B2 (en) | 2015-06-14 | 2021-10-26 | Facense Ltd. | Detecting fever from images and temperatures |
US10791938B2 (en) | 2015-06-14 | 2020-10-06 | Facense Ltd. | Smartglasses for detecting congestive heart failure |
US20170225033A1 (en) * | 2015-06-23 | 2017-08-10 | Ipcomm Llc | Method and Apparatus for Analysis of Gait and to Provide Haptic and Visual Corrective Feedback |
US10332122B1 (en) | 2015-07-27 | 2019-06-25 | Intuit Inc. | Obtaining and analyzing user physiological data to determine whether a user would benefit from user support |
US20170061823A1 (en) * | 2015-09-02 | 2017-03-02 | Hello Doctor Ltd. | System for tracking and monitoring personal medical data and encouraging to follow personalized condition-based profile and method thereof |
US11272864B2 (en) | 2015-09-14 | 2022-03-15 | Health Care Originals, Inc. | Respiratory disease monitoring wearable apparatus |
WO2017065061A1 (en) * | 2015-10-13 | 2017-04-20 | アルプス電気株式会社 | Walking measurement device, walking measurement method, and program |
US10014967B2 (en) * | 2015-11-23 | 2018-07-03 | Huami Inc. | System and method for authenticating a broadcast device using facial recognition |
US10096383B2 (en) * | 2015-11-24 | 2018-10-09 | International Business Machines Corporation | Performing a health analysis using a smart floor mat |
US10105095B2 (en) * | 2015-11-30 | 2018-10-23 | Oura Health Oy | Method and system for defining balance between physical activity and rest |
EP3178379A1 (en) * | 2015-12-09 | 2017-06-14 | Rythm | Method and device for bioelectric physiological signal acquisition and processing |
US10610146B1 (en) * | 2015-12-21 | 2020-04-07 | Dp Technologies, Inc. | Utilizing wearable devices in an internet of things environment |
US11094418B2 (en) * | 2015-12-31 | 2021-08-17 | Nokia Technologies Oy | Optimized biological measurement |
WO2017163227A1 (en) * | 2016-03-25 | 2017-09-28 | Randolph Andrae | User authentication using biometric information |
US10251597B2 (en) * | 2016-04-21 | 2019-04-09 | Viavi Solutions Inc. | Health tracking device |
JP6992003B2 (en) * | 2016-04-22 | 2022-02-03 | ビオプティックス・インコーポレイテッド | Determining Absolute Tissue Oxygen Saturation and Relative Tissue Oxygen Saturation |
US10955269B2 (en) | 2016-05-20 | 2021-03-23 | Health Care Originals, Inc. | Wearable apparatus |
US20170345274A1 (en) * | 2016-05-27 | 2017-11-30 | General Scientific Corporation | Neck posture recording and warning device |
US10002515B2 (en) * | 2016-06-01 | 2018-06-19 | Tile, Inc. | User intervention based on tracking device location |
CN109310913B (en) * | 2016-08-09 | 2021-07-06 | 株式会社比弗雷斯 | Three-dimensional simulation method and device |
WO2018027253A1 (en) * | 2016-08-11 | 2018-02-15 | Bloomfield Lochlan John | Health management system and method |
JP6854900B2 (en) | 2016-09-14 | 2021-04-14 | エフ ホフマン−ラ ロッシュ アクチェン ゲゼルシャフト | Digital biomarkers for progressive MS |
CN106510719B (en) * | 2016-09-30 | 2023-11-28 | 歌尔股份有限公司 | User gesture monitoring method and wearable device |
US10539549B2 (en) * | 2016-10-13 | 2020-01-21 | Worcester Polytechnic Institute | Mobile blood alcohol content and impairment sensing device |
EP3529726A4 (en) * | 2016-10-20 | 2020-06-10 | Datapult, Inc. | Sickness prediction application system |
CN106377271A (en) * | 2016-10-20 | 2017-02-08 | 中国矿业大学 | Wearable method and wearable device for monitoring and regulating physiology and psychology |
WO2018085271A1 (en) | 2016-11-01 | 2018-05-11 | Polyvagal Science LLC | Methods and systems for reducing sound sensitivities and improving auditory processing, behavioral state regulation and social engagement |
WO2018097204A1 (en) * | 2016-11-22 | 2018-05-31 | 国立研究開発法人宇宙航空研究開発機構 | System, method, and program for predicting diminished attentiveness state, and storage medium on which program is stored |
GB201620638D0 (en) | 2016-12-05 | 2017-01-18 | Equi+Poise Ltd | A gait analysis system |
KR20180089803A (en) | 2017-02-01 | 2018-08-09 | 삼성전자주식회사 | Electronic apparatus and method for processing authentication |
US10973446B2 (en) * | 2017-02-13 | 2021-04-13 | David Schie | Device to extract physiological information and method therefor |
US11622716B2 (en) * | 2017-02-13 | 2023-04-11 | Health Care Originals, Inc. | Wearable physiological monitoring systems and methods |
US20180228405A1 (en) | 2017-02-13 | 2018-08-16 | Starkey Laboratories, Inc. | Fall prediction system including an accessory and method of using same |
JP6894252B2 (en) * | 2017-02-16 | 2021-06-30 | 日本光電工業株式会社 | Sensor device and watching device |
US20180241973A1 (en) * | 2017-02-21 | 2018-08-23 | Janet Newell | Video and audio recording system and method |
SE541712C2 (en) * | 2017-02-22 | 2019-12-03 | Next Step Dynamics Ab | Method and apparatus for health prediction |
US20210141453A1 (en) * | 2017-02-23 | 2021-05-13 | Charles Robert Miller, III | Wearable user mental and contextual sensing device and system |
WO2018182159A1 (en) * | 2017-03-28 | 2018-10-04 | 문명일 | Smart glasses capable of processing virtual object |
KR102651253B1 (en) * | 2017-03-31 | 2024-03-27 | 삼성전자주식회사 | An electronic device for determining user's emotions and a control method thereof |
CN107049338A (en) * | 2017-04-12 | 2017-08-18 | 河南工业大学 | A kind of medical use mood detection means communicated based on computer |
US11559252B2 (en) | 2017-05-08 | 2023-01-24 | Starkey Laboratories, Inc. | Hearing assistance device incorporating virtual audio interface for therapy guidance |
US10534203B2 (en) * | 2017-07-31 | 2020-01-14 | Snap Inc. | Near-field antenna for eyewear |
US11373450B2 (en) * | 2017-08-11 | 2022-06-28 | Tectus Corporation | Eye-mounted authentication system |
US11062572B1 (en) * | 2017-09-20 | 2021-07-13 | Amazon Technologies, Inc. | Visual indicator for head-mounted device |
US10073998B1 (en) * | 2017-11-29 | 2018-09-11 | The United States Of America As Representd By The Secretary Of The Navy | Multifunction wearable object identified glasses for the visually handicapped |
US20190167226A1 (en) * | 2017-12-04 | 2019-06-06 | International Business Machines Corporation | Infant gastrointestinal monitor |
US11092998B1 (en) | 2018-01-12 | 2021-08-17 | Snap Inc. | Eyewear device with fingerprint sensor for user input |
CN108089326B (en) * | 2018-02-01 | 2023-12-26 | 北京七鑫易维信息技术有限公司 | Device suitable for being used with glasses |
WO2019157450A1 (en) * | 2018-02-12 | 2019-08-15 | Cornell University | Methods and systems for concussion management using cold stimulus |
WO2019169220A1 (en) * | 2018-03-01 | 2019-09-06 | Polyvagal Science LLC | Systems and methods for modulating physiological state |
CN108401129A (en) * | 2018-03-22 | 2018-08-14 | 广东小天才科技有限公司 | Video call method, device, terminal based on Wearable and storage medium |
US11331003B2 (en) | 2018-03-27 | 2022-05-17 | Samsung Electronics Co., Ltd. | Context-aware respiration rate determination using an electronic device |
JP7019796B2 (en) * | 2018-03-30 | 2022-02-15 | 株式会社日立製作所 | Physical function independence support device and its method |
US10424035B1 (en) * | 2018-05-16 | 2019-09-24 | Trungram Gyaltrul R. Sherpa | Monitoring conditions associated with remote individuals over a data communication network and automatically notifying responsive to detecting customized emergency conditions |
US10438479B1 (en) * | 2018-05-21 | 2019-10-08 | International Business Machines Corporation | Safety enhancement for persons under the care of others |
US11054638B2 (en) * | 2018-06-13 | 2021-07-06 | Reavire, Inc. | Tracking pointing direction of device |
US10121355B1 (en) * | 2018-06-26 | 2018-11-06 | Capital One Services, Llc | Condition-responsive wearable device for sensing and indicating proximity of an article with a specific characteristic |
US10722128B2 (en) | 2018-08-01 | 2020-07-28 | Vision Service Plan | Heart rate detection system and method |
EP3895141B1 (en) | 2018-12-15 | 2024-01-24 | Starkey Laboratories, Inc. | Hearing assistance system with enhanced fall detection features |
US11638563B2 (en) | 2018-12-27 | 2023-05-02 | Starkey Laboratories, Inc. | Predictive fall event management system and method of using same |
WO2020144175A1 (en) * | 2019-01-07 | 2020-07-16 | Metralabs Gmbh Neue Technologien Und Systeme | Method and system for capturing the sequence of movement of a person |
US11156856B2 (en) | 2019-05-02 | 2021-10-26 | Tom Nepola | Eyewear with wearing status detector |
US11007406B2 (en) * | 2019-05-03 | 2021-05-18 | Xperience Robotics, Inc. | Wearable device systems and methods for guiding physical movements |
ES1231504Y (en) * | 2019-05-14 | 2019-09-13 | I4Life Innovacion Y Desarrollos S L | Multifunctional Unlocking Device |
US11786694B2 (en) | 2019-05-24 | 2023-10-17 | NeuroLight, Inc. | Device, method, and app for facilitating sleep |
US11223939B1 (en) * | 2019-07-31 | 2022-01-11 | United Services Automobile Association (Usaa) | Environmental conditions monitoring system |
JPWO2021084903A1 (en) * | 2019-11-01 | 2021-05-06 | ||
US11310322B2 (en) | 2019-11-21 | 2022-04-19 | Blackberry Limited | Method and system for pairing a chassis and container in an asset tracking system |
DE102020102281A1 (en) * | 2019-12-19 | 2021-06-24 | USound GmbH | Glasses with charging interface |
US11948672B2 (en) * | 2020-02-27 | 2024-04-02 | Todd Martin | Mobile intelligent injury minimization system and method |
US20220020481A1 (en) | 2020-07-20 | 2022-01-20 | Abbott Laboratories | Digital pass verification systems and methods |
CN112150767B (en) * | 2020-09-27 | 2021-11-23 | 河南城建学院 | Fatigue driving monitoring system based on Internet of things and computer |
KR102525485B1 (en) * | 2020-12-02 | 2023-04-25 | 임지호 | Self-authentication wearable device for the visually impaired |
CN112562260B (en) * | 2020-12-16 | 2022-08-09 | 浙江大华技术股份有限公司 | Anti-lost method and device |
DE202021100976U1 (en) | 2021-02-26 | 2021-05-18 | seiwo Technik GmbH | System for recording the state of health of a person |
DE102021104705A1 (en) | 2021-02-26 | 2022-09-01 | seiwo Technik GmbH | Method and system for recording the state of health of a person |
WO2023279082A1 (en) * | 2021-07-02 | 2023-01-05 | Research Triangle Institute | Systems, methods, and devices for detecting viral respiratory illness in presymptomatic and asymptomatic infected persons |
CN113457108B (en) * | 2021-07-07 | 2022-07-15 | 首都体育学院 | Cognitive characterization-based exercise performance improving method and device |
US20230028690A1 (en) * | 2021-07-23 | 2023-01-26 | Consumer Safety Technology Llc | Method and system of deploying a wearable transdermal based vehicle ignition interlock |
WO2023021509A1 (en) * | 2021-08-17 | 2023-02-23 | Brainwatch Tech Ltd. | Methods, systems, and devices for determining a status of brain and/or nerve functions of a patient |
US11806078B1 (en) | 2022-05-01 | 2023-11-07 | Globe Biomedical, Inc. | Tear meniscus detection and evaluation system |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100042430A1 (en) * | 2008-08-12 | 2010-02-18 | Irody Inc | System and method for collecting and authenticating medication consumption |
US20100136508A1 (en) * | 2008-10-23 | 2010-06-03 | Damir Zekhtser | Meal Plan Management |
US20100332571A1 (en) * | 2009-06-30 | 2010-12-30 | Jennifer Healey | Device augmented food identification |
US20110301446A1 (en) * | 2007-01-15 | 2011-12-08 | Deka Products Limited Partnership | Device and Method for Food Management |
US20120135384A1 (en) * | 2010-11-26 | 2012-05-31 | Terumo Kabushiki Kaisha | Portable terminal, calorie estimation method, and calorie estimation program |
US20130157232A1 (en) * | 2011-12-09 | 2013-06-20 | Joel Ehrenkranz | System and methods for monitoring food consumption |
Family Cites Families (247)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3505879A (en) | 1968-07-31 | 1970-04-14 | Ford Motor Co | Electrical generator including components of an automotive vehicle mechanical speedometer |
US3548663A (en) | 1968-07-31 | 1970-12-22 | Ford Motor Co | Electrical generator including components of an automotive vehicle mechanical speedometer |
US3972038A (en) | 1975-03-28 | 1976-07-27 | Nasa | Accelerometer telemetry system |
US4100401A (en) | 1977-01-13 | 1978-07-11 | Tutt Eugene F | Calorie calculator-chronometer |
DE2709412A1 (en) | 1977-03-04 | 1978-09-07 | Max Baermann | Eddy current tachometer with temperature compensation |
US4195642A (en) | 1978-01-03 | 1980-04-01 | Beehive International | Wearable heart rate monitor |
GB1593839A (en) | 1978-05-26 | 1981-07-22 | Pringle R D | Performance testing device |
US4434801A (en) | 1980-04-30 | 1984-03-06 | Biotechnology, Inc. | Apparatus for testing physical condition of a self-propelled vehicle rider |
US4407295A (en) | 1980-10-16 | 1983-10-04 | Dna Medical, Inc. | Miniature physiological monitor with interchangeable sensors |
US4855942A (en) | 1987-10-28 | 1989-08-08 | Elexis Corporation | Pedometer and/or calorie measuring device and method |
US4878749A (en) | 1988-06-29 | 1989-11-07 | Mcgee James E | Protective eyewear with interchangeable decorative frames |
US4919530A (en) | 1989-01-25 | 1990-04-24 | Hyman Roger L | Eyeglass assembly |
US5670872A (en) | 1992-06-22 | 1997-09-23 | U.S. Philips Corporation | System and device with vertical and rotary wheel-velocity-measuring for determining vehicle displacement |
US5497143A (en) | 1993-06-24 | 1996-03-05 | Casio Computer Co., Ltd. | Electronic device for a vehicle |
US5422816A (en) | 1994-02-22 | 1995-06-06 | Trimble Navigation Limited | Portable personal navigation tracking system |
US5452480A (en) | 1994-04-15 | 1995-09-26 | Electric Eyewear Systems, Inc. | Ski goggles |
US7386401B2 (en) | 1994-11-21 | 2008-06-10 | Phatrat Technology, Llc | Helmet that reports impact information, and associated methods |
US5585871A (en) | 1995-05-26 | 1996-12-17 | Linden; Harry | Multi-function display apparatus |
US5746501A (en) | 1995-09-01 | 1998-05-05 | Chien; Tseng Lu | Portable object having a fastening band illuminated by a super thin lighting element |
US6183425B1 (en) | 1995-10-13 | 2001-02-06 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Method and apparatus for monitoring of daily activity in terms of ground reaction forces |
US5966680A (en) | 1996-02-15 | 1999-10-12 | Butnaru; Hanan | Motion sickness/vertigo prevention device and method |
CN1310616C (en) | 1996-04-08 | 2007-04-18 | 精工爱普生株式会社 | Motion prescription support device |
US5976083A (en) | 1997-07-30 | 1999-11-02 | Living Systems, Inc. | Portable aerobic fitness monitor for walking and running |
US5891042A (en) | 1997-09-09 | 1999-04-06 | Acumen, Inc. | Fitness monitoring device having an electronic pedometer and a wireless heart rate monitor |
US6381482B1 (en) | 1998-05-13 | 2002-04-30 | Georgia Tech Research Corp. | Fabric or garment with integrated flexible information infrastructure |
US6013007A (en) | 1998-03-26 | 2000-01-11 | Liquid Spark, Llc | Athlete's GPS-based performance monitor |
US5931764A (en) | 1998-06-24 | 1999-08-03 | Viztec, Inc. | Wearable device with flexible display |
US7376238B1 (en) | 1998-09-18 | 2008-05-20 | Rivas Technologies International, Inc. | Pulse rate, pressure and heart condition monitoring glasses |
US6218958B1 (en) | 1998-10-08 | 2001-04-17 | International Business Machines Corporation | Integrated touch-skin notification system for wearable computing devices |
US6532298B1 (en) | 1998-11-25 | 2003-03-11 | Iridian Technologies, Inc. | Portable authentication device and method using iris patterns |
WO2000046581A1 (en) | 1999-02-05 | 2000-08-10 | Curtis Instruments, Inc. | Shaft sensor for angular velocity, torque, power |
EP1217942A1 (en) | 1999-09-24 | 2002-07-03 | Healthetech, Inc. | Physiological monitor and associated computation, display and communication unit |
US6736759B1 (en) | 1999-11-09 | 2004-05-18 | Paragon Solutions, Llc | Exercise monitoring system and methods |
US6431705B1 (en) | 1999-11-10 | 2002-08-13 | Infoeye | Eyewear heart rate monitor |
US7156809B2 (en) | 1999-12-17 | 2007-01-02 | Q-Tec Systems Llc | Method and apparatus for health and disease management combining patient data monitoring with wireless internet connectivity |
US7454002B1 (en) | 2000-01-03 | 2008-11-18 | Sportbrain, Inc. | Integrating personal data capturing functionality into a portable computing device and a wireless communication device |
US6513532B2 (en) | 2000-01-19 | 2003-02-04 | Healthetech, Inc. | Diet and activity-monitoring device |
JP2001297318A (en) | 2000-04-14 | 2001-10-26 | Omron Corp | Pedometer |
WO2001088857A1 (en) | 2000-05-16 | 2001-11-22 | Swisscom Mobile Ag | Biometric method for identification and authorisation |
JP2001344352A (en) | 2000-05-31 | 2001-12-14 | Toshiba Corp | Life assisting device, life assisting method and advertisement information providing method |
US6325507B1 (en) | 2000-06-02 | 2001-12-04 | Oakley, Inc. | Eyewear retention system extending across the top of a wearer's head |
DE10043797A1 (en) | 2000-09-06 | 2002-03-28 | Daimler Chrysler Ag | Integrated traffic monitoring system |
US20050054942A1 (en) | 2002-01-22 | 2005-03-10 | Melker Richard J. | System and method for therapeutic drug monitoring |
EP1344092A2 (en) | 2000-12-15 | 2003-09-17 | Daniel Rosenfeld | Location-based weather nowcast system and method |
US20020151810A1 (en) | 2001-04-16 | 2002-10-17 | Acumen, Inc. | Wrist-based fitness monitoring devices |
US6769767B2 (en) | 2001-04-30 | 2004-08-03 | Qr Spex, Inc. | Eyewear with exchangeable temples housing a transceiver forming ad hoc networks with other devices |
JP2003033328A (en) | 2001-07-19 | 2003-02-04 | Nippon Seimitsu Sokki Kk | Heart rate monitor and method for measuring heart rate |
JP2003242584A (en) | 2002-02-13 | 2003-08-29 | Seiko Instruments Inc | Wearable electronic equipment system and wearable electronic equipment |
AU2003272191B2 (en) | 2002-04-22 | 2008-07-03 | Abreu, Marcio Marc Aurelio Martins | Apparatus and method for measuring biologic parameters |
US9153074B2 (en) | 2011-07-18 | 2015-10-06 | Dylan T X Zhou | Wearable augmented reality eyeglass communication device including mobile phone and mobile computing via virtual touch screen gesture control and neuron command |
GB2396421A (en) | 2002-12-16 | 2004-06-23 | Orange Personal Comm Serv Ltd | Head-worn device measuring brain and facial muscle activity |
MXPA05009159A (en) | 2003-02-26 | 2005-12-05 | Marc Aurelio Martins Ab Marcio | Appartus and method for measuring biologic parameters. |
US7792552B2 (en) | 2003-04-15 | 2010-09-07 | Ipventure, Inc. | Eyeglasses for wireless communications |
US7922321B2 (en) | 2003-10-09 | 2011-04-12 | Ipventure, Inc. | Eyewear supporting after-market electrical components |
US7806525B2 (en) | 2003-10-09 | 2010-10-05 | Ipventure, Inc. | Eyeglasses having a camera |
US8109629B2 (en) | 2003-10-09 | 2012-02-07 | Ipventure, Inc. | Eyewear supporting electrical components and apparatus therefor |
US7500747B2 (en) | 2003-10-09 | 2009-03-10 | Ipventure, Inc. | Eyeglasses with electrical components |
US7500746B1 (en) | 2004-04-15 | 2009-03-10 | Ip Venture, Inc. | Eyewear with radiation detection system |
US7380936B2 (en) | 2003-10-09 | 2008-06-03 | Ipventure, Inc. | Eyeglasses with a clock or other electrical component |
US8465151B2 (en) | 2003-04-15 | 2013-06-18 | Ipventure, Inc. | Eyewear with multi-part temple for supporting one or more electrical components |
US7255437B2 (en) | 2003-10-09 | 2007-08-14 | Howell Thomas A | Eyeglasses with activity monitoring |
US7192136B2 (en) | 2003-04-15 | 2007-03-20 | Howell Thomas A | Tethered electrical components for eyeglasses |
US7581833B2 (en) | 2003-10-09 | 2009-09-01 | Ipventure, Inc. | Eyewear supporting after-market electrical components |
US20050033200A1 (en) | 2003-08-05 | 2005-02-10 | Soehren Wayne A. | Human motion identification and measurement system and method |
US7059717B2 (en) | 2003-08-11 | 2006-06-13 | Bloch Nigel K | Eyeglasses with interchangable temple-members |
US7438410B1 (en) | 2003-10-09 | 2008-10-21 | Ip Venture, Inc. | Tethered electrical components for eyeglasses |
US10310296B2 (en) | 2003-10-09 | 2019-06-04 | Ingeniospec, Llc | Eyewear with printed circuit board |
US7677723B2 (en) | 2003-10-09 | 2010-03-16 | Ipventure, Inc. | Eyeglasses with a heart rate monitor |
FR2860700B1 (en) | 2003-10-10 | 2005-12-09 | Commissariat Energie Atomique | CROWN CONTROL DEVICE |
CN102670163B (en) | 2004-04-01 | 2016-04-13 | 威廉·C·托奇 | The system and method for controlling calculation device |
US8337013B2 (en) | 2004-07-28 | 2012-12-25 | Ipventure, Inc. | Eyeglasses with RFID tags or with a strap |
US8915588B2 (en) | 2004-11-02 | 2014-12-23 | E-Vision Smart Optics, Inc. | Eyewear including a heads up display |
US7793361B2 (en) | 2004-11-12 | 2010-09-14 | Nike, Inc. | Article of apparel incorporating a separable electronic device |
US20060115130A1 (en) | 2004-11-29 | 2006-06-01 | Douglas Kozlay | Eyewear with biometrics to protect displayed data |
US7457434B2 (en) | 2005-04-04 | 2008-11-25 | Massachusetts Eye & Ear Infirmary | Adaptively focusing extra-ocular vision prostheses |
US7400257B2 (en) | 2005-04-06 | 2008-07-15 | Rivas Victor A | Vital signals and glucose monitoring personal wireless system |
WO2006119186A2 (en) * | 2005-05-02 | 2006-11-09 | University Of Virginia Patent Foundation | Systems, devices, and methods for interpreting movement |
US8979295B2 (en) | 2005-05-17 | 2015-03-17 | Michael Waters | Rechargeable lighted glasses |
US20070052672A1 (en) | 2005-09-08 | 2007-03-08 | Swisscom Mobile Ag | Communication device, system and method |
US20070112287A1 (en) | 2005-09-13 | 2007-05-17 | Fancourt Craig L | System and method for detecting deviations in nominal gait patterns |
US20070081123A1 (en) | 2005-10-07 | 2007-04-12 | Lewis Scott W | Digital eyewear |
US9658473B2 (en) | 2005-10-07 | 2017-05-23 | Percept Technologies Inc | Enhanced optical and perceptual digital eyewear |
US8696113B2 (en) | 2005-10-07 | 2014-04-15 | Percept Technologies Inc. | Enhanced optical and perceptual digital eyewear |
US11428937B2 (en) | 2005-10-07 | 2022-08-30 | Percept Technologies | Enhanced optical and perceptual digital eyewear |
US7648463B1 (en) | 2005-12-15 | 2010-01-19 | Impact Sports Technologies, Inc. | Monitoring device, method and system |
GB0602127D0 (en) | 2006-02-02 | 2006-03-15 | Imp Innovations Ltd | Gait analysis |
US8188868B2 (en) | 2006-04-20 | 2012-05-29 | Nike, Inc. | Systems for activating and/or authenticating electronic devices for operation with apparel |
US7558622B2 (en) | 2006-05-24 | 2009-07-07 | Bao Tran | Mesh network stroke monitoring appliance |
US7539533B2 (en) | 2006-05-16 | 2009-05-26 | Bao Tran | Mesh network monitoring appliance |
US7543934B2 (en) | 2006-09-20 | 2009-06-09 | Ipventures, Inc. | Eyeglasses with activity monitoring and acoustic dampening |
WO2008073806A1 (en) | 2006-12-08 | 2008-06-19 | Sabic Innovative Plastics Ip B.V. | Active transdermal drug delivery system |
US8157730B2 (en) * | 2006-12-19 | 2012-04-17 | Valencell, Inc. | Physiological and environmental monitoring systems and methods |
JP2008198028A (en) | 2007-02-14 | 2008-08-28 | Sony Corp | Wearable device, authentication method and program |
EP2149068B1 (en) | 2007-04-23 | 2021-06-09 | Huawei Technologies Co., Ltd. | Eyewear having human activity monitoring device |
US8944590B2 (en) | 2010-07-02 | 2015-02-03 | Mitsui Chemicals, Inc. | Electronic spectacle frames |
US7634379B2 (en) | 2007-05-18 | 2009-12-15 | Ultimate Balance, Inc. | Newtonian physical activity monitor |
US8750971B2 (en) | 2007-05-24 | 2014-06-10 | Bao Tran | Wireless stroke monitoring |
EP2165234A1 (en) | 2007-06-07 | 2010-03-24 | Panagiotis Pavlopoulos | An eyewear comprising at least one display device |
US9254100B2 (en) | 2007-09-12 | 2016-02-09 | Cardiac Pacemakers, Inc. | Logging daily average metabolic activity using a motion sensor |
EP2233081B2 (en) | 2007-09-21 | 2018-03-28 | Covidien LP | Surgical device |
US8448846B2 (en) | 2007-11-18 | 2013-05-28 | Intel-Ge Care Innovations Llc | Medication recording device |
US8202148B2 (en) | 2007-12-03 | 2012-06-19 | Julius Young | Machine and method for caddying and golf instruction |
US20090195747A1 (en) | 2008-02-04 | 2009-08-06 | Insua Luisa M | Interchangeable eyeglass temples |
US20090227853A1 (en) | 2008-03-03 | 2009-09-10 | Ravindra Wijesiriwardana | Wearable optical pulse plethysmography sensors or pulse oximetry sensors based wearable heart rate monitoring systems |
US20120142443A1 (en) | 2008-03-17 | 2012-06-07 | Chris Savarese | Golf club apparatuses and methods |
CN101566874A (en) | 2008-04-24 | 2009-10-28 | 鸿富锦精密工业(深圳)有限公司 | Control device and electronic equipment using same |
RU2011106031A (en) | 2008-07-18 | 2012-08-27 | Опталерт Пти Лтд (Au) | SENSITIVE DEVICE FOR ACTIVE STATE |
US8011242B2 (en) | 2008-07-29 | 2011-09-06 | Garmin Switzerland Gmbh | System and device for measuring and analyzing forces applied by a cyclist on a pedal of a bicycle |
US20100045928A1 (en) | 2008-08-25 | 2010-02-25 | Tri-Specs, Inc. | Fashion eyewear frame that houses circuitry to effect wireless audio communication while providing extraneous background noise cancellation capability |
CN103119512A (en) | 2008-11-02 | 2013-05-22 | 大卫·乔姆 | Near to eye display system and appliance |
US8494507B1 (en) | 2009-02-16 | 2013-07-23 | Handhold Adaptive, LLC | Adaptive, portable, multi-sensory aid for the disabled |
WO2010096691A2 (en) | 2009-02-20 | 2010-08-26 | The Regents Of The University Of Colorado, A Body Corporate | Footwear-based body weight monitor and postural allocation, physical activity classification, and energy expenditure calculator |
US8700111B2 (en) | 2009-02-25 | 2014-04-15 | Valencell, Inc. | Light-guiding devices and monitoring devices incorporating same |
US20130024211A1 (en) | 2009-04-09 | 2013-01-24 | Access Mobility, Inc. | Active learning and advanced relationship marketing and health interventions |
US20100280336A1 (en) | 2009-04-30 | 2010-11-04 | Medtronic, Inc. | Anxiety disorder monitoring |
US8081082B2 (en) | 2009-05-27 | 2011-12-20 | International Business Machines Corporation | Monitoring patterns of motion |
US20100308999A1 (en) | 2009-06-05 | 2010-12-09 | Chornenky Todd E | Security and monitoring apparatus |
US8253561B2 (en) | 2009-06-10 | 2012-08-28 | Betty L. Bowers | Medication management apparatus and system |
US10748447B2 (en) | 2013-05-24 | 2020-08-18 | Lincoln Global, Inc. | Systems and methods providing a computerized eyewear device to aid in welding |
US20130009907A1 (en) | 2009-07-31 | 2013-01-10 | Rosenberg Ilya D | Magnetic Stylus |
WO2011040259A1 (en) * | 2009-09-30 | 2011-04-07 | 三菱化学株式会社 | Body movement signal information processing method, information processing system and information processing device |
US8303311B2 (en) | 2009-09-30 | 2012-11-06 | Forest Carl A | Sport personal coach system |
JP5504810B2 (en) | 2009-10-06 | 2014-05-28 | オムロンヘルスケア株式会社 | Walking posture determination device, control program, and control method |
US8605165B2 (en) | 2010-10-06 | 2013-12-10 | Ai Cure Technologies Llc | Apparatus and method for assisting monitoring of medication adherence |
US8290558B1 (en) | 2009-11-23 | 2012-10-16 | Vioptix, Inc. | Tissue oximeter intraoperative sensor |
FR2953284A1 (en) | 2009-12-02 | 2011-06-03 | Movea Sa | SYSTEM AND METHOD FOR DRIVER ASSISTANCE OF BIOMECHANIC DRIVE VEHICLE COMPRISING AT LEAST ONE WHEEL |
EP2333778A1 (en) | 2009-12-04 | 2011-06-15 | Lg Electronics Inc. | Digital data reproducing apparatus and method for controlling the same |
US20110169932A1 (en) | 2010-01-06 | 2011-07-14 | Clear View Technologies Inc. | Wireless Facial Recognition |
IT1397737B1 (en) | 2010-01-18 | 2013-01-24 | Giovanni Saggio | EQUIPMENT AND METHOD OF DETECTION, TRAINING AND TRAINING OF MOTOR ACTIVITIES |
CN102906623A (en) | 2010-02-28 | 2013-01-30 | 奥斯特豪特集团有限公司 | Local advertising content on an interactive head-mounted eyepiece |
US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
US20120206485A1 (en) | 2010-02-28 | 2012-08-16 | Osterhout Group, Inc. | Ar glasses with event and sensor triggered user movement control of ar eyepiece facilities |
US8964298B2 (en) | 2010-02-28 | 2015-02-24 | Microsoft Corporation | Video display modification based on sensor input for a see-through near-to-eye display |
US20110224505A1 (en) | 2010-03-12 | 2011-09-15 | Rajendra Padma Sadhu | User wearable portable communicative device |
US9652965B2 (en) | 2010-03-12 | 2017-05-16 | Rajendra Padma Sadhu | System and method for transmitting alerts and notifications to a user |
US8568313B2 (en) | 2010-03-12 | 2013-10-29 | Rajendra Padma Sadhu | User wearable portable communication device for collection and transmission of physiological data |
WO2011143655A1 (en) | 2010-05-14 | 2011-11-17 | Advitech, Inc. | System and method for prevention and control of the effects of spatial disorientation |
WO2011158965A1 (en) | 2010-06-17 | 2011-12-22 | 日本電気株式会社 | Sensitivity evaluation system, sensitivity evaluation method, and program |
US8531355B2 (en) | 2010-07-23 | 2013-09-10 | Gregory A. Maltz | Unitized, vision-controlled, wireless eyeglass transceiver |
US20120029367A1 (en) | 2010-07-31 | 2012-02-02 | Hobeika Hind Louis | Heart rate waterproof measuring apparatus |
US9247212B2 (en) | 2010-08-26 | 2016-01-26 | Blast Motion Inc. | Intelligent motion capture element |
US8594971B2 (en) | 2010-09-22 | 2013-11-26 | Invensense, Inc. | Deduced reckoning navigation without a constraint relationship between orientation of a sensor platform and a direction of travel of an object |
US9241635B2 (en) | 2010-09-30 | 2016-01-26 | Fitbit, Inc. | Portable monitoring devices for processing applications and processing analysis of physiological conditions of a user associated with the portable monitoring device |
US8849610B2 (en) | 2010-09-30 | 2014-09-30 | Fitbit, Inc. | Tracking user physical activity with multiple devices |
US8615377B1 (en) | 2010-09-30 | 2013-12-24 | Fitbit, Inc. | Methods and systems for processing social interactive data and sharing of tracked activity associated with locations |
US8762102B2 (en) | 2010-09-30 | 2014-06-24 | Fitbit, Inc. | Methods and systems for generation and rendering interactive events having combined activity and location information |
US8738323B2 (en) | 2010-09-30 | 2014-05-27 | Fitbit, Inc. | Methods and systems for metrics analysis and interactive rendering, including events having combined activity and location information |
US10216893B2 (en) | 2010-09-30 | 2019-02-26 | Fitbit, Inc. | Multimode sensor devices |
US8712724B2 (en) | 2010-09-30 | 2014-04-29 | Fitbit, Inc. | Calendar integration methods and systems for presentation of events having combined activity and location information |
US8738321B2 (en) | 2010-09-30 | 2014-05-27 | Fitbit, Inc. | Methods and systems for classification of geographic locations for tracked activity |
EP2439580A1 (en) | 2010-10-01 | 2012-04-11 | Ophtimalia | Data exchange system |
KR101346661B1 (en) | 2010-11-15 | 2014-02-06 | 부경대학교 산학협력단 | Cosmetic composition for preventing skin aging comprising chitooligosaccharides |
US9113793B2 (en) | 2010-12-10 | 2015-08-25 | Rohm Co., Ltd. | Pulse wave sensor |
US20120169990A1 (en) | 2011-01-05 | 2012-07-05 | Burnstein Tracey E | Electronic eyewear and footwear |
US20120191016A1 (en) * | 2011-01-25 | 2012-07-26 | Harris Corporation | Gait based notification and control of portable devices |
US20120203310A1 (en) | 2011-02-04 | 2012-08-09 | Pugh Randall B | Spectacles for light therapy |
US9785242B2 (en) | 2011-03-12 | 2017-10-10 | Uday Parshionikar | Multipurpose controllers and methods |
US9317660B2 (en) | 2011-03-31 | 2016-04-19 | Adidas Ag | Group performance monitoring system and method |
US8510166B2 (en) | 2011-05-11 | 2013-08-13 | Google Inc. | Gaze tracking system |
US8911087B2 (en) | 2011-05-20 | 2014-12-16 | Eyefluence, Inc. | Systems and methods for measuring reactions of head, eyes, eyelids and pupils |
US9256711B2 (en) | 2011-07-05 | 2016-02-09 | Saudi Arabian Oil Company | Systems, computer medium and computer-implemented methods for providing health information to employees via augmented reality display |
US8184067B1 (en) | 2011-07-20 | 2012-05-22 | Google Inc. | Nose bridge sensor |
US9342610B2 (en) | 2011-08-25 | 2016-05-17 | Microsoft Technology Licensing, Llc | Portals: registered objects as virtualized, personalized displays |
TWI455342B (en) | 2011-08-30 | 2014-10-01 | Nat Univ Tsing Hua | Solar cell with selective emitter structure and manufacturing method thereof |
KR20130025675A (en) | 2011-09-02 | 2013-03-12 | 삼성전자주식회사 | User health monitoring system which comprises 3d glasses and display apparatus, and display apparatus and control method thereof |
US8941560B2 (en) | 2011-09-21 | 2015-01-27 | Google Inc. | Wearable computer with superimposed controls and instructions for external device |
US20150057512A1 (en) | 2011-11-16 | 2015-02-26 | Rijuven Corporation | Wearable heart failure monitor patch |
US20130138413A1 (en) | 2011-11-24 | 2013-05-30 | Auckland Uniservices Limited | System and Method for Determining Motion |
US8540583B2 (en) | 2011-12-30 | 2013-09-24 | Nike, Inc. | System for tracking a golf ball and displaying an enhanced image of the golf ball |
US9141194B1 (en) | 2012-01-04 | 2015-09-22 | Google Inc. | Magnetometer-based gesture sensing with a wearable device |
US9529197B2 (en) | 2012-03-21 | 2016-12-27 | Google Inc. | Wearable device with input and output structures |
US9737261B2 (en) | 2012-04-13 | 2017-08-22 | Adidas Ag | Wearable athletic activity monitoring systems |
US9504414B2 (en) | 2012-04-13 | 2016-11-29 | Adidas Ag | Wearable athletic activity monitoring methods and systems |
US20130307670A1 (en) | 2012-05-15 | 2013-11-21 | Jonathan E. Ramaci | Biometric authentication system |
US9763592B2 (en) | 2012-05-25 | 2017-09-19 | Emotiv, Inc. | System and method for instructing a behavior change in a user |
US9001427B2 (en) | 2012-05-30 | 2015-04-07 | Microsoft Technology Licensing, Llc | Customized head-mounted display device |
US20130329183A1 (en) | 2012-06-11 | 2013-12-12 | Pixeloptics, Inc. | Adapter For Eyewear |
US8948832B2 (en) | 2012-06-22 | 2015-02-03 | Fitbit, Inc. | Wearable heart rate monitor |
US9005129B2 (en) | 2012-06-22 | 2015-04-14 | Fitbit, Inc. | Wearable heart rate monitor |
US9599632B2 (en) | 2012-06-22 | 2017-03-21 | Fitbit, Inc. | Fitness monitoring device with altimeter |
US9035970B2 (en) | 2012-06-29 | 2015-05-19 | Microsoft Technology Licensing, Llc | Constraint based information inference |
US9579048B2 (en) | 2012-07-30 | 2017-02-28 | Treefrog Developments, Inc | Activity monitoring system with haptic feedback |
WO2014022438A1 (en) | 2012-07-30 | 2014-02-06 | Treefrog Developments, Inc. | Athletic monitoring |
US20150156196A1 (en) | 2012-07-31 | 2015-06-04 | Intellectual Discovery Co., Ltd. | Wearable electronic device and method for controlling same |
US9720231B2 (en) | 2012-09-26 | 2017-08-01 | Dolby Laboratories Licensing Corporation | Display, imaging system and controller for eyewear display device |
JP6021582B2 (en) | 2012-10-24 | 2016-11-09 | オリンパス株式会社 | Glasses-type wearable device and front part of glasses-type wearable device |
US10143405B2 (en) | 2012-11-14 | 2018-12-04 | MAD Apparel, Inc. | Wearable performance monitoring, analysis, and feedback systems and methods |
ITMI20121957A1 (en) | 2012-11-16 | 2014-05-17 | Marco Carrara | GLASSES WITH HIGH FLEXIBILITY OF USE |
US10045718B2 (en) | 2012-11-22 | 2018-08-14 | Atheer, Inc. | Method and apparatus for user-transparent system control using bio-input |
US20140218281A1 (en) | 2012-12-06 | 2014-08-07 | Eyefluence, Inc. | Systems and methods for eye gaze determination |
GB2496064B (en) | 2012-12-31 | 2015-03-11 | Nicholas Jamie Marston | Video camera shooting glasses |
ITMI20130024A1 (en) | 2013-01-10 | 2014-07-11 | Marco Carrara | METHOD OF ACQUISITION AND TREATMENT OF HEART RATE DATA |
US9520638B2 (en) | 2013-01-15 | 2016-12-13 | Fitbit, Inc. | Hybrid radio frequency / inductive loop antenna |
US20140204334A1 (en) | 2013-01-18 | 2014-07-24 | William Anthony Stoll | Bio-sensors in eyeglasses |
US9370302B2 (en) | 2014-07-08 | 2016-06-21 | Wesley W. O. Krueger | System and method for the measurement of vestibulo-ocular reflex to improve human performance in an occupational environment |
US9848776B2 (en) | 2013-03-04 | 2017-12-26 | Hello Inc. | Methods using activity manager for monitoring user activity |
US9500464B2 (en) | 2013-03-12 | 2016-11-22 | Adidas Ag | Methods of determining performance information for individuals and sports objects |
US20140276096A1 (en) * | 2013-03-15 | 2014-09-18 | Bonutti Research, Inc. | Systems and methods for use in diagnosing a medical condition of a patient |
ES2900423T3 (en) | 2013-03-15 | 2022-03-16 | Percept Tech Inc | Enhanced optical and perceptual digital glasses |
WO2014144918A2 (en) | 2013-03-15 | 2014-09-18 | Percept Technologies, Inc. | Enhanced optical and perceptual digital eyewear |
WO2014145166A2 (en) | 2013-03-15 | 2014-09-18 | Eyecam, LLC | Autonomous computing and telecommunications head-up displays glasses |
US9341526B2 (en) | 2013-04-01 | 2016-05-17 | Saris Cycling Group, Inc. | System for speed-based power calculation |
DE102013207064A1 (en) | 2013-04-19 | 2014-10-23 | Bayerische Motoren Werke Aktiengesellschaft | Method for selecting an information source for display on data glasses |
US20140324459A1 (en) | 2013-04-30 | 2014-10-30 | Hti Ip, L.L.C | Automatic health monitoring alerts |
US10930174B2 (en) | 2013-05-24 | 2021-02-23 | Lincoln Global, Inc. | Systems and methods providing a computerized eyewear device to aid in welding |
US20140375470A1 (en) | 2013-06-20 | 2014-12-25 | Chester Charles Malveaux | Wearable networked and standalone biometric sensor system to record and transmit biometric data for multiple applications |
US10512407B2 (en) | 2013-06-24 | 2019-12-24 | Fitbit, Inc. | Heart rate data collection |
ES2530421B1 (en) | 2013-07-30 | 2015-07-09 | Ion Eyewear, S.L. | ACCESSORY GLASSES FOR MOBILE DEVICES AND PC¿S |
FR3009270B1 (en) | 2013-07-31 | 2016-09-09 | Michelin & Cie | DEVICE AND METHOD FOR CONTROLLING THE POWER OF ASSISTANCE OF A POWER-ASSISTED VELO |
EP2833196B1 (en) | 2013-08-02 | 2016-03-16 | ESSILOR INTERNATIONAL (Compagnie Générale d'Optique) | A method of controlling a programmable ophthalmic lens device |
US9704412B2 (en) | 2013-08-26 | 2017-07-11 | John Andrew Wells | Biometric data gathering |
US20150065889A1 (en) | 2013-09-02 | 2015-03-05 | Life Beam Technologies Ltd. | Bodily worn multiple optical sensors heart rate measuring device and method |
US20150148632A1 (en) | 2013-11-26 | 2015-05-28 | David Alan Benaron | Calorie Monitoring Sensor And Method For Cell Phones, Smart Watches, Occupancy Sensors, And Wearables |
US9595181B2 (en) | 2013-12-20 | 2017-03-14 | Invensense, Inc. | Wearable device assisting smart media application and vice versa |
TWI548438B (en) | 2013-12-20 | 2016-09-11 | 岱宇國際股份有限公司 | Exercise device providing symmetry index |
US9947012B2 (en) | 2013-12-26 | 2018-04-17 | Intel Corporation | Secure transactions using a personal device |
WO2015095924A1 (en) | 2013-12-27 | 2015-07-02 | Koonung Heights Pty Ltd | A biofeedback, stress management and cognitive enhancement system |
US9579060B1 (en) | 2014-02-18 | 2017-02-28 | Orbitol Research Inc. | Head-mounted physiological signal monitoring system, devices and methods |
KR102204265B1 (en) | 2014-02-24 | 2021-01-15 | 소니 주식회사 | Smart wearable devices and methods for acquisition of sensorial information from wearable devices to activate functions in other devices |
CA2939922A1 (en) | 2014-02-24 | 2015-08-27 | Brain Power, Llc | Systems, environment and methods for evaluation and management of autism spectrum disorder using a wearable data collection device |
US9031812B2 (en) * | 2014-02-27 | 2015-05-12 | Fitbit, Inc. | Notifications on a user device based on activity detected by an activity monitoring device |
FR3019346B1 (en) | 2014-03-31 | 2016-07-08 | Withings | METHOD FOR COUNTING STEPS PERFORMED BY A USER |
CN105893721A (en) | 2014-05-13 | 2016-08-24 | 陈威宇 | Adaptive skin care information prompt system and adaptive skin care prompt method |
US20150332149A1 (en) | 2014-05-15 | 2015-11-19 | Red Lozenge, Inc. | Tracking behavior and goal achievement |
US10478127B2 (en) | 2014-06-23 | 2019-11-19 | Sherlock Solutions, LLC | Apparatuses, methods, processes, and systems related to significant detrimental changes in health parameters and activating lifesaving measures |
TWI530276B (en) | 2014-07-08 | 2016-04-21 | 原相科技股份有限公司 | Biometric detection module with denoising function and biometric detection method thereof |
WO2016017997A1 (en) | 2014-07-31 | 2016-02-04 | Samsung Electronics Co., Ltd. | Wearable glasses and method of providing content using the same |
US20160041404A1 (en) | 2014-08-08 | 2016-02-11 | Marchon Eyewear, Inc. | Eyewear with interchangeable temples and brow bar |
HK1203120A2 (en) | 2014-08-26 | 2015-10-16 | 高平 | A gait monitor and a method of monitoring the gait of a person |
US10617342B2 (en) | 2014-09-05 | 2020-04-14 | Vision Service Plan | Systems, apparatus, and methods for using a wearable device to monitor operator alertness |
US10448867B2 (en) | 2014-09-05 | 2019-10-22 | Vision Service Plan | Wearable gait monitoring apparatus, systems, and related methods |
US20160066848A1 (en) | 2014-09-05 | 2016-03-10 | Vision Service Plan | Wearable environmental pollution monitor computer apparatus, systems, and related methods |
US20160117937A1 (en) | 2014-10-27 | 2016-04-28 | Bloom Technologies NV | System and method for providing biometric and context based messaging |
US9566033B2 (en) | 2014-11-03 | 2017-02-14 | Phillip Bogdanovich | Garment system with electronic components and associated methods |
KR102313220B1 (en) | 2015-01-09 | 2021-10-15 | 삼성전자주식회사 | Wearable device and method for controlling thereof |
US20160223577A1 (en) | 2015-01-30 | 2016-08-04 | Vision Service Plan | Systems and methods for tracking motion of a bicycle or other vehicles |
CN107205673A (en) | 2015-01-30 | 2017-09-26 | 皇家飞利浦有限公司 | Photo-plethysmographic device |
US10349887B1 (en) | 2015-06-14 | 2019-07-16 | Facense Ltd. | Blood pressure measuring smartglasses |
US10398328B2 (en) | 2015-08-25 | 2019-09-03 | Koninklijke Philips N.V. | Device and system for monitoring of pulse-related information of a subject |
US9726904B1 (en) | 2015-09-29 | 2017-08-08 | Snap Inc. | Eyewear with conductive temple joint |
FR3043245B1 (en) | 2015-11-03 | 2017-10-27 | Stmicroelectronics Rousset | METHOD FOR READING AN EEPROM MEMORY AND CORRESPONDING DEVICE |
US9610476B1 (en) | 2016-05-02 | 2017-04-04 | Bao Tran | Smart sport device |
US20170255029A1 (en) | 2016-03-03 | 2017-09-07 | Vision Service Plan | Systems and methods for charging eyewear |
US20180064399A1 (en) | 2016-09-07 | 2018-03-08 | Heptagon Micro Optics Pte. Ltd. | Imaging systems including multi-tap demodulation pixels for biometric measurements |
EP3919969A1 (en) | 2016-09-22 | 2021-12-08 | Essilor International | Health monitoring device and wearing detection module for spectacles frame |
US20180206735A1 (en) | 2017-01-26 | 2018-07-26 | Microsoft Technology Licensing, Llc | Head-mounted device for capturing pulse data |
US10874305B2 (en) | 2018-01-15 | 2020-12-29 | Microsoft Technology Licensing, Llc | Sensor device |
-
2014
- 2014-11-21 US US14/550,406 patent/US10448867B2/en active Active
- 2014-12-05 US US14/562,454 patent/US9795324B2/en active Active
- 2014-12-19 US US14/578,039 patent/US20160066829A1/en not_active Abandoned
- 2014-12-31 US US14/588,122 patent/US9649052B2/en active Active
-
2015
- 2015-01-30 US US14/610,439 patent/US10307085B2/en active Active
- 2015-01-30 US US14/610,589 patent/US20160071423A1/en not_active Abandoned
- 2015-01-30 US US14/610,628 patent/US20160070122A1/en not_active Abandoned
- 2015-09-04 WO PCT/US2015/048612 patent/WO2016037091A1/en active Application Filing
- 2015-09-04 CA CA2960425A patent/CA2960425A1/en not_active Abandoned
- 2015-09-04 CA CA2960429A patent/CA2960429A1/en not_active Abandoned
- 2015-09-04 EP EP15839047.6A patent/EP3189371A4/en not_active Withdrawn
- 2015-09-04 EP EP15838411.5A patent/EP3148435B1/en active Active
- 2015-09-04 WO PCT/US2015/048656 patent/WO2016037117A1/en active Application Filing
- 2015-09-04 CA CA2953856A patent/CA2953856A1/en not_active Abandoned
- 2015-09-04 EP EP15838234.1A patent/EP3189367B1/en active Active
- 2015-09-04 WO PCT/US2015/048662 patent/WO2016037120A1/en active Application Filing
-
2017
- 2017-05-15 US US15/594,898 patent/US10188323B2/en active Active
- 2017-10-23 US US15/791,196 patent/US20180042523A1/en not_active Abandoned
-
2019
- 2019-01-28 US US16/259,646 patent/US10542915B2/en active Active
- 2019-06-03 US US16/429,480 patent/US10694981B2/en active Active
- 2019-10-18 US US16/657,982 patent/US20200046260A1/en not_active Abandoned
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110301446A1 (en) * | 2007-01-15 | 2011-12-08 | Deka Products Limited Partnership | Device and Method for Food Management |
US20100042430A1 (en) * | 2008-08-12 | 2010-02-18 | Irody Inc | System and method for collecting and authenticating medication consumption |
US20100136508A1 (en) * | 2008-10-23 | 2010-06-03 | Damir Zekhtser | Meal Plan Management |
US20100332571A1 (en) * | 2009-06-30 | 2010-12-30 | Jennifer Healey | Device augmented food identification |
US20120135384A1 (en) * | 2010-11-26 | 2012-05-31 | Terumo Kabushiki Kaisha | Portable terminal, calorie estimation method, and calorie estimation program |
US20130157232A1 (en) * | 2011-12-09 | 2013-06-20 | Joel Ehrenkranz | System and methods for monitoring food consumption |
Non-Patent Citations (2)
Title |
---|
Phend, Crystal. "Calorie Intake Rises as Sleep Time Drops." Medpage Today. MedpageToday, LLC, 15 Mar. 2012. Web. 19 Dec. 2016. <http://www.medpagetoday.com/cardiology/prevention/31663>. * |
Shankland, Stephen. "Google's electronic eyewear gets 'OK Glass' voice commands'. Published Feb 20 2013. Cnet.com. <https://www.cnet.com/news/googles-electronic-eyewear-gets-ok-glass-voice-commands/>. * |
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210307642A1 (en) * | 2015-07-27 | 2021-10-07 | Massachusetts Institute Of Technology | Systems, apparatus, and methods related to modeling, monitoring, and/or managing metabolism |
US10861594B2 (en) | 2015-10-01 | 2020-12-08 | Dnanudge Limited | Product recommendation system and method |
US10650919B2 (en) | 2015-10-01 | 2020-05-12 | Dnanudge Limited | Wearable device |
US11133095B2 (en) | 2015-10-01 | 2021-09-28 | Dnanudge Limited | Wearable device |
US11589807B2 (en) * | 2017-05-11 | 2023-02-28 | The Regents Of The University Of California | Biosensor for monitoring eyedrop usage compliance |
US10699247B2 (en) | 2017-05-16 | 2020-06-30 | Under Armour, Inc. | Systems and methods for providing health task notifications |
US20180374386A1 (en) * | 2017-06-27 | 2018-12-27 | International Business Machines Corporation | Physical activity and dietary based services |
US10685585B2 (en) | 2017-06-27 | 2020-06-16 | International Business Machines Corporation | Physical activity and dietary based services |
US10262553B2 (en) * | 2017-06-27 | 2019-04-16 | International Business Machines Corporation | Physical activity and dietary based services |
US10832590B2 (en) * | 2017-09-13 | 2020-11-10 | At&T Intellectual Property I, L.P. | Monitoring food intake |
US20190080629A1 (en) * | 2017-09-13 | 2019-03-14 | At&T Intellectual Property I, L.P. | Monitoring food intake |
US11295860B2 (en) | 2018-02-12 | 2022-04-05 | Zoe Limited | Using at home measures to predict clinical state and improving the accuracy of at home measurements/predictions data associated with circadian rhythm and meal timing |
US11183291B2 (en) | 2018-02-12 | 2021-11-23 | Zoe Limited | Generating personalized nutritional recommendations using predicted values of biomarkers |
US11183080B2 (en) | 2018-02-12 | 2021-11-23 | Zoe Limited | Generating predicted values of biomarkers for scoring food |
US11348479B2 (en) | 2018-05-23 | 2022-05-31 | Zoe Limited | Accuracy of measuring nutritional responses in a non-clinical setting |
US11720659B2 (en) | 2018-07-24 | 2023-08-08 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US10582897B2 (en) | 2018-07-24 | 2020-03-10 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US10922397B2 (en) | 2018-07-24 | 2021-02-16 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
US11915151B2 (en) | 2018-08-27 | 2024-02-27 | Zoe Limited | Accuracy of test data outside the clinic |
US11380215B2 (en) * | 2018-08-30 | 2022-07-05 | Kyndryl, Inc. | Reward-based ecosystem for tracking nutritional consumption |
WO2020053307A1 (en) * | 2018-09-12 | 2020-03-19 | Dnanudge | Product recommendation system and method |
CN113168903A (en) * | 2018-09-12 | 2021-07-23 | 基因动力公司 | Product recommendation system and method |
US20200086133A1 (en) * | 2018-09-18 | 2020-03-19 | Biointellisense, Inc. | Validation, compliance, and/or intervention with ear device |
CN109543546A (en) * | 2018-10-26 | 2019-03-29 | 复旦大学 | The gait age estimation method returned based on the distribution of depth sequence |
US10811140B2 (en) | 2019-03-19 | 2020-10-20 | Dnanudge Limited | Secure set-up of genetic related user account |
US11901082B2 (en) | 2019-03-19 | 2024-02-13 | Dnanudge Limited | Secure set-up of genetic related user account |
US10699806B1 (en) | 2019-04-15 | 2020-06-30 | Dnanudge Limited | Monitoring system, wearable monitoring device and method |
US20200372826A1 (en) * | 2019-05-10 | 2020-11-26 | Tso-Cheng Chien | Food quantum tracking tools and methods related thereto |
US11663927B2 (en) * | 2019-05-10 | 2023-05-30 | Tso-Cheng Chien | Food quantum tracking tools and methods related thereto |
US11403382B2 (en) | 2019-07-24 | 2022-08-02 | Dnanudge Limited | Method and device for comparing personal biological data of two users |
Also Published As
Publication number | Publication date |
---|---|
EP3148435B1 (en) | 2023-10-18 |
US20190159700A1 (en) | 2019-05-30 |
US20160066829A1 (en) | 2016-03-10 |
WO2016037091A1 (en) | 2016-03-10 |
EP3189371A1 (en) | 2017-07-12 |
CA2960425A1 (en) | 2016-03-10 |
US10448867B2 (en) | 2019-10-22 |
US20160070121A1 (en) | 2016-03-10 |
EP3189367B1 (en) | 2024-04-10 |
EP3148435A4 (en) | 2018-01-17 |
EP3189371A4 (en) | 2018-05-30 |
EP3189367A1 (en) | 2017-07-12 |
WO2016037117A1 (en) | 2016-03-10 |
US20200046260A1 (en) | 2020-02-13 |
US10188323B2 (en) | 2019-01-29 |
US10307085B2 (en) | 2019-06-04 |
US10542915B2 (en) | 2020-01-28 |
US9795324B2 (en) | 2017-10-24 |
US20170245757A1 (en) | 2017-08-31 |
WO2016037120A1 (en) | 2016-03-10 |
CA2960429A1 (en) | 2016-03-10 |
US20180042523A1 (en) | 2018-02-15 |
US20160071390A1 (en) | 2016-03-10 |
EP3148435A1 (en) | 2017-04-05 |
US20160066847A1 (en) | 2016-03-10 |
US20160070122A1 (en) | 2016-03-10 |
EP3189367A4 (en) | 2018-05-30 |
US10694981B2 (en) | 2020-06-30 |
US9649052B2 (en) | 2017-05-16 |
US20160066820A1 (en) | 2016-03-10 |
CA2953856A1 (en) | 2016-03-10 |
US20190298228A1 (en) | 2019-10-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20160071423A1 (en) | Systems and method for monitoring an individual's compliance with a weight loss plan | |
JP7446295B2 (en) | Automatic detection of physical behavioral events and corresponding adjustment of drug delivery systems | |
US11754542B2 (en) | System for nutritional monitoring and management | |
US20200311609A1 (en) | Adaptive model-based system to automatically quantify fall risk | |
CN109068983B (en) | Method and apparatus for tracking food intake and other behaviors and providing relevant feedback | |
US11341167B2 (en) | Data tagging | |
US9442100B2 (en) | Caloric intake measuring system using spectroscopic and 3D imaging analysis | |
US20160034764A1 (en) | Wearable Imaging Member and Spectroscopic Optical Sensor for Food Identification and Nutrition Modification | |
US9536449B2 (en) | Smart watch and food utensil for monitoring food consumption | |
US9254099B2 (en) | Smart watch and food-imaging member for monitoring food consumption | |
US9529385B2 (en) | Smart watch and human-to-computer interface for monitoring food consumption | |
US20170039336A1 (en) | Health maintenance advisory technology | |
US20160148535A1 (en) | Tracking Nutritional Information about Consumed Food | |
US20180116560A1 (en) | Method and apparatus for monitoring body parts of an individual | |
CN107924720A (en) | Client computing device for healthy related advisory | |
US20150126873A1 (en) | Wearable Spectroscopy Sensor to Measure Food Consumption | |
CN104834827A (en) | Health data processing method and system based on terminal equipment and cloud computation storage | |
US20230245741A1 (en) | Information processing device, information processing system, and information processing method | |
CN114616627A (en) | Automatic detection of physical behavioral events and corresponding adjustment of medication dispensing systems | |
Boland et al. | eNutrition-The next dimension for eHealth? | |
US20220415476A1 (en) | Wearable Device and System for Nutritional Intake Monitoring and Management | |
JP2017102614A (en) | Electronic instrument |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |