WO2008042121A2 - Systems and methods for secure voice identification and medical device interface - Google Patents
Systems and methods for secure voice identification and medical device interface Download PDFInfo
- Publication number
- WO2008042121A2 WO2008042121A2 PCT/US2007/020380 US2007020380W WO2008042121A2 WO 2008042121 A2 WO2008042121 A2 WO 2008042121A2 US 2007020380 W US2007020380 W US 2007020380W WO 2008042121 A2 WO2008042121 A2 WO 2008042121A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- medical device
- set forth
- voice commands
- unlock
- receiving
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/145—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6813—Specially adapted to be attached to a specific body part
- A61B5/6825—Hand
- A61B5/6826—Finger
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/683—Means for maintaining contact with the body
- A61B5/6838—Clamps or clips
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/7475—User input or interface means, e.g. keyboard, pointing device, joystick
- A61B5/749—Voice-controlled interfaces
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
Definitions
- the present invention relates generally to medical devices and, more particularly, to secure voice identification and interface in such medical devices.
- pulse oximetry one technique for monitoring certain physiological characteristics of a patient
- pulse oximeters the devices built based upon pulse oximetry techniques
- Pulse oximetry may be used to measure various blood flow characteristics, such as the oxygen saturation of hemoglobin in arterial blood, the volume of individual blood pulsations supplying the tissue, and/or the rate of blood pulsations corresponding to each heartbeat of a patient.
- Pulse oximeters and other types of medical devices are typically mounted on stands that are positioned around a patient's bed or around an operating room table.
- a caregiver desires to command the medical device (e.g., program, configure, and so-forth) they manipulate controls or push buttons on the medical device itself.
- the medical device typically provides results or responses to commands on a liquid crystal display (“LCD”) screen mounted in an externally visible position within the medical device.
- LCD liquid crystal display
- this conventional configuration has several disadvantages.
- this conventional configuration relies upon physical contact with the medical device to input commands (e.g., pushing a button, turning a knob, and the like).
- Such physical contact raises several concerns. Among these concerns are that in making contact with the medical device, the caregiver may spread illness or disease from room to room. More specifically, a caregiver may accidentally deposit germs (e.g., bacteria, viruses, and so forth) on the medical device while manipulating the device's controls. These germs may then be spread to the patient when a subsequent caregiver touches the medical device and then touches the patient.
- germs e.g., bacteria, viruses, and so forth
- germs e.g., bacteria, viruses, and so forth
- medical devices that rely on physical contact for command input may clutter the caregiver's workspace. For example, because the medical device must be within an arm's length of the caregiver, the medical device may crowd the caregiver - potentially even restricting free movement of the caregiver.
- caregivers may have difficulty manipulating controls with gloved hands. For example, it may be difficult to grasp a knob or press a small button due to the added encumbrance of a glove.
- Fig. 1 is a diagrammatical representation of a medical device featuring an integral microphone in accordance with one embodiment of the present invention
- Fig. 2 is a diagrammatical representation of a medical device featuring an external microphone in accordance with one embodiment of the present invention
- Fig. 3 is a flow chart illustrating an exemplary technique for secure user identification in accordance with one embodiment of the present invention
- Fig. 4 is a block diagram of a medical device configured for secure user identification in accordance with one embodiment of the present invention.
- Fig. 5 is a diagrammatical representation of a medical device featuring an biometric reader in accordance with one embodiment of the present invention.
- Fig. 6 is a diagrammatical representation of a patient room containing a plurality of patients and a plurality of medical devices in accordance with one embodiment of the present invention. MODES FOR CARRYING OUT THE INVENTION
- the medical device should not react to simple complaints from patients or visitors. Medical devices often have loud alarms to alert caregivers that something requires their attention. If an alarm sounds, the patient or a visitor typically should not be allowed to tell the medical device to quiet the alarm. In addition, it may be desirable to prevent some medical personnel from performing all of the functions on a given medical device. For example, an orderly typically may not be able to command a medical device to alter a patient's treatment.
- a pulse oximeter may record trends in the patient's blood-oxygen saturation level, so that a caregiver can determine whether treatment is improving the patient's condition. Accordingly, it may be desirable to allow only certain personnel to access the patient's medical history.
- One or more of the embodiments set forth below may be directed towards one or more of the issues discussed above.
- the medical device 10 comprises a pulse oximeter.
- the medical device 10 may include a main unit 12 that houses hardware and/or software configured to calculate various physiological parameters.
- the main unit 12 may include a display 14 for displaying the calculated physiological parameters, such as oxygen saturation or pulse rate, to a caregiver or patient.
- the display 14 may be omitted from the main unit 12.
- the medical device 10 may also include a sensor 16 that may be connected to a body part (e.g., finger, forehead, toe, or earlobe) of a patient or a user.
- the sensor 16 may be configured to emit signals or waves into the patient's or user's tissue and detect these signals or waves after dispersion and/or reflection by the tissue.
- the sensor 16 may be configured to emit light from two or more light emitting diodes ("LEDs") into pulsatile tissue (e.g., finger, forehead, toe, or earlobe) and then detect the transmitted light with a light detector (e.g., a photodiode or photo-detector) after the light has passed through the pulsatile tissue.
- LEDs light emitting diodes
- the amount of transmitted light that passes through the tissue generally varies in accordance with a changing amount of blood constituent in the tissue and the related light absorption.
- the heart pumps an incremental amount of arterial blood into the pulsatile tissue, which then drains back through the venous system.
- the amount of light that passes through the blood-perfused tissue varies with the cardiac-induced cycling arterial blood volume. For example, when the cardiac cycle causes more light-absorbing blood to be present in the tissue, less light travels through the tissue to strike the sensor's photo-detector.
- These pulsatile signals allow the medical device 10 to measure signal continuation caused by the tissue's arterial blood, because light absorption from other tissues remains generally unchanged in the relevant time span.
- the senor 16 may take other suitable forms beside the form illustrated in Fig. 1.
- the sensor 16 may be configured to be clipped onto a finger or earlobe or may be configured to be secured with tape or another static mounting technique.
- the sensor 16 may be connected to the main unit 12 via a cable 18 and a connector 20.
- the medical device 10 may also include an integral microphone 22.
- the integral microphone 22 may be configured to receive voice commands from a caregiver or user that can be processed into commands for the medical device 10.
- Fig. 1 illustrates the integral microphone 22 as being located on a front facade of the main unit 12, it will be appreciated that in alternate embodiments, the integral microphone 22 may be located at another suitable location on or within the main unit 12.
- the medical device 10 may also include a speaker 23, which may be configured to broadcast alerts to a caregiver or user.
- Fig. 1 illustrates the speaker 23 as being located on a side facade of the main unit 12, it will be appreciated that in alternate embodiments, the speaker 23 may be located at another suitable location on or within the main unit 12.
- Fig. 2 another embodiment of the exemplary medical device 10 featuring an external microphone and speaker in accordance with one embodiment.
- the medical device 10 includes the main unit 12, the screen 14, the sensor 16, the cable 18, and the connector 20.
- the medical device 10 illustrated in Fig. 2 includes an audio connector 24 suitable for coupling a head set 26 to the main unit 12.
- the headset 26 may include one or more speakers 28 and an external microphone 30.
- the one or more external speakers 28 may be employed by the medical device 10 to broadcast suitable alerts to a caregiver or user.
- the external microphone 30 may be employed to receive voice commands for the medical device 10.
- the medical device 10 may be configured to execute voice commands from users, such as caregivers.
- many other unauthorized individuals e.g., patients, guests
- a patient, annoyed with an alert may try to use voice commands to silence an alert before the caregiver hears the alarm.
- a conversation around the medical device could be interpreted by the medical device 10 as a voice command.
- the medical device 10 could accidentally interpret the question "When will my ventilator be turned off," as command to turn off the patient's ventilator. This is clearly undesirable.
- the medical device 10 may be configured to prevent unauthorized or accidental voice commands from occurring.
- the technique 40 may include using one or more sounds and/or keywords to unlock and/or lock a voice interface on the medical device 10.
- voice interface is locked, authorized or accidental commands would be ignored, but when a caregiver unlocks medical device 10, it would be free to accept and execute voice commands.
- the technique 40 may be executed by the medical device 10.
- other suitable medical devices may be configured to execute the technique 40
- the technique 40 may begin by receiving an "unlock" keyword or sound, as indicated by block 42.
- the keyword may be a simple word, such as "unlock” or "start.”
- the keyword may be a more linguistically complex word, such as a nonsensical or foreign word to discourage unauthorized access to the medical device.
- the medical device 10 may recognize a keyword using the techniques for voice recognition set forth in commonly assigned U.S. Patent Application Serial No. 11/540,457 entitled SYSTEM AND METHOD FOR INTEGRATING VOICE WITH A MEDICAL DEVICE and filed on September 29, 2006, which is hereby incorporated by reference.
- the technique 40 may include unlocking a voice interface of the medical device 10 to voice commands, as set indicated in block 44.
- the medical device 10 may start to execute voice commands after receiving the unlock keyword.
- the technique 40 may continue by executing subsequent voice commands, as indicated by block 46.
- the technique 40 may continue to execute voice commands until a "lock" keyword is received, as indicated by block 48.
- the lock keyword may be the same as the unlock keyword or it may be different.
- the words "unlock” and "lock” may be used or a single foreign word may be used to both unlock and lock the medical device 10.
- the technique 40 may lock the medical device 10 to future voice commands until the unlock keyword is received again, as indicated by block 50.
- Fig. 4 is a block diagram of the medical device 10 configured for secure user identification in accordance with one embodiment.
- the medical device 10 may include a plurality of modules (blocks 60-70). These modules may be hardware, software, or some combination of hardware and software. Additionally, it will be appreciated that the modules shown in Fig. 4 are merely one example and other embodiments can be envisaged wherein the module functions are split up differently or wherein some modules are not included or other modules are included.
- the blocks 60-70 may be employed in a plurality of other suitable medical devices in addition to the medical device 10.
- the blocks 60-70 may be employed in respirators, ventilators, electroencephalogram ("EEG”) devices, medical cutting devices, and so-forth.
- EEG electroencephalogram
- the medical device 10 may include an audio signal receiver 60.
- the audio signal receiver 60 may include any suitable form of microphone or voice recording device, such as the integral microphone 22 (illustrated in Fig. 1) or the external microphone 30 (illustrated in Fig. 2).
- the audio signal receiver 60 may be configured to receive an audio signal (i.e., an acoustic wave) and to convert the audio signal into an electronic analog waveform.
- the audio signal receiver 60 may be configured to transmit the analog electrical wave to an audio sampling system 62.
- the audio sampling system 62 may be configured to sample the electronic analog waveform to create digital voice data.
- the audio sampling system 62 may be configured to sample the electronic analog waveform 16,000 times per second to create a digital waveform of pulse amplitudes.
- other suitable sampling techniques may be employed.
- An audio processing system 64 may be configured to receive the digital waveform and to convert the digital waveform into frequencies that can be recognized by an audio recognition system 66.
- the audio processing system 64 may be configured to perform a fast fourier transform on the incoming digital waveform to generate a plurality of frequencies. The audio processing system 64 may then transmit the plurality of frequencies to the audio recognition system 66.
- the audio recognition system 66 may be pre-populated or programmed with a plurality of frequency combinations that are associated with commands for the medical device 10. For example, frequency combinations associated with the audio command "turn off alarm" may be associated with a command for the medical device 10 to silence an alarm. As mentioned above, in one embodiment, the particular frequency combinations may be pre-programmed or pre-configured. However, in alternate embodiments, the frequency combinations may be programmed into the audio recognition system by another suitable system.
- the audio recognition system 66 may be configured to identify keywords to unlock and/or lock the voice interface of the medical device 10.
- the audio recognition system 66 may be configured to access a keyword database 68 and to compare spoken combinations of frequencies as keywords that unlock and/or lock the medical device 10.
- the audio recognition system 66 may be configured to instruct a medical device control system 70 to accept voice commands after an unlock keyword is spoken and/or to stop executing voice commands after a lock keyword is spoken.
- a medical device may employ one or more suitable forms of biometric recognition to unlock and/or lock itself.
- Fig 5 is a diagrammatical representation of a medical device 80 featuring a biometric reader 82 in accordance with one embodiment.
- the medical device 80 may include the main unit 12, the display 14, the sensor 16, the cable 18, the connector 20, the microphone 22, and the speaker 23, as outlined above.
- the medical device 80 may include the biometric reader 82, which may be configured to recognize a biometric signature of a user (e.g., fingerprint, eye shape, facial contours, and the like).
- the biometric reader 82 may be configured to unlock and/or lock the medical device 80 to voice commands based on the biometric signature.
- the biometric reader 82 may include a fingerprint scanner.
- the biometric reader 82 may comprise a retinal scanner configured to conditioned access based upon the shape of a user's eye.
- the biometric reader 82 may comprise a video camera configured to conditioned access based on one or more facial or body features.
- the biometric reader 82 may include an optical spectrometer.
- the biometric reader 82 may be configured to recognize a user voiceprint.
- the biometric reader 82 may be configured to be able to receive and recognize an authorized user by their voice.
- the medical device may be configured to unlock when it detects an authorized user voice and to lock in the absence of that voice. It will be appreciated, however, that the above-described examples of the biometric reader 82 are merely exemplary, and, as such, not intended to be exclusive. Accordingly, in other embodiments, other alternate suitable biometric readers 82 may be employed.
- the medical device 10 may be configured to execute voice commands from caregivers.
- modern patient rooms, operating rooms, doctors' offices, and the like include a plurality of different medical devices.
- a patient's room may have a pulse oximeter, a respirator, and a multi-parameter monitoring system.
- a single patient room may include multiple patients and, as such, multiple incarnations of each device, which are each configured to employ voice commands.
- having multiple medical devices that each accept voice commands may introduce some measure of confusion between the medical devices. For example, a voice command directed to an oximeter to "turn off could be executed by a ventilator with disastrous results.
- an unlock command to one pulse oximeter in a patient's room may be received by another pulse oximeter in the same room - allowing potentially unauthorized access to the other pulse oximeter.
- one or more of the embodiments set forth below may be directed toward identification and specification of voice commands for a particular medical device.
- Fig. 6 is a diagrammatical representation of a patient room 90 containing a plurality of patients 92a and 92b and a plurality of medical devices 94a and 94b in accordance with one embodiment.
- the medical devices 94a and 94b may include the main units 12a and 12b, as described above with regard to Figs. 1 and 2 as well as external screens 96a and 96b.
- the patient room 90 is illustrated as including one medical device 94a corresponding to the patient 92a, and one medical device 94b corresponding to the patient 92b, it will be appreciated that this configuration and correspondence is merely exemplary.
- the patients 92a and 92b may each be supported by a plurality of similar and/or different medical devices.
- the voice commands from a caregiver directed towards the medical device 94a may be received by the medical device 94b.
- the medical devices 94a and 94b may be configured to employ preparatory identification commands. More specifically, in one embodiment, the medical devices 94a and 94b may be configured to only execute those voice commands preceded by a medical device specific identifier. For example, in one embodiment, each of medical devices 94a and 94b may have a unique numerical designation.
- a caregiver may direct voice commands to the medical device 94a by prefacing voice commands with the number one and may direct voice commands to the medical device 94b by prefacing commands with the number two.
- the medical devices 94a and 94b may also include labels that visually designate their unique identifiers to the caregiver.
- the medical device 94a may have a large sticker with the number one and the medical device 94b have a large sticker with the number two.
- numerical identifiers are merely one suitable type of unique identifier.
- the medical devices 94a and 94b may be assigned other suitable numbers, letters, words, and so-forth.
- each of the medical devices 94a and 94b may be assigned a unique name to identify it.
- the medical devices 94a and 94b may include a sensor configured to receive a laser or other light beam.
- the caregiver may use a laser pointer to designate the intended medical device.
- medical devices 94a and 94b may be configured to receive voice commands and/or unlock (as described above) when the caregiver directs a laser beam at that particular medical device 94a and 94b. In this way, medical devices 94a and 94b will only execute those voice commands intended for them. While the invention may be susceptible to various modifications and alternative forms, specific embodiments have been shown by way of example in the drawings and have been described in detail herein.
Abstract
There is provided a system and method for secure voice identification and medical device interface. More specifically, in one embodiment, there is provided a medical device (10) comprising an audio recognition system (66) configured to receive an unlock keyword for the medical device (42) and to unlock the medical device to voice commands in response to the unlock keyword (44), and a control system (70) coupled to the audio recognition system (66) configured to execute one or more voice commands (46) on the medical device (10) after the medical device is unlocked.
Description
SYSTEMS AND METHODS FOR SECURE VOICE IDENTIFICATION AND MEDICAL DEVICE INTERFACE
1. Technical Field
The present invention relates generally to medical devices and, more particularly, to secure voice identification and interface in such medical devices.
2. Background Art
This section is intended to introduce the reader to various aspects of art that may be related to various aspects of the present invention, which are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present invention. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art.
In the field of medicine, doctors often desire to monitor certain physiological characteristics of their patients. Accordingly, a wide variety of medical devices have been developed for monitoring physiological characteristics. Such devices provide caregivers, such as doctors, nurses, and/or other healthcare personnel, with the information they need to provide the best possible healthcare for their patients. As a result, such monitoring devices have become an indispensable part of modern medicine.
For example, one technique for monitoring certain physiological characteristics of a patient is commonly referred to as pulse oximetry, and the devices built based upon pulse oximetry techniques are commonly referred to as pulse oximeters. Pulse oximetry may be used to measure various blood flow characteristics, such as the oxygen saturation of hemoglobin in arterial blood, the volume of individual blood pulsations supplying the tissue, and/or the rate of blood pulsations corresponding to each heartbeat of a patient.
Pulse oximeters and other types of medical devices are typically mounted on stands that are positioned around a patient's bed or around an operating room table. When a caregiver desires to command the medical device (e.g., program, configure, and so-forth) they manipulate controls or push buttons on the medical device itself. The medical device typically provides results or responses to commands on a liquid crystal display ("LCD") screen mounted in an externally visible position within the medical device.
This conventional configuration, however, has several disadvantages. First, as described above, this conventional configuration relies upon physical contact with the medical device to input commands (e.g., pushing a button, turning a knob, and the like). Such physical contact, however, raises several concerns. Among these concerns are that in making contact with the medical device, the caregiver may spread illness or disease from room to room. More specifically, a caregiver may accidentally deposit germs (e.g., bacteria, viruses, and so forth) on the medical device while manipulating the device's controls. These germs may then be spread to the patient when a subsequent caregiver
touches the medical device and then touches the patient. Moreover, if medical devices are moved from one patient room to another, germs transferred to the medical device via touch may be carried from one patient room to another. Even in operating rooms where medical devices are typically static, germs may be transferred onto a medical device during one surgery and subsequently transferred off the medical device during a later performed surgery.
Second, beyond contamination, medical devices that rely on physical contact for command input may clutter the caregiver's workspace. For example, because the medical device must be within an arm's length of the caregiver, the medical device may crowd the caregiver - potentially even restricting free movement of the caregiver. In addition, caregivers may have difficulty manipulating controls with gloved hands. For example, it may be difficult to grasp a knob or press a small button due to the added encumbrance of a glove.
Third, current trends in general medical device design focus on miniaturizing overall medical device size. However, as controls which rely on physical contact must be large enough for most, if not all, caregivers to manipulate with their hands, medical devices that employ these types of controls are limited in their possible miniaturization. For example, even if it were possible to produce a conventional oximeter that was the size of a postage stamp, it would be difficult to control this theoretical postage stamp-sized pulse oximeter with currently available techniques.
BRIEF DESCRIPTION OF DRAWINGS
Advantages of the invention may become apparent upon reading the following detailed description and upon reference to the drawings in which:
Fig. 1 is a diagrammatical representation of a medical device featuring an integral microphone in accordance with one embodiment of the present invention;
Fig. 2 is a diagrammatical representation of a medical device featuring an external microphone in accordance with one embodiment of the present invention;
Fig. 3 is a flow chart illustrating an exemplary technique for secure user identification in accordance with one embodiment of the present invention;
Fig. 4 is a block diagram of a medical device configured for secure user identification in accordance with one embodiment of the present invention;
Fig. 5 is a diagrammatical representation of a medical device featuring an biometric reader in accordance with one embodiment of the present invention; and
Fig. 6 is a diagrammatical representation of a patient room containing a plurality of patients and a plurality of medical devices in accordance with one embodiment of the present invention.
MODES FOR CARRYING OUT THE INVENTION
One or more specific embodiments of the present invention will be described below. In an effort to provide a concise description of these embodiments, not all features of an actual implementation are described in the specification. It should be appreciated that in the development of any such actual implementation, as in any engineering or design project, numerous implementation-specific decisions must be made to achieve the developers' specific goals, such as compliance with system-related and business-related constraints, which may vary from one implementation to another. Moreover, it should be appreciated that such a development effort might be complex and time consuming, but would nevertheless be a routine undertaking of design, fabrication, and manufacture for those of ordinary skill having the benefit of this disclosure.
For at least the reasons set forth above, an improved system or method for interacting with a medical monitoring device would be desirable. A possible solution for resolving one or more of the issues set forth above involves issuing voice commands to the medical device rather than physically manipulating controls. However, this solution raises new concerns.
First, the medical device should not react to simple complaints from patients or visitors. Medical devices often have loud alarms to alert caregivers that something requires their attention. If an alarm sounds, the patient or a visitor typically should not be allowed to tell the medical device to quiet the alarm. In addition, it may be desirable to
prevent some medical personnel from performing all of the functions on a given medical device. For example, an orderly typically may not be able to command a medical device to alter a patient's treatment.
Second, in medical devices that store patient information, security measures typically should ensure patient privacy. Some medical devices store historical data on the patient which the caregiver can reference for comparisons. For example, a pulse oximeter may record trends in the patient's blood-oxygen saturation level, so that a caregiver can determine whether treatment is improving the patient's condition. Accordingly, it may be desirable to allow only certain personnel to access the patient's medical history.
One or more of the embodiments set forth below may be directed towards one or more of the issues discussed above.
Turning initially to Fig. 1 , an exemplary medical device featuring an integral microphone in accordance with one embodiment is illustrated and generally designated by the reference numeral 10. For example, in the illustrated embodiment, the medical device 10 comprises a pulse oximeter. The medical device 10 may include a main unit 12 that houses hardware and/or software configured to calculate various physiological parameters. As illustrated, the main unit 12 may include a display 14 for displaying the calculated physiological parameters, such as oxygen saturation or pulse rate, to a caregiver or patient. In alternate embodiments, as described in further detail below, the display 14 may be omitted from the main unit 12.
The medical device 10 may also include a sensor 16 that may be connected to a body part (e.g., finger, forehead, toe, or earlobe) of a patient or a user. The sensor 16 may be configured to emit signals or waves into the patient's or user's tissue and detect these signals or waves after dispersion and/or reflection by the tissue. For example, the sensor 16 may be configured to emit light from two or more light emitting diodes ("LEDs") into pulsatile tissue (e.g., finger, forehead, toe, or earlobe) and then detect the transmitted light with a light detector (e.g., a photodiode or photo-detector) after the light has passed through the pulsatile tissue.
As those of ordinary skill in the art will appreciate, the amount of transmitted light that passes through the tissue generally varies in accordance with a changing amount of blood constituent in the tissue and the related light absorption. On a beat-by- beat basis, the heart pumps an incremental amount of arterial blood into the pulsatile tissue, which then drains back through the venous system. The amount of light that passes through the blood-perfused tissue varies with the cardiac-induced cycling arterial blood volume. For example, when the cardiac cycle causes more light-absorbing blood to be present in the tissue, less light travels through the tissue to strike the sensor's photo-detector. These pulsatile signals allow the medical device 10 to measure signal continuation caused by the tissue's arterial blood, because light absorption from other tissues remains generally unchanged in the relevant time span.
In alternate embodiments, the sensor 16 may take other suitable forms beside the form illustrated in Fig. 1. For example, the sensor 16 may be configured to be clipped onto a finger or earlobe or may be configured to be secured with tape or another static mounting technique. The sensor 16 may be connected to the main unit 12 via a cable 18 and a connector 20.
The medical device 10 may also include an integral microphone 22. As will be described further below, the integral microphone 22 may be configured to receive voice commands from a caregiver or user that can be processed into commands for the medical device 10. Although Fig. 1 illustrates the integral microphone 22 as being located on a front facade of the main unit 12, it will be appreciated that in alternate embodiments, the integral microphone 22 may be located at another suitable location on or within the main unit 12.
The medical device 10 may also include a speaker 23, which may be configured to broadcast alerts to a caregiver or user. Although Fig. 1 illustrates the speaker 23 as being located on a side facade of the main unit 12, it will be appreciated that in alternate embodiments, the speaker 23 may be located at another suitable location on or within the main unit 12.
Turning next to Fig. 2, another embodiment of the exemplary medical device 10 featuring an external microphone and speaker in accordance with one embodiment. For simplicity, like reference numerals have been used to designate those features
previously described in regard to Fig. 1. As illustrated in Fig. 2, the medical device 10 includes the main unit 12, the screen 14, the sensor 16, the cable 18, and the connector 20. However, in place of or in addition to the integral microphone 22, the medical device 10 illustrated in Fig. 2 includes an audio connector 24 suitable for coupling a head set 26 to the main unit 12.
As illustrated in Fig. 2, the headset 26 may include one or more speakers 28 and an external microphone 30. As will be described further below, the one or more external speakers 28 may be employed by the medical device 10 to broadcast suitable alerts to a caregiver or user. In addition, the external microphone 30 may be employed to receive voice commands for the medical device 10.
As will be described further below, the medical device 10 may be configured to execute voice commands from users, such as caregivers. However, as will be appreciated, many other unauthorized individuals (e.g., patients, guests) may also try to give voice commands to the medical device. For example, a patient, annoyed with an alert, may try to use voice commands to silence an alert before the caregiver hears the alarm. Alternatively, a conversation around the medical device could be interpreted by the medical device 10 as a voice command. For example, the medical device 10 could accidentally interpret the question "When will my ventilator be turned off," as command to turn off the patient's ventilator. This is clearly undesirable.
For at least these reasons, the medical device 10 may be configured to prevent unauthorized or accidental voice commands from occurring. For example, Fig. 3 is a flow chart illustrating an exemplary technique 40 for secure user identification in accordance with one embodiment. In particular, as described further below, the technique 40 may include using one or more sounds and/or keywords to unlock and/or lock a voice interface on the medical device 10. When voice interface is locked, authorized or accidental commands would be ignored, but when a caregiver unlocks medical device 10, it would be free to accept and execute voice commands. In one embodiment, the technique 40 may be executed by the medical device 10. However, in alternate embodiments, other suitable medical devices may be configured to execute the technique 40
The technique 40 may begin by receiving an "unlock" keyword or sound, as indicated by block 42. For example, in one embodiment, the keyword may be a simple word, such as "unlock" or "start." However, in alternate embodiments, the keyword may be a more linguistically complex word, such as a nonsensical or foreign word to discourage unauthorized access to the medical device. In one embodiment, the medical device 10 may recognize a keyword using the techniques for voice recognition set forth in commonly assigned U.S. Patent Application Serial No. 11/540,457 entitled SYSTEM AND METHOD FOR INTEGRATING VOICE WITH A MEDICAL DEVICE and filed on September 29, 2006, which is hereby incorporated by reference.
After receiving the unlock keyword, the technique 40 may include unlocking a voice interface of the medical device 10 to voice commands, as set indicated in block 44. For example, in one embodiment, the medical device 10 may start to execute voice commands after receiving the unlock keyword. After the medical device 10 has been unlocked, the technique 40 may continue by executing subsequent voice commands, as indicated by block 46.
As illustrated in Fig. 3, the technique 40 may continue to execute voice commands until a "lock" keyword is received, as indicated by block 48. In various embodiments, the lock keyword may be the same as the unlock keyword or it may be different. For example, the words "unlock" and "lock" may be used or a single foreign word may be used to both unlock and lock the medical device 10. After the lock keyword is received, the technique 40 may lock the medical device 10 to future voice commands until the unlock keyword is received again, as indicated by block 50.
As described above, the medical device 10 may be configured to unlock and/or lock its voice interface based on one or more keywords. Accordingly, Fig. 4 is a block diagram of the medical device 10 configured for secure user identification in accordance with one embodiment. For simplicity, like reference numerals have been used to designate those features previously described with regard to Figs. 1 and 2. As illustrated in Fig. 4, the medical device 10 may include a plurality of modules (blocks 60-70). These modules may be hardware, software, or some combination of hardware and software. Additionally, it will be appreciated that the modules shown in Fig. 4 are
merely one example and other embodiments can be envisaged wherein the module functions are split up differently or wherein some modules are not included or other modules are included. Moreover, it will be appreciated that the blocks 60-70 may be employed in a plurality of other suitable medical devices in addition to the medical device 10. For example, the blocks 60-70 may be employed in respirators, ventilators, electroencephalogram ("EEG") devices, medical cutting devices, and so-forth.
As illustrated in Fig. 4, the medical device 10 may include an audio signal receiver 60. The audio signal receiver 60 may include any suitable form of microphone or voice recording device, such as the integral microphone 22 (illustrated in Fig. 1) or the external microphone 30 (illustrated in Fig. 2). As those of ordinary skill in the art will appreciate, the audio signal receiver 60 may be configured to receive an audio signal (i.e., an acoustic wave) and to convert the audio signal into an electronic analog waveform.
The audio signal receiver 60 may be configured to transmit the analog electrical wave to an audio sampling system 62. The audio sampling system 62 may be configured to sample the electronic analog waveform to create digital voice data. For example, in one embodiment, the audio sampling system 62 may be configured to sample the electronic analog waveform 16,000 times per second to create a digital waveform of pulse amplitudes. In alternate embodiments, other suitable sampling techniques may be employed.
An audio processing system 64 may be configured to receive the digital waveform and to convert the digital waveform into frequencies that can be recognized by an audio recognition system 66. In one embodiment, the audio processing system 64 may be configured to perform a fast fourier transform on the incoming digital waveform to generate a plurality of frequencies. The audio processing system 64 may then transmit the plurality of frequencies to the audio recognition system 66.
The audio recognition system 66 may be pre-populated or programmed with a plurality of frequency combinations that are associated with commands for the medical device 10. For example, frequency combinations associated with the audio command "turn off alarm" may be associated with a command for the medical device 10 to silence an alarm. As mentioned above, in one embodiment, the particular frequency combinations may be pre-programmed or pre-configured. However, in alternate embodiments, the frequency combinations may be programmed into the audio recognition system by another suitable system.
In addition, besides recognizing a command for the oximeter 10, the audio recognition system 66 may be configured to identify keywords to unlock and/or lock the voice interface of the medical device 10. For example, the audio recognition system 66 may be configured to access a keyword database 68 and to compare spoken combinations of frequencies as keywords that unlock and/or lock the medical device 10. As such, the audio recognition system 66 may be configured to instruct a medical
device control system 70 to accept voice commands after an unlock keyword is spoken and/or to stop executing voice commands after a lock keyword is spoken.
In another embodiment, a medical device may employ one or more suitable forms of biometric recognition to unlock and/or lock itself. For example, Fig 5 is a diagrammatical representation of a medical device 80 featuring a biometric reader 82 in accordance with one embodiment. For simplicity, like reference numerals have been used to designate those features previously described in relation to Figs. 1 and 2. As illustrated by Fig. 5, the medical device 80 may include the main unit 12, the display 14, the sensor 16, the cable 18, the connector 20, the microphone 22, and the speaker 23, as outlined above.
In addition, the medical device 80 may include the biometric reader 82, which may be configured to recognize a biometric signature of a user (e.g., fingerprint, eye shape, facial contours, and the like). The biometric reader 82 may be configured to unlock and/or lock the medical device 80 to voice commands based on the biometric signature. In one embodiment, the biometric reader 82 may include a fingerprint scanner. In another embodiment, the biometric reader 82 may comprise a retinal scanner configured to conditioned access based upon the shape of a user's eye. In still other embodiments, the biometric reader 82 may comprise a video camera configured to conditioned access based on one or more facial or body features. In yet another embodiment, the biometric reader 82 may include an optical spectrometer.
In yet another embodiment, the biometric reader 82 may be configured to recognize a user voiceprint. For example, the biometric reader 82 may be configured to be able to receive and recognize an authorized user by their voice. Accordingly, in this embodiment, the medical device may be configured to unlock when it detects an authorized user voice and to lock in the absence of that voice. It will be appreciated, however, that the above-described examples of the biometric reader 82 are merely exemplary, and, as such, not intended to be exclusive. Accordingly, in other embodiments, other alternate suitable biometric readers 82 may be employed.
As described above, the medical device 10 may be configured to execute voice commands from caregivers. As will be appreciated, however, modern patient rooms, operating rooms, doctors' offices, and the like, include a plurality of different medical devices. For example, a patient's room may have a pulse oximeter, a respirator, and a multi-parameter monitoring system. Moreover, a single patient room may include multiple patients and, as such, multiple incarnations of each device, which are each configured to employ voice commands. Although such a configuration is advantageous from a cross-contamination standpoint, having multiple medical devices that each accept voice commands may introduce some measure of confusion between the medical devices. For example, a voice command directed to an oximeter to "turn off could be executed by a ventilator with disastrous results. Alternatively, an unlock command to one pulse oximeter in a patient's room may be received by another pulse oximeter in the same room - allowing potentially unauthorized access to the other pulse oximeter. For at least these reasons, one or more of the embodiments set forth below may be
directed toward identification and specification of voice commands for a particular medical device.
Fig. 6 is a diagrammatical representation of a patient room 90 containing a plurality of patients 92a and 92b and a plurality of medical devices 94a and 94b in accordance with one embodiment. As illustrated, the medical devices 94a and 94b may include the main units 12a and 12b, as described above with regard to Figs. 1 and 2 as well as external screens 96a and 96b. Although the patient room 90 is illustrated as including one medical device 94a corresponding to the patient 92a, and one medical device 94b corresponding to the patient 92b, it will be appreciated that this configuration and correspondence is merely exemplary. As such, in alternate embodiments, the patients 92a and 92b may each be supported by a plurality of similar and/or different medical devices.
As described above, the voice commands from a caregiver directed towards the medical device 94a, for example, may be received by the medical device 94b. As such, the medical devices 94a and 94b may be configured to employ preparatory identification commands. More specifically, in one embodiment, the medical devices 94a and 94b may be configured to only execute those voice commands preceded by a medical device specific identifier. For example, in one embodiment, each of medical devices 94a and 94b may have a unique numerical designation. In this embodiment, a caregiver may direct voice commands to the medical device 94a by prefacing voice
commands with the number one and may direct voice commands to the medical device 94b by prefacing commands with the number two.
In one embodiment, the medical devices 94a and 94b may also include labels that visually designate their unique identifiers to the caregiver. For example, the medical device 94a may have a large sticker with the number one and the medical device 94b have a large sticker with the number two. It will be appreciated, however, that numerical identifiers are merely one suitable type of unique identifier. As such, in alternate embodiments, the medical devices 94a and 94b may be assigned other suitable numbers, letters, words, and so-forth. For example, each of the medical devices 94a and 94b may be assigned a unique name to identify it.
In still other embodiments, other techniques for identifying the target medical device 10 for a voice command may be employed. For example, in one embodiment, the medical devices 94a and 94b may include a sensor configured to receive a laser or other light beam. In this embodiment, the caregiver may use a laser pointer to designate the intended medical device. For example, medical devices 94a and 94b may be configured to receive voice commands and/or unlock (as described above) when the caregiver directs a laser beam at that particular medical device 94a and 94b. In this way, medical devices 94a and 94b will only execute those voice commands intended for them.
While the invention may be susceptible to various modifications and alternative forms, specific embodiments have been shown by way of example in the drawings and have been described in detail herein. However, it should be understood that the invention is not intended to be limited to the particular forms disclosed. Rather, the invention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the invention as defined by the following appended claims. Indeed, as described above the present techniques may not only be applied to pulse oximeters, but also to a number of other suitable medical devices
Claims
1. A method comprising: receiving an unlock sound for a medical device; unlocking the medical device to voice commands in response to receiving the unlock sound; and executing one or more voice commands on the medical device after the medical device is unlocked.
2. The method, as set forth in claim 1 , comprising: receiving a lock sound for the medical device; and locking the medical device to subsequent voice commands.
3. The method, as set forth in claim 2, wherein the lock sound and the unlock sound are different words.
4. The method, as set forth in claim 1 , wherein receiving the unlock sound comprises receiving a non-english word.
5. The method, as set forth in claim 4, The method, as set forth in claim 1, wherein receiving the unlock sound comprises receiving a nonsensical word.
6. The method, as set forth in claim 1, wherein unlocking the medical device to voice commands comprises unlocking a pulse oximeter to voice commands.
7. A medical device comprising: an audio recognition system configured to receive an unlock keyword for the medical device; and to unlock the medical device to voice commands in response to the unlock keyword; and a control system coupled to the audio recognition system configured to execute one or more voice commands on the medical device after the medical device is unlocked.
8. The medical device, as set forth in claim 7, wherein the audio recognition is configured to receive a lock keyword for the medical device and wherein the control system is configured to lock the medical device in response to the lock keyword.
9. The medical device, as set forth in claim 7, wherein the medical device comprises a pulse oximeter.
10. A medical device comprising: a biometric reader configured to recognize a biometric signature of an authorized user; and a control system configured to execute voice commands in response to the biometric reader recognizing the authorized user.
1 1. The medical device, as set forth in claim 10, wherein the biometric reader comprises a fingerprint scanner.
12. The medical device, as set forth in claim 10, wherein the biometric reader comprises a retinal scanner.
13. The medical device, as set forth in claim 10, wherein the biometric reader comprises a camera.
14. The medical device, as set forth in claim 10, wherein the medical device comprises a pulse oximeter.
15. The medical device, as set forth in claim 10, wherein the biometric reader comprises an optical spectrometer.
16. A medical device configured to execute voice commands preceded by a unique identifier associated with the medical device, wherein the medical device is configured to ignore voice commands not preceded by the unique identifier.
17. The medical device, as set forth in claim 16, wherein the unique identifier comprises a numerical designation.
18. The medical device, as set forth in claim 17, comprising a label comprising the numerical designation associated with the medical device.
19. The medical device, as set forth in claim 16, wherein the medical device comprises a pulse oximeter.
20. A medical device comprising: a laser sensor configured to detect the presence of a laser beam on the laser sensor; and a control system coupled to the laser sensor and configured to execute voice commands after the laser sensor has detected the presence of the laser beam.
21. The medical device, as set forth in claim 20, wherein the control system is configured to only execute commands while the laser is detecting the presence of the laser beam.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/540,015 | 2006-09-29 | ||
US11/540,015 US20080082338A1 (en) | 2006-09-29 | 2006-09-29 | Systems and methods for secure voice identification and medical device interface |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2008042121A2 true WO2008042121A2 (en) | 2008-04-10 |
WO2008042121A3 WO2008042121A3 (en) | 2008-10-30 |
Family
ID=39033530
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2007/020380 WO2008042121A2 (en) | 2006-09-29 | 2007-09-19 | Systems and methods for secure voice identification and medical device interface |
Country Status (2)
Country | Link |
---|---|
US (1) | US20080082338A1 (en) |
WO (1) | WO2008042121A2 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2027844A1 (en) * | 2007-08-20 | 2009-02-25 | Hill-Rom Services, Inc. | Proximity activiation of voice operation of hospital bed |
US8544126B2 (en) | 2005-12-19 | 2013-10-01 | Stryker Corporation | Patient support with improved control |
US9020823B2 (en) | 2009-10-30 | 2015-04-28 | Continental Automotive Gmbh | Apparatus, system and method for voice dialogue activation and/or conduct |
US11246776B2 (en) | 2005-12-19 | 2022-02-15 | Stryker Corporation | Patient support with improved control |
Families Citing this family (177)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US7706896B2 (en) * | 2006-09-29 | 2010-04-27 | Nellcor Puritan Bennett Llc | User interface and identification in a medical device system and method |
US20080081956A1 (en) * | 2006-09-29 | 2008-04-03 | Jayesh Shah | System and method for integrating voice with a medical device |
US9591392B2 (en) * | 2006-11-06 | 2017-03-07 | Plantronics, Inc. | Headset-derived real-time presence and communication systems and methods |
US20080260169A1 (en) * | 2006-11-06 | 2008-10-23 | Plantronics, Inc. | Headset Derived Real Time Presence And Communication Systems And Methods |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10002189B2 (en) | 2007-12-20 | 2018-06-19 | Apple Inc. | Method and apparatus for searching using an active ontology |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US8255225B2 (en) * | 2008-08-07 | 2012-08-28 | Vocollect Healthcare Systems, Inc. | Voice assistant system |
US20100081891A1 (en) * | 2008-09-30 | 2010-04-01 | Nellcor Puritan Bennett Llc | System And Method For Displaying Detailed Information For A Data Point |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US9747417B2 (en) * | 2013-11-14 | 2017-08-29 | Mores, Inc. | Method and apparatus for enhanced personal care |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10706373B2 (en) | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
US20100328034A1 (en) * | 2009-06-30 | 2010-12-30 | Nellcor Puritan Bennett Llc | System and Method for Controlling One or Both of Sensor Functionality and Data Access Based on Biometrics Data |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
WO2011063106A1 (en) * | 2009-11-18 | 2011-05-26 | Nellcor Puritan Bennett Llc | Intelligent user interface for medical monitors |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US8442835B2 (en) | 2010-06-17 | 2013-05-14 | At&T Intellectual Property I, L.P. | Methods, systems, and products for measuring health |
US8666768B2 (en) | 2010-07-27 | 2014-03-04 | At&T Intellectual Property I, L. P. | Methods, systems, and products for measuring health |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10672399B2 (en) | 2011-06-03 | 2020-06-02 | Apple Inc. | Switching between text data and audio data based on a mapping |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
KR101889836B1 (en) * | 2012-02-24 | 2018-08-20 | 삼성전자주식회사 | Method and apparatus for cotrolling lock/unlock state of terminal through voice recognition |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
KR20130133629A (en) | 2012-05-29 | 2013-12-09 | 삼성전자주식회사 | Method and apparatus for executing voice command in electronic device |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
US10438591B1 (en) | 2012-10-30 | 2019-10-08 | Google Llc | Hotword-based speaker recognition |
EP2954514B1 (en) | 2013-02-07 | 2021-03-31 | Apple Inc. | Voice trigger for a digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10642574B2 (en) | 2013-03-14 | 2020-05-05 | Apple Inc. | Device, method, and graphical user interface for outputting captions |
US9977779B2 (en) | 2013-03-14 | 2018-05-22 | Apple Inc. | Automatic supplementation of word correction dictionaries |
US10572476B2 (en) | 2013-03-14 | 2020-02-25 | Apple Inc. | Refining a search based on schedule items |
US9733821B2 (en) | 2013-03-14 | 2017-08-15 | Apple Inc. | Voice control to diagnose inadvertent activation of accessibility features |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
AU2014233517B2 (en) | 2013-03-15 | 2017-05-25 | Apple Inc. | Training an at least partial voice command system |
CN105144133B (en) | 2013-03-15 | 2020-11-20 | 苹果公司 | Context-sensitive handling of interrupts |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
WO2014144395A2 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | User training by intelligent digital assistant |
WO2014144579A1 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
DE112014002747T5 (en) | 2013-06-09 | 2016-03-03 | Apple Inc. | Apparatus, method and graphical user interface for enabling conversation persistence over two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
KR101809808B1 (en) | 2013-06-13 | 2017-12-15 | 애플 인크. | System and method for emergency calls initiated by voice command |
AU2014306221B2 (en) | 2013-08-06 | 2017-04-06 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
AU2015266863B2 (en) | 2014-05-30 | 2018-03-15 | Apple Inc. | Multi-command single utterance input method |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9263042B1 (en) * | 2014-07-25 | 2016-02-16 | Google Inc. | Providing pre-computed hotword models |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US10152299B2 (en) | 2015-03-06 | 2018-12-11 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10127906B1 (en) | 2015-12-28 | 2018-11-13 | Amazon Technologies, Inc. | Naming devices via voice commands |
US10026401B1 (en) | 2015-12-28 | 2018-07-17 | Amazon Technologies, Inc. | Naming devices via voice commands |
US10185544B1 (en) | 2015-12-28 | 2019-01-22 | Amazon Technologies, Inc. | Naming devices via voice commands |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179309B1 (en) | 2016-06-09 | 2018-04-23 | Apple Inc | Intelligent automated assistant in a home environment |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
US10258295B2 (en) * | 2017-05-09 | 2019-04-16 | LifePod Solutions, Inc. | Voice controlled assistance for monitoring adverse events of a user and/or coordinating emergency actions such as caregiver communication |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | User interface for correcting recognition errors |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
DK201770427A1 (en) | 2017-05-12 | 2018-12-20 | Apple Inc. | Low-latency intelligent automated assistant |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
DK179549B1 (en) | 2017-05-16 | 2019-02-12 | Apple Inc. | Far-field extension for digital assistant services |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10283117B2 (en) * | 2017-06-19 | 2019-05-07 | Lenovo (Singapore) Pte. Ltd. | Systems and methods for identification of response cue at peripheral device |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
US11076039B2 (en) | 2018-06-03 | 2021-07-27 | Apple Inc. | Accelerated task performance |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030149342A1 (en) * | 2002-02-01 | 2003-08-07 | Hanover Barry Keith | System and method for wireless voice control of an interventional or diagnostic medical device |
US20050202843A1 (en) * | 2004-03-15 | 2005-09-15 | Fors Steven L. | Method and system for utilizing wireless voice technology within a radiology workflow |
WO2006057839A1 (en) * | 2004-11-23 | 2006-06-01 | Medical Digital Developers Llc | Medical media capture system and method |
Family Cites Families (76)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US1986543A (en) * | 1931-01-24 | 1935-01-01 | Ralph O Dulany | Method of canning sweet potatoes |
US1905356A (en) * | 1932-01-02 | 1933-04-25 | George E West | Cap for containers |
US3638640A (en) * | 1967-11-01 | 1972-02-01 | Robert F Shaw | Oximeter and method for in vivo determination of oxygen saturation in blood using three or more different wavelengths |
US4770179A (en) * | 1982-09-02 | 1988-09-13 | Nellcor Incorporated | Calibrated optical oximeter probe |
US4653498A (en) * | 1982-09-13 | 1987-03-31 | Nellcor Incorporated | Pulse oximeter monitor |
US4911167A (en) * | 1985-06-07 | 1990-03-27 | Nellcor Incorporated | Method and apparatus for detecting optical pulses |
US4936679A (en) * | 1985-11-12 | 1990-06-26 | Becton, Dickinson And Company | Optical fiber transducer driving and measuring circuit and method for using same |
SE451371B (en) * | 1986-03-27 | 1987-10-05 | Harry Kihlberg | SET AND DEVICE FOR VERTICAL MANUFACTURE OF CONSTRUCTION ELEMENTS IN S SANDWICH |
US4805623A (en) * | 1987-09-04 | 1989-02-21 | Vander Corporation | Spectrophotometric method for quantitatively determining the concentration of a dilute component in a light- or other radiation-scattering environment |
US5078136A (en) * | 1988-03-30 | 1992-01-07 | Nellcor Incorporated | Method and apparatus for calculating arterial oxygen saturation based plethysmographs including transients |
US5873821A (en) * | 1992-05-18 | 1999-02-23 | Non-Invasive Technology, Inc. | Lateralization spectrophotometer |
US5122974A (en) * | 1989-02-06 | 1992-06-16 | Nim, Inc. | Phase modulated spectrophotometry |
US5119815A (en) * | 1988-12-21 | 1992-06-09 | Nim, Incorporated | Apparatus for determining the concentration of a tissue pigment of known absorbance, in vivo, using the decay characteristics of scintered electromagnetic radiation |
US6708048B1 (en) * | 1989-02-06 | 2004-03-16 | Non-Invasive Technology, Inc. | Phase modulation spectrophotometric apparatus |
US6183414B1 (en) * | 1999-04-26 | 2001-02-06 | Michael S. Wysor | Technique for restoring plasticity to tissues of a male or female organ |
US6681128B2 (en) * | 1990-10-06 | 2004-01-20 | Hema Metrics, Inc. | System for noninvasive hematocrit monitoring |
US6266546B1 (en) * | 1990-10-06 | 2001-07-24 | In-Line Diagnostics Corporation | System for noninvasive hematocrit monitoring |
DE69229994T2 (en) * | 1991-03-07 | 2000-04-27 | Masimo Corp | DEVICE AND METHOD FOR SIGNAL PROCESSING |
US6549795B1 (en) * | 1991-05-16 | 2003-04-15 | Non-Invasive Technology, Inc. | Spectrophotometer for tissue examination |
US5351685A (en) * | 1991-08-05 | 1994-10-04 | Nellcor Incorporated | Condensed oximeter system with noise reduction software |
EP0548012B1 (en) * | 1991-12-16 | 1997-09-03 | Novartis AG | Endoplastic reticulum-located recombinant dibasic endoprotease and uses thereof |
US5385143A (en) * | 1992-02-06 | 1995-01-31 | Nihon Kohden Corporation | Apparatus for measuring predetermined data of living tissue |
US5297548A (en) * | 1992-02-07 | 1994-03-29 | Ohmeda Inc. | Arterial blood monitoring probe |
US6785568B2 (en) * | 1992-05-18 | 2004-08-31 | Non-Invasive Technology Inc. | Transcranial examination of the brain |
US5355880A (en) * | 1992-07-06 | 1994-10-18 | Sandia Corporation | Reliable noninvasive measurement of blood gases |
US5645059A (en) * | 1993-12-17 | 1997-07-08 | Nellcor Incorporated | Medical sensor with modulated encoding scheme |
US5758644A (en) * | 1995-06-07 | 1998-06-02 | Masimo Corporation | Manual and automatic probe calibration |
US5645060A (en) * | 1995-06-14 | 1997-07-08 | Nellcor Puritan Bennett Incorporated | Method and apparatus for removing artifact and noise from pulse oximetry |
US6163715A (en) * | 1996-07-17 | 2000-12-19 | Criticare Systems, Inc. | Direct to digital oximeter and method for calculating oxygenation levels |
US6544193B2 (en) * | 1996-09-04 | 2003-04-08 | Marcio Marc Abreu | Noninvasive measurement of chemical substances |
JPH10127651A (en) * | 1996-09-04 | 1998-05-19 | Olympus Optical Co Ltd | Treatment tool for operation |
US6009830A (en) * | 1997-11-21 | 2000-01-04 | Applied Materials Inc. | Independent gas feeds in a plasma reactor |
DE69940053D1 (en) * | 1998-02-05 | 2009-01-22 | Hema Metrics Inc | METHOD AND DEVICE FOR NON-INVASIVE OBSERVATION OF BLOOD COMPONENTS |
US6662030B2 (en) * | 1998-05-18 | 2003-12-09 | Abbott Laboratories | Non-invasive sensor having controllable temperature feature |
JP2000083933A (en) * | 1998-07-17 | 2000-03-28 | Nippon Koden Corp | Instrument for measuring concentration of light absorptive material in vital tissue |
EP1104254A2 (en) * | 1998-08-13 | 2001-06-06 | Whitland Research Limited | Optical device |
US6199550B1 (en) * | 1998-08-14 | 2001-03-13 | Bioasyst, L.L.C. | Integrated physiologic sensor system |
US6064898A (en) * | 1998-09-21 | 2000-05-16 | Essential Medical Devices | Non-invasive blood component analyzer |
US6684090B2 (en) * | 1999-01-07 | 2004-01-27 | Masimo Corporation | Pulse oximetry data confidence indicator |
US6438399B1 (en) * | 1999-02-16 | 2002-08-20 | The Children's Hospital Of Philadelphia | Multi-wavelength frequency domain near-infrared cerebral oximeter |
US6675029B2 (en) * | 1999-07-22 | 2004-01-06 | Sensys Medical, Inc. | Apparatus and method for quantification of tissue hydration using diffuse reflectance spectroscopy |
US6515273B2 (en) * | 1999-08-26 | 2003-02-04 | Masimo Corporation | System for indicating the expiration of the useful operating life of a pulse oximetry sensor |
US7904139B2 (en) * | 1999-08-26 | 2011-03-08 | Non-Invasive Technology Inc. | Optical examination of biological tissue using non-contact irradiation and detection |
US6708049B1 (en) * | 1999-09-28 | 2004-03-16 | Nellcor Puritan Bennett Incorporated | Sensor with signature of data relating to sensor |
US6524239B1 (en) * | 1999-11-05 | 2003-02-25 | Wcr Company | Apparatus for non-instrusively measuring health parameters of a subject and method of use thereof |
JP2001149349A (en) * | 1999-11-26 | 2001-06-05 | Nippon Koden Corp | Sensor for living body |
US6622095B2 (en) * | 1999-11-30 | 2003-09-16 | Nihon Kohden Corporation | Apparatus for determining concentrations of hemoglobins |
US6415236B2 (en) * | 1999-11-30 | 2002-07-02 | Nihon Kohden Corporation | Apparatus for determining concentrations of hemoglobins |
WO2001045553A1 (en) * | 1999-12-22 | 2001-06-28 | Orsense Ltd. | A method of optical measurements for determining various parameters of the patient's blood |
US6594513B1 (en) * | 2000-01-12 | 2003-07-15 | Paul D. Jobsis | Method and apparatus for determining oxygen saturation of blood in body organs |
AU2001251654B2 (en) * | 2000-04-17 | 2005-03-03 | Nellcor Puritan Bennett Incorporated | Pulse oximeter sensor with piece-wise function |
US6591123B2 (en) * | 2000-08-31 | 2003-07-08 | Mallinckrodt Inc. | Oximeter sensor with digital memory recording sensor data |
US6606510B2 (en) * | 2000-08-31 | 2003-08-12 | Mallinckrodt Inc. | Oximeter sensor with digital memory encoding patient data |
IL138683A0 (en) * | 2000-09-25 | 2001-10-31 | Vital Medical Ltd | Apparatus and method for monitoring tissue vitality parameters |
IL138884A (en) * | 2000-10-05 | 2006-07-05 | Conmed Corp | Pulse oximeter and a method of its operation |
WO2002050817A1 (en) * | 2000-12-20 | 2002-06-27 | Koninklijke Philips Electronics N.V. | Speechdriven setting of a language of interaction |
US6591122B2 (en) * | 2001-03-16 | 2003-07-08 | Nellcor Puritan Bennett Incorporated | Device and method for monitoring body fluid and electrolyte disorders |
US6606509B2 (en) * | 2001-03-16 | 2003-08-12 | Nellcor Puritan Bennett Incorporated | Method and apparatus for improving the accuracy of noninvasive hematocrit measurements |
IL145445A (en) * | 2001-09-13 | 2006-12-31 | Conmed Corp | Signal processing method and device for signal-to-noise improvement |
US20030065626A1 (en) * | 2001-09-28 | 2003-04-03 | Allen Karl H. | User verification for conducting health-related transactions |
US6748254B2 (en) * | 2001-10-12 | 2004-06-08 | Nellcor Puritan Bennett Incorporated | Stacked adhesive optical sensor |
JP4031438B2 (en) * | 2002-02-14 | 2008-01-09 | 俊徳 加藤 | Biological function diagnostic device |
WO2003071939A1 (en) * | 2002-02-22 | 2003-09-04 | Masimo Corporation | Active pulse spectraphotometry |
DE10213692B4 (en) * | 2002-03-27 | 2013-05-23 | Weinmann Diagnostics Gmbh & Co. Kg | Method for controlling a device and device for measuring ingredients in the blood |
US6690958B1 (en) * | 2002-05-07 | 2004-02-10 | Nostix Llc | Ultrasound-guided near infrared spectrophotometer |
US6711425B1 (en) * | 2002-05-28 | 2004-03-23 | Ob Scientific, Inc. | Pulse oximeter with calibration stabilization |
US6693812B1 (en) * | 2002-08-12 | 2004-02-17 | Tyco Electronics Power Systems Inc. | Bias supply selection circuit, method of operation thereof and power supply employing the same |
JP3905005B2 (en) * | 2002-09-18 | 2007-04-18 | 富士通株式会社 | Portable device and semiconductor integrated circuit device |
US7027849B2 (en) * | 2002-11-22 | 2006-04-11 | Masimo Laboratories, Inc. | Blood parameter measurement system |
KR100571811B1 (en) * | 2003-05-09 | 2006-04-17 | 삼성전자주식회사 | Ear type measurement apparatus for bio signal |
US7017777B1 (en) * | 2003-08-15 | 2006-03-28 | Dixon Carolyn S | Automated vending machine |
US7551950B2 (en) * | 2004-06-29 | 2009-06-23 | O2 Medtech, Inc,. | Optical apparatus and method of use for non-invasive tomographic scan of biological tissues |
US7343186B2 (en) * | 2004-07-07 | 2008-03-11 | Masimo Laboratories, Inc. | Multi-wavelength physiological monitor |
EP1781162A1 (en) * | 2004-07-09 | 2007-05-09 | Tadiran Spectralink Ltd. | Wearable device, system and method for measuring vital parameters |
US7976472B2 (en) * | 2004-09-07 | 2011-07-12 | Masimo Corporation | Noninvasive hypovolemia monitor |
JP4503401B2 (en) * | 2004-09-08 | 2010-07-14 | 株式会社荏原製作所 | Method for forming metal film and method for forming wiring |
-
2006
- 2006-09-29 US US11/540,015 patent/US20080082338A1/en not_active Abandoned
-
2007
- 2007-09-19 WO PCT/US2007/020380 patent/WO2008042121A2/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030149342A1 (en) * | 2002-02-01 | 2003-08-07 | Hanover Barry Keith | System and method for wireless voice control of an interventional or diagnostic medical device |
US20050202843A1 (en) * | 2004-03-15 | 2005-09-15 | Fors Steven L. | Method and system for utilizing wireless voice technology within a radiology workflow |
WO2006057839A1 (en) * | 2004-11-23 | 2006-06-01 | Medical Digital Developers Llc | Medical media capture system and method |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10052249B2 (en) | 2004-10-29 | 2018-08-21 | Stryker Corporation | Patient support with improved control |
US11382813B2 (en) | 2004-10-29 | 2022-07-12 | Stryker Corporation | Patient support with improved control |
US8544126B2 (en) | 2005-12-19 | 2013-10-01 | Stryker Corporation | Patient support with improved control |
US9038217B2 (en) | 2005-12-19 | 2015-05-26 | Stryker Corporation | Patient support with improved control |
US11246776B2 (en) | 2005-12-19 | 2022-02-15 | Stryker Corporation | Patient support with improved control |
EP2027844A1 (en) * | 2007-08-20 | 2009-02-25 | Hill-Rom Services, Inc. | Proximity activiation of voice operation of hospital bed |
US9020823B2 (en) | 2009-10-30 | 2015-04-28 | Continental Automotive Gmbh | Apparatus, system and method for voice dialogue activation and/or conduct |
Also Published As
Publication number | Publication date |
---|---|
WO2008042121A3 (en) | 2008-10-30 |
US20080082338A1 (en) | 2008-04-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20080082338A1 (en) | Systems and methods for secure voice identification and medical device interface | |
US7925511B2 (en) | System and method for secure voice identification in a medical device | |
US8160683B2 (en) | System and method for integrating voice with a medical device | |
US11812229B2 (en) | Patient monitor alarm speaker analyzer | |
US7698002B2 (en) | Systems and methods for user interface and identification in a medical device | |
US7706896B2 (en) | User interface and identification in a medical device system and method | |
US11272883B2 (en) | Physiological sensor | |
US9380982B2 (en) | Adaptive alarm system and method | |
US20080097176A1 (en) | User interface and identification in a medical device systems and methods | |
US20080097177A1 (en) | System and method for user interface and identification in a medical device | |
US8968193B2 (en) | System and method for enabling a research mode on physiological monitors | |
US20180192953A1 (en) | Nose sensor | |
WO2008042219A2 (en) | User interface and identification in a medical device systems and methods | |
US11355232B2 (en) | Automated personnel identification and location, and automated procedure monitoring | |
CN110522413A (en) | A kind of life sign monitor system | |
CN111755089A (en) | Safety checking system for ophthalmic surgery | |
KR20160107885A (en) | Door lock system that can open with thoughts | |
US20150150518A1 (en) | Multimedia terminal with vital parameters monitoring | |
WO2022044129A1 (en) | Pulmonary sound analysis system | |
US20240020092A1 (en) | Contactless control of physiological monitors | |
WO2022044126A1 (en) | Lung sound analysis system | |
WO2022044128A1 (en) | Lung sound analysis system | |
WO2022044127A1 (en) | Lung sound analysis system | |
JP2017170009A (en) | Biological information measurement device and control method thereof | |
PT104038B (en) | PHYSIOLOGICAL PARAMETERS METER BASED ON AURICULAR TYPE SENSORS, DISTRIBUTED BLUETOOTH ARCHITECTURE AND MOBILE PHONES |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: COMMUNICATION NOT DELIVERED. NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112 EPC (EPO FORM 1205A DATED 19.06.2009) |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07838566 Country of ref document: EP Kind code of ref document: A2 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 07838566 Country of ref document: EP Kind code of ref document: A2 |