US20080074387A1 - Method and Apparatus for Voice-Controlled Graphical User Interface Pointing Device - Google Patents
Method and Apparatus for Voice-Controlled Graphical User Interface Pointing Device Download PDFInfo
- Publication number
- US20080074387A1 US20080074387A1 US11/948,263 US94826307A US2008074387A1 US 20080074387 A1 US20080074387 A1 US 20080074387A1 US 94826307 A US94826307 A US 94826307A US 2008074387 A1 US2008074387 A1 US 2008074387A1
- Authority
- US
- United States
- Prior art keywords
- user
- circuit board
- printed circuit
- signal
- tracking
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
Definitions
- the present invention generally relates to the field of computers, and more particularly relates to a pointing device for a computer.
- a pointing device for use with a computer is well known.
- Existing pointing devices such as a mouse, combine tracking and selecting functions into one device. Tracking involves physically moving the mouse to control the motion of an on-screen pointer or other icon. Physical movement of the mouse is typically accomplished by moving the wrist and palm of the hand gripping the mouse. Once tracking has moved the pointer, an operation may be selected for execution by the computer, typically by depressing a button on the mouse.
- a computer user relies significantly on his or her hand (particularly on the wrist, palm and fingers) to use a mouse for executing computer operations.
- RMI repetitive motion injuries
- CTS Carpal Tunnel Syndrome
- One common solution for relieving strain on a computer user's hands and wrists is to enable the computer system to be at least partially operated by the user's voice commands.
- voice recognition software allows a user to perform word processing operations by dictation rather than by typing.
- a significant drawback to conventional voice operated software applications is that they are susceptible to receiving background noise as well as the user's voice commands, distorting the audible input into the system. This leads to inaccuracy in the performance of the user's commands, and the mistakes can be very time consuming to fix.
- the present invention overcomes the above-described limitations by providing a pointing device for use in computing devices comprising a printed circuit board, a tracking device adapted to generate a tracking signal in response to a user vocal input and relay the tracking signal to the printed circuit board, and a selection device adapted to generate a selection signal in response to a user manipulation and relay the selection signal to the printed circuit board.
- FIG. 1 depicts a diagram graphically illustrating components of a pointing device system according to one embodiment of the invention
- FIG. 2 depicts a flow diagram of a method for acquiring user input via a pointing device according to one embodiment of the invention
- FIG. 3 depicts a flow diagram of a method of indexing user input according to one embodiment of the invention
- FIG. 4A depicts a hash table illustrating an indexing database according to one embodiment of the invention
- FIG. 4B depicts a flow diagram of a method for updating a hash table such as that illustrated in FIG. 4A ;
- FIG. 5 depicts a flow diagram of a processing sequence by which computer applications are launched according to one embodiment of the invention.
- FIG. 6 depicts a graphical representation of a pipeline sequence for acquiring and indexing input and launching applications according to one embodiment of the invention
- the present invention provides an improved pointing device for assisting computer users having poor or limited hand motor skill control or eyesight and for reducing the occurrence of repetitive motion injuries in physically unimpaired users.
- the present invention achieves this aim allows a user to manipulate an on-screen pointer and launch system applications by means of vocal commands. Thus both tracking and selecting can be achieved with little or no need to engage other parts of the body such as the hands and wrists.
- embodiments of the invention provide an improved system for acquiring vocal/audible input from a computer user, so that erroneous inputs such as background noise and/or vocal irregularities of the user are filtered out to optimize system accuracy.
- FIG. 1 The basic operation of one embodiment of a pointing device system according the present invention is illustrated in FIG. 1 .
- Operation of the pointing device system 100 is accomplished at least in part by the vocal commands of a computer user.
- the pointing device system 100 includes at least one vocal cord movement (VCM) sensor 102 , a voice box converter (VBC) 104 , and a keyword database 106 .
- the pointing device system 100 may determine the position of the icon corresponding to application matching keyword.
- the pointing device system 100 generates an output to the processor 108 , which executes the indicated commands or pointer or icon movements. Once the pointer or icon has been moved to the desired on-screen location, the user generates a signal to launch (or “select”) the chosen application.
- the system 100 optionally generates an audible output to indicate the new pointer position or open application, so that, for example, a blind user may use the pointing device system 100 effectively.
- FIG. 2 depicts a flow diagram of one embodiment of a method 200 for acquiring user input via a pointing device according to one embodiment of the invention.
- At least one VCM sensor 102 is positioned near a user's voice box (larynx).
- the user starts acquisition of vocal command data by activating a start delimiter 202 .
- Activation of the start delimiter 202 starts acquisition/recording of vocal cord movements.
- the user provides audible input 204 to the VCM sensor 102 by speaking into a converter system the keyword(s) corresponding to the movement or application the user wishes to execute.
- the sensor 102 discerns movement of the user's vocal cords and records these movements.
- the user activates a stop delimiter 206 that causes the VCM sensor 102 to cease recording vocal cord movements.
- the start and stop delimiters 202 , 206 may comprise one delimiter device that is activated once to commence and once more to cease recording.
- the recorded vocal cord movements are then sent as input to the VBC 104 .
- VBC that may be advantageously adapted for use with the present invention is commercially available from Lucent Technologies Inc. of Murray Hill, N.J.
- the VBC 104 converts the input from the VCM sensor(s) 102 into text output, and sends this text output to the keyword database 106 for further processing and indexing.
- the VBC output is also used by software to train a personal user dictionary for alphabets, words and mouse commands, so that vocal input from a particular user is easily and accurately converted to output that the processor 108 can recognize and use to execute commands.
- Such software is well known in the art.
- the keyword database 106 is coupled to the VBC 104 and further includes a desktop icon database 350 and a menu search tree 360 (see FIG. 3 ).
- the keyword database 106 indexes text output from the VBC 104 to determine the position of an on-screen icon or operation corresponding to the application that matches the keyword (user vocal input) processed by the VCM sensor/VBC operations.
- FIG. 3 depicts a flow diagram of a method 300 of indexing user input according to one embodiment of the invention.
- a spoken keyword corresponds either to a desktop icon or to a menu.
- the keyword database 106 asks whether the spoken keyword corresponds to a desktop icon ( 302 ), a submenu ( 304 ), or the start menu ( 306 ), or whether the vocal input does not match any keyword at all ( 308 ).
- the appropriate database 350 or search tree 360 is searched in steps 310 - 314 for an exact match to an icon or menu.
- FIG. 4A depicts a hash table 400 illustrating an indexing database according to one embodiment of the invention.
- the desktop icon database 350 is a hash table 400 that maps icon names or keywords 402 to specific positions (X, Y) 404 within the desktop area. For example, if the keyword 402 corresponds to a desktop icon, a search in the desktop icon database 350 will determine the actual position 404 of the indicated icon in the desktop area, and the on-screen pointer will be moved to that position on the desktop screen (step 316 in FIG. 3 ).
- the generation of the process of the hash table 400 is illustrated in FIG. 4B and will be discussed in further detail herein.
- the menu search tree 360 operates in a manner similar to the desktop icon database 350 . However, the menu search tree 360 maps names and keywords of submenus within the start menu (for example, the Accessories or Multimedia submenus in Windows), rather than names and locations of desktop icons.
- the on-screen pointer is then positioned on the appropriate submenu (step 318 in FIG. 3 ). If the user does not know the keyword for the desired submenu, the “Start” keyword will retrieve the root node of the menu search tree 360 (step 314 ) and position the pointer over the Start menu (step 320 ), which may be searched for the appropriate submenu.
- a signal is generated in step 308 to indicate to the user that the input is not valid.
- the signal is an audible signal, such as a beep.
- the user Once the pointer has been moved to the desired on-screen location, the user generates a signal to launch the chosen application.
- the launch signal is indicated by depressing a button on a conventional mouse.
- depression of the left mouse button on a conventional two- or three-button mouse optionally indicates selection of an application.
- alternate means of generating selection input to the mouse PCB may be used.
- FIG. 5 depicts a flow diagram of a processing sequence 500 by which a processor 108 launches computer applications according to one embodiment of the invention.
- the processing sequence 500 operates on a series of timed intervals during which the user must activate a launch signal or allow a default operation to be executed. For example, once the on-screen pointer has been moved to an indicated location in step 502 , (based on the database keyword matches described herein), a first timed interval begins to run. If the user wishes to launch the application on which the pointer is positioned, he signals so, for example by depressing a mouse button as described, before the first timeout in step 504 . A selection signal before the first timeout prompts the processor 108 to launch the chosen application (step 350 ).
- the processor 108 determines in step 506 whether the pointer is positioned on a desktop icon ( 506 a ) or a submenu ( 506 b ). If the pointer is positioned on a submenu ( 506 b ), that application is launched ( 350 ). If the pointer is positioned on a desktop icon ( 506 a ), the processor 108 moves the pointer, in step 508 , to a text area associated with the icon (located, for example, below the icon), and a second timed interval begins to run.
- the processor 108 receives a signal from the user before the second timeout (step 510 ), the user is enabled to update or alter the text in the text area in step 512 , and a database entry for the icon is updated accordingly in step 514 .
- the process 450 by which the icon database hash table 400 is updated is illustrated in FIG. 4B . If no signal is received from the user, the application is launched ( 350 ). In this manner, if no signal or signals are received from the user at any point in the processing sequence 500 , the default operations of the processing sequence 500 are programmed to execute the probable intentions of the user in pointing to the indicated icon or submenu.
- FIG. 6 depicts a graphical representation of a pipeline sequence for acquiring and indexing input and launching applications according to one embodiment of the invention.
- the three main stages 200 , 300 , 500 in the pointing device system processing sequence are performed concurrently in a pipeline system. That is, the stages 200 , 300 , 500 overlap so that after three cycles n 0 , n i , n i+1 of vocal command acquisition 200 ( FIG. 2 ), database indexing 300 ( FIG. 3 ), and application launch 500 ( FIG. 5 ), the stages 200 , 300 , 500 operate and process input concurrently at time t. Therefore, if a subsequent vocal command is acquired 200 immediately after a previous acquired command has been processed (i.e., indexed 300 ), the execution speed of the system 100 is approximately three times faster than a conventional sequential processing system.
- the launched application may be any application capable of being operated by the computer, such as a word processing program, a video game or a web browser.
- the launched application is a web browser such as Microsoft Internet Explorer.
- the keyword database 106 stores keywords for all hyperlinks, buttons and menus in all active (launched) applications, as well as the respective positions (X, Y) of each hyperlink, button and menu within the application.
- the keyword database 106 stores keywords and positions for all similar executable icons in other applications (i.e., word processing programs, video games, etc.).
- the method of tracking (positioning) and selecting (launching) with the pointer to launch a particular action within an application is substantially similar to the method illustrated in FIGS. 1-5 to launch the application.
- a user wishes to position the pointer on a particular hyperlink or button, he starts acquisition of vocal command data 200 by activating the start delimiter 202 .
- the user speaks into the converter system (i.e., VCM sensor 102 ) the keyword(s) corresponding to the hyperlink or button the user wishes to launch.
- the user activates the stop delimiter 206 .
- the acquired audible input is then converted into text form by the VBC 104 , which launches the corresponding hyperlink or button by the method discussed with reference to FIG. 5 .
- a user knows the keyword for a particular submenu (for example, the Save command within the File menu) he wishes to launch or explore, he can launch it by the same method by which hyperlinks and buttons are launched (i.e., speak the keyword—“Save”—in between activating the delimiters 202 , 206 ). If the user does not know the keyword, he may speak the keyword for the whole menu (“File”). Exploration of the menu is then automatic; the pointer is positioned upon a submenu for brief interval of time (such as in step 504 of FIG. 5 ), during which the user may signal (for example, by clicking a button) to launch the submenu or take no action. Taking no action will cause the processor to position the pointer upon the next submenu and repeat the process of allowing a brief interval for user input.
- the keyword “Active” is used to distinguish between a launched application and a new (but as yet unlaunched) application. If multiple applications of the same type are already launched, the keyword prompts the processor 108 to move the on-screen pointer from one application to the next. The pointer remains on each application for a brief interval (again similar to step 504 in FIG. 5 ) to allow the user to signal acceptance of an application. If no signal is given, the pointer is moved to the next application upon timeout.
- a graphical user interface pointing device system is provided that substantially reduces the occurrence of repetitive motion injuries caused by conventional pointing devices.
- the system is suitable for use with standard desktop and laptop computers, as well as next-generation embedded computing devices like hand held devices and web browsing devices that operate through television sets.
- next-generation embedded computing devices like hand held devices and web browsing devices that operate through television sets.
- the inaccuracies associated with conventional voice-operated software are substantially reduced by the delimiting functions of the present invention, which filter background noise and vocal irregularities in the system input.
- the present invention may be particularly well-suited for use by users who are deaf, ill, or have poor or limited hand motor skill control, as well as physically unimpaired users who may benefit from a hand-free pointing device system (such as those dictating in public places or while otherwise occupied).
Abstract
Description
- This application is a divisional of commonly owned, pending, U.S. patent application Ser. No. 10/638,183, filed Aug. 8, 2003, entitled METHOD AND APPARATUS FOR VOICE-CONTROLLED GRAPHICAL USER INTERFACE POINTING DEVICE, in the names of Narayan L. Gehlot and Victor B. Lawrence, which application is incorporated herein by reference as if set forth in its entirety.
- The present invention generally relates to the field of computers, and more particularly relates to a pointing device for a computer.
- A pointing device for use with a computer is well known. Existing pointing devices, such as a mouse, combine tracking and selecting functions into one device. Tracking involves physically moving the mouse to control the motion of an on-screen pointer or other icon. Physical movement of the mouse is typically accomplished by moving the wrist and palm of the hand gripping the mouse. Once tracking has moved the pointer, an operation may be selected for execution by the computer, typically by depressing a button on the mouse.
- A computer user relies significantly on his or her hand (particularly on the wrist, palm and fingers) to use a mouse for executing computer operations. Unfortunately, the use of a computer mouse over extended periods of time has been shown to be a leading cause of many different types of repetitive motion injuries (RMI) to body parts including the wrists, fingers, shoulders, and elbows, e.g., Carpal Tunnel Syndrome (CTS). Individuals in many fields of work rely on computers in their daily work and are thus forced to use a mouse quite extensively. Early injuries to children may even be incurable, rendering the injured child permanently disabled.
- One common solution for relieving strain on a computer user's hands and wrists is to enable the computer system to be at least partially operated by the user's voice commands. For example, commercially available voice recognition software allows a user to perform word processing operations by dictation rather than by typing. However, a significant drawback to conventional voice operated software applications is that they are susceptible to receiving background noise as well as the user's voice commands, distorting the audible input into the system. This leads to inaccuracy in the performance of the user's commands, and the mistakes can be very time consuming to fix.
- The present invention overcomes the above-described limitations by providing a pointing device for use in computing devices comprising a printed circuit board, a tracking device adapted to generate a tracking signal in response to a user vocal input and relay the tracking signal to the printed circuit board, and a selection device adapted to generate a selection signal in response to a user manipulation and relay the selection signal to the printed circuit board.
- So that the manner in which the above recited embodiments of the invention are attained and can be understood in detail, a more particular description of the invention, briefly summarized above, may be had by reference to the embodiments thereof which are illustrated in the appended drawings. It is to be noted, however, that the appended drawings illustrate only typical embodiments of this invention and are therefore not to be considered limiting of its scope, for the invention may admit to other equally effective embodiments.
-
FIG. 1 depicts a diagram graphically illustrating components of a pointing device system according to one embodiment of the invention; -
FIG. 2 depicts a flow diagram of a method for acquiring user input via a pointing device according to one embodiment of the invention; -
FIG. 3 depicts a flow diagram of a method of indexing user input according to one embodiment of the invention; -
FIG. 4A depicts a hash table illustrating an indexing database according to one embodiment of the invention; -
FIG. 4B depicts a flow diagram of a method for updating a hash table such as that illustrated inFIG. 4A ; -
FIG. 5 depicts a flow diagram of a processing sequence by which computer applications are launched according to one embodiment of the invention; and -
FIG. 6 depicts a graphical representation of a pipeline sequence for acquiring and indexing input and launching applications according to one embodiment of the invention; - To facilitate understanding, identical reference numerals have been used, where possible, to designate identical elements that are common to the figures.
- The present invention provides an improved pointing device for assisting computer users having poor or limited hand motor skill control or eyesight and for reducing the occurrence of repetitive motion injuries in physically unimpaired users. The present invention achieves this aim allows a user to manipulate an on-screen pointer and launch system applications by means of vocal commands. Thus both tracking and selecting can be achieved with little or no need to engage other parts of the body such as the hands and wrists. Furthermore, embodiments of the invention provide an improved system for acquiring vocal/audible input from a computer user, so that erroneous inputs such as background noise and/or vocal irregularities of the user are filtered out to optimize system accuracy.
- The basic operation of one embodiment of a pointing device system according the present invention is illustrated in
FIG. 1 . Operation of the pointingdevice system 100 is accomplished at least in part by the vocal commands of a computer user. Thepointing device system 100 includes at least one vocal cord movement (VCM)sensor 102, a voice box converter (VBC) 104, and akeyword database 106. Thepointing device system 100 may determine the position of the icon corresponding to application matching keyword. Thepointing device system 100 generates an output to theprocessor 108, which executes the indicated commands or pointer or icon movements. Once the pointer or icon has been moved to the desired on-screen location, the user generates a signal to launch (or “select”) the chosen application. Thesystem 100 optionally generates an audible output to indicate the new pointer position or open application, so that, for example, a blind user may use thepointing device system 100 effectively. -
FIG. 2 depicts a flow diagram of one embodiment of amethod 200 for acquiring user input via a pointing device according to one embodiment of the invention. At least oneVCM sensor 102 is positioned near a user's voice box (larynx). The user starts acquisition of vocal command data by activating astart delimiter 202. Activation of thestart delimiter 202 starts acquisition/recording of vocal cord movements. Next, the user providesaudible input 204 to theVCM sensor 102 by speaking into a converter system the keyword(s) corresponding to the movement or application the user wishes to execute. As the user speaks, thesensor 102 discerns movement of the user's vocal cords and records these movements. To stop the acquisition ofvocal input 204 by the system, the user activates astop delimiter 206 that causes theVCM sensor 102 to cease recording vocal cord movements. Optionally, the start andstop delimiters delimiter vocal input 204. The recorded vocal cord movements are then sent as input to the VBC 104. - One VBC that may be advantageously adapted for use with the present invention is commercially available from Lucent Technologies Inc. of Murray Hill, N.J. The VBC 104 converts the input from the VCM sensor(s) 102 into text output, and sends this text output to the
keyword database 106 for further processing and indexing. The VBC output is also used by software to train a personal user dictionary for alphabets, words and mouse commands, so that vocal input from a particular user is easily and accurately converted to output that theprocessor 108 can recognize and use to execute commands. Such software is well known in the art. - The
keyword database 106 is coupled to the VBC 104 and further includes adesktop icon database 350 and a menu search tree 360 (seeFIG. 3 ). Thekeyword database 106 indexes text output from the VBC 104 to determine the position of an on-screen icon or operation corresponding to the application that matches the keyword (user vocal input) processed by the VCM sensor/VBC operations. -
FIG. 3 depicts a flow diagram of amethod 300 of indexing user input according to one embodiment of the invention. In general, a spoken keyword corresponds either to a desktop icon or to a menu. In steps 302-308, thekeyword database 106 asks whether the spoken keyword corresponds to a desktop icon (302), a submenu (304), or the start menu (306), or whether the vocal input does not match any keyword at all (308). Once the keyword has been categorized, theappropriate database 350 orsearch tree 360 is searched in steps 310-314 for an exact match to an icon or menu. -
FIG. 4A depicts a hash table 400 illustrating an indexing database according to one embodiment of the invention. Referring simultaneously toFIGS. 3 and 4 A, thedesktop icon database 350 is a hash table 400 that maps icon names orkeywords 402 to specific positions (X, Y) 404 within the desktop area. For example, if thekeyword 402 corresponds to a desktop icon, a search in thedesktop icon database 350 will determine theactual position 404 of the indicated icon in the desktop area, and the on-screen pointer will be moved to that position on the desktop screen (step 316 inFIG. 3 ). The generation of the process of the hash table 400 is illustrated inFIG. 4B and will be discussed in further detail herein. - The
menu search tree 360 operates in a manner similar to thedesktop icon database 350. However, themenu search tree 360 maps names and keywords of submenus within the start menu (for example, the Accessories or Multimedia submenus in Windows), rather than names and locations of desktop icons. The on-screen pointer is then positioned on the appropriate submenu (step 318 inFIG. 3 ). If the user does not know the keyword for the desired submenu, the “Start” keyword will retrieve the root node of the menu search tree 360 (step 314) and position the pointer over the Start menu (step 320), which may be searched for the appropriate submenu. - In the event that the user input does not match a keyword in either the
desktop icon database 350 or themenu search tree 360, a signal is generated instep 308 to indicate to the user that the input is not valid. In one embodiment, the signal is an audible signal, such as a beep. - Once the pointer has been moved to the desired on-screen location, the user generates a signal to launch the chosen application. In one embodiment, the launch signal is indicated by depressing a button on a conventional mouse. For example, depression of the left mouse button on a conventional two- or three-button mouse optionally indicates selection of an application. In further embodiments, alternate means of generating selection input to the mouse PCB may be used.
-
FIG. 5 depicts a flow diagram of aprocessing sequence 500 by which aprocessor 108 launches computer applications according to one embodiment of the invention. Theprocessing sequence 500 operates on a series of timed intervals during which the user must activate a launch signal or allow a default operation to be executed. For example, once the on-screen pointer has been moved to an indicated location instep 502, (based on the database keyword matches described herein), a first timed interval begins to run. If the user wishes to launch the application on which the pointer is positioned, he signals so, for example by depressing a mouse button as described, before the first timeout instep 504. A selection signal before the first timeout prompts theprocessor 108 to launch the chosen application (step 350). If no signal is received from the user, theprocessor 108 determines in step 506 whether the pointer is positioned on a desktop icon (506 a) or a submenu (506 b). If the pointer is positioned on a submenu (506 b), that application is launched (350). If the pointer is positioned on a desktop icon (506 a), theprocessor 108 moves the pointer, instep 508, to a text area associated with the icon (located, for example, below the icon), and a second timed interval begins to run. If theprocessor 108 receives a signal from the user before the second timeout (step 510), the user is enabled to update or alter the text in the text area instep 512, and a database entry for the icon is updated accordingly instep 514. Theprocess 450 by which the icon database hash table 400 is updated is illustrated inFIG. 4B . If no signal is received from the user, the application is launched (350). In this manner, if no signal or signals are received from the user at any point in theprocessing sequence 500, the default operations of theprocessing sequence 500 are programmed to execute the probable intentions of the user in pointing to the indicated icon or submenu. -
FIG. 6 depicts a graphical representation of a pipeline sequence for acquiring and indexing input and launching applications according to one embodiment of the invention. The threemain stages stages FIG. 2 ), database indexing 300 (FIG. 3 ), and application launch 500 (FIG. 5 ), thestages system 100 is approximately three times faster than a conventional sequential processing system. - Once an application is launched 500, the pointer is repositioned within the launched application. The launched application may be any application capable of being operated by the computer, such as a word processing program, a video game or a web browser. In one embodiment, the launched application is a web browser such as Microsoft Internet Explorer.
- One embodiment of a typical web browser includes such user-executable icons as hyperlinks, buttons and menus. The purposes and operations of such icons are well known in the art. The
keyword database 106 stores keywords for all hyperlinks, buttons and menus in all active (launched) applications, as well as the respective positions (X, Y) of each hyperlink, button and menu within the application. In addition, thekeyword database 106 stores keywords and positions for all similar executable icons in other applications (i.e., word processing programs, video games, etc.). - The method of tracking (positioning) and selecting (launching) with the pointer to launch a particular action within an application is substantially similar to the method illustrated in
FIGS. 1-5 to launch the application. For example, if a user wishes to position the pointer on a particular hyperlink or button, he starts acquisition ofvocal command data 200 by activating thestart delimiter 202. Next, the user speaks into the converter system (i.e., VCM sensor 102) the keyword(s) corresponding to the hyperlink or button the user wishes to launch. To stop the acquisition of vocal input by the system, the user activates thestop delimiter 206. The acquired audible input is then converted into text form by theVBC 104, which launches the corresponding hyperlink or button by the method discussed with reference toFIG. 5 . - If a user knows the keyword for a particular submenu (for example, the Save command within the File menu) he wishes to launch or explore, he can launch it by the same method by which hyperlinks and buttons are launched (i.e., speak the keyword—“Save”—in between activating the
delimiters 202, 206). If the user does not know the keyword, he may speak the keyword for the whole menu (“File”). Exploration of the menu is then automatic; the pointer is positioned upon a submenu for brief interval of time (such as instep 504 ofFIG. 5 ), during which the user may signal (for example, by clicking a button) to launch the submenu or take no action. Taking no action will cause the processor to position the pointer upon the next submenu and repeat the process of allowing a brief interval for user input. - Switching between active applications (multitasking) may also be accomplished by vocal command. In one embodiment, the keyword “Active” is used to distinguish between a launched application and a new (but as yet unlaunched) application. If multiple applications of the same type are already launched, the keyword prompts the
processor 108 to move the on-screen pointer from one application to the next. The pointer remains on each application for a brief interval (again similar to step 504 inFIG. 5 ) to allow the user to signal acceptance of an application. If no signal is given, the pointer is moved to the next application upon timeout. - Thus the present invention represents a significant advancement in the field of computer usage and computer pointing device systems. A graphical user interface pointing device system is provided that substantially reduces the occurrence of repetitive motion injuries caused by conventional pointing devices. The system is suitable for use with standard desktop and laptop computers, as well as next-generation embedded computing devices like hand held devices and web browsing devices that operate through television sets. Furthermore, the inaccuracies associated with conventional voice-operated software are substantially reduced by the delimiting functions of the present invention, which filter background noise and vocal irregularities in the system input. Therefore, the present invention may be particularly well-suited for use by users who are deaf, ill, or have poor or limited hand motor skill control, as well as physically unimpaired users who may benefit from a hand-free pointing device system (such as those dictating in public places or while otherwise occupied).
- While the foregoing is directed to embodiments of the invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/948,263 US20080074387A1 (en) | 2003-08-08 | 2007-11-30 | Method and Apparatus for Voice-Controlled Graphical User Interface Pointing Device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/638,183 US7307615B2 (en) | 2003-08-08 | 2003-08-08 | Method and apparatus for voice-controlled graphical user interface pointing device |
US11/948,263 US20080074387A1 (en) | 2003-08-08 | 2007-11-30 | Method and Apparatus for Voice-Controlled Graphical User Interface Pointing Device |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/638,183 Division US7307615B2 (en) | 2003-08-08 | 2003-08-08 | Method and apparatus for voice-controlled graphical user interface pointing device |
Publications (1)
Publication Number | Publication Date |
---|---|
US20080074387A1 true US20080074387A1 (en) | 2008-03-27 |
Family
ID=34116737
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/638,183 Active 2025-04-11 US7307615B2 (en) | 2003-08-08 | 2003-08-08 | Method and apparatus for voice-controlled graphical user interface pointing device |
US11/948,263 Abandoned US20080074387A1 (en) | 2003-08-08 | 2007-11-30 | Method and Apparatus for Voice-Controlled Graphical User Interface Pointing Device |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/638,183 Active 2025-04-11 US7307615B2 (en) | 2003-08-08 | 2003-08-08 | Method and apparatus for voice-controlled graphical user interface pointing device |
Country Status (1)
Country | Link |
---|---|
US (2) | US7307615B2 (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005288265A (en) * | 2004-03-31 | 2005-10-20 | Procter & Gamble Co | Aqueous liquid absorbent and its manufacturing method |
US7921364B2 (en) * | 2005-11-03 | 2011-04-05 | Nuance Communications, Inc. | Controlling a computer user interface with sound |
US20090013255A1 (en) * | 2006-12-30 | 2009-01-08 | Matthew John Yuschik | Method and System for Supporting Graphical User Interfaces |
US8663182B2 (en) | 2009-06-02 | 2014-03-04 | The Procter & Gamble Company | Disposable absorbent article with absorbent waistcap or waistband and method for making the same |
US20100313133A1 (en) * | 2009-06-08 | 2010-12-09 | Microsoft Corporation | Audio and position control of user interface |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5631669A (en) * | 1994-02-04 | 1997-05-20 | Stobbs; Gregory A. | Pointing device with integral microphone |
US5818423A (en) * | 1995-04-11 | 1998-10-06 | Dragon Systems, Inc. | Voice controlled cursor movement |
US5874939A (en) * | 1996-12-10 | 1999-02-23 | Motorola, Inc. | Keyboard apparatus and method with voice recognition |
US6230287B1 (en) * | 1997-09-04 | 2001-05-08 | Mitel Corporation | Web based help desk |
US6232539B1 (en) * | 1998-06-17 | 2001-05-15 | Looney Productions, Llc | Music organizer and entertainment center |
US6233559B1 (en) * | 1998-04-01 | 2001-05-15 | Motorola, Inc. | Speech control of multiple applications using applets |
US6424357B1 (en) * | 1999-03-05 | 2002-07-23 | Touch Controls, Inc. | Voice input system and method of using same |
US20030156130A1 (en) * | 2002-02-15 | 2003-08-21 | Frankie James | Voice-controlled user interfaces |
US6771250B1 (en) * | 1999-07-27 | 2004-08-03 | Samsung Electronics Co., Ltd. | Portable computer system having application program launcher for low power consumption and method of operating the same |
US20040260438A1 (en) * | 2003-06-17 | 2004-12-23 | Chernetsky Victor V. | Synchronous voice user interface/graphical user interface |
US7109970B1 (en) * | 2000-07-01 | 2006-09-19 | Miller Stephen S | Apparatus for remotely controlling computers and other electronic appliances/devices using a combination of voice commands and finger movements |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1999005671A1 (en) * | 1997-07-24 | 1999-02-04 | Knowles Electronics, Inc. | Universal voice operated command and control engine |
-
2003
- 2003-08-08 US US10/638,183 patent/US7307615B2/en active Active
-
2007
- 2007-11-30 US US11/948,263 patent/US20080074387A1/en not_active Abandoned
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5631669A (en) * | 1994-02-04 | 1997-05-20 | Stobbs; Gregory A. | Pointing device with integral microphone |
US5818423A (en) * | 1995-04-11 | 1998-10-06 | Dragon Systems, Inc. | Voice controlled cursor movement |
US5874939A (en) * | 1996-12-10 | 1999-02-23 | Motorola, Inc. | Keyboard apparatus and method with voice recognition |
US6230287B1 (en) * | 1997-09-04 | 2001-05-08 | Mitel Corporation | Web based help desk |
US6233559B1 (en) * | 1998-04-01 | 2001-05-15 | Motorola, Inc. | Speech control of multiple applications using applets |
US6232539B1 (en) * | 1998-06-17 | 2001-05-15 | Looney Productions, Llc | Music organizer and entertainment center |
US6424357B1 (en) * | 1999-03-05 | 2002-07-23 | Touch Controls, Inc. | Voice input system and method of using same |
US6771250B1 (en) * | 1999-07-27 | 2004-08-03 | Samsung Electronics Co., Ltd. | Portable computer system having application program launcher for low power consumption and method of operating the same |
US7109970B1 (en) * | 2000-07-01 | 2006-09-19 | Miller Stephen S | Apparatus for remotely controlling computers and other electronic appliances/devices using a combination of voice commands and finger movements |
US20030156130A1 (en) * | 2002-02-15 | 2003-08-21 | Frankie James | Voice-controlled user interfaces |
US20040260438A1 (en) * | 2003-06-17 | 2004-12-23 | Chernetsky Victor V. | Synchronous voice user interface/graphical user interface |
Also Published As
Publication number | Publication date |
---|---|
US7307615B2 (en) | 2007-12-11 |
US20050030280A1 (en) | 2005-02-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Schmandt et al. | Augmenting a window system with speech input | |
US11710482B2 (en) | Natural assistant interaction | |
US8150699B2 (en) | Systems and methods of a structured grammar for a speech recognition command system | |
AU2017234428B2 (en) | Identification of voice inputs providing credentials | |
US6963937B1 (en) | Method and apparatus for providing configurability and customization of adaptive user-input filtration | |
US5377303A (en) | Controlled computer interface | |
RU2349969C2 (en) | Synchronous understanding of semantic objects realised by means of tags of speech application | |
RU2352979C2 (en) | Synchronous comprehension of semantic objects for highly active interface | |
US20100169098A1 (en) | System and method of a list commands utility for a speech recognition command system | |
US20150025885A1 (en) | System and method of dictation for a speech recognition command system | |
US10115398B1 (en) | Simple affirmative response operating system | |
WO2016190950A1 (en) | Device voice control | |
US20110264452A1 (en) | Audio output of text data using speech control commands | |
EP4075426B1 (en) | Low-latency intelligent automated assistant | |
AU2008222600A1 (en) | Method system and apparatus for entering text on a computing device | |
US20130218573A1 (en) | Voice command recognition method and related electronic device and computer-readable medium | |
US20080074387A1 (en) | Method and Apparatus for Voice-Controlled Graphical User Interface Pointing Device | |
JPH08263258A (en) | Input device, input method, information processing system and management method for input information | |
JP2005025760A (en) | Combined use of stepwise markup language and object oriented development tool | |
US20060089834A1 (en) | Verb error recovery in speech recognition | |
WO2023235010A1 (en) | Application vocabulary integration with a digital assistant | |
Perzanowski et al. | Towards seamless integration in a multi-modal interface | |
JPH1124813A (en) | Multi-modal input integration system | |
KR20000014620A (en) | Method for controlling a voice mouse | |
Hansen et al. | Gaze communication systems for people with ALS |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CREDIT SUISSE AG, NEW YORK Free format text: SECURITY AGREEMENT;ASSIGNOR:LUCENT, ALCATEL;REEL/FRAME:029821/0001 Effective date: 20130130 Owner name: CREDIT SUISSE AG, NEW YORK Free format text: SECURITY AGREEMENT;ASSIGNOR:ALCATEL LUCENT;REEL/FRAME:029821/0001 Effective date: 20130130 |
|
AS | Assignment |
Owner name: SOUND VIEW INNOVATIONS, LLC, NEW JERSEY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ALCATEL LUCENT;REEL/FRAME:033416/0763 Effective date: 20140630 |
|
AS | Assignment |
Owner name: ALCATEL LUCENT, FRANCE Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:CREDIT SUISSE AG;REEL/FRAME:033868/0555 Effective date: 20140819 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION |