CA2378535C - System and method for transmitting voice input from a remote location over a wireless data channel - Google Patents
System and method for transmitting voice input from a remote location over a wireless data channel Download PDFInfo
- Publication number
- CA2378535C CA2378535C CA002378535A CA2378535A CA2378535C CA 2378535 C CA2378535 C CA 2378535C CA 002378535 A CA002378535 A CA 002378535A CA 2378535 A CA2378535 A CA 2378535A CA 2378535 C CA2378535 C CA 2378535C
- Authority
- CA
- Canada
- Prior art keywords
- user
- sending
- server
- speech
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/487—Arrangements for providing information services, e.g. recorded voice services or time announcements
- H04M3/493—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
- H04M3/4938—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals comprising a voice browser which renders and interprets, e.g. VoiceXML
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2201/00—Electronic components, circuits, software, systems or apparatus used in telephone systems
- H04M2201/40—Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
Abstract
A system and method for improving voice recognition processing at a server system that receives voice input from a remotely located user system. The user system includes a microphone, a processor that performs front-end voice recognition processing of the received user voice input, and a communication component configured to send the front- end processed user voice input to a destination wirelessly over a network. The server system includes a communication component configured to receive the sent front-end processed user voice input, and a processor configured to complete voice recognition processing of the sent front-end processed user voice input.
Description
S
SYSTEM AND METHOD FOR TRANSMITTING VOICE INPUT FROM
A REMOTE LOCATION OVER A WIRELESS DATA CHANNEL
INVENTORS
Gilad Odinak Thomas R. McCann :lulien Rivarol Vergin FIELD OF THE INVENTION
This invention relates generally to the field of wireless communication, digital signal processing, and to a system and method for transmitting data derived from a voice input at a remote, mobile location over a wireless data channel.
BACKGROUND OF THE INVENTION
Cellular telephone transmission has long been used in mobile communication.
Traditionally, cellular telephone transmission has been used to facilitate conversations between remote individuals. More recently, these same systems have been modified to facilitate the communication of verbal instructions to remote computer systems using speech recognition programs. In these modified systems, the speaker's verbal instructions are converted to digital data instructions, which in turn are used by a computer system to carry out the desired operation. The ability to communicate verbal instructions "hands-free" carnes obvious advantages, not the least of which include safety advantages, especially when the speaker is concurrently attempting to operate a vehicle.
The traditional implementation of this type of speech transmission and recognition occurs in one of two ways: (1) sending raw audio (i.e., the speaker's verbal instructions) from a receiver in the vehicle, such as a cellular telephone, to the remote computer system, where the verbal instructions are converted to data instructions; or (2) performing extensive automated speech recognition (ASR) in the receiver (e.g., recognizing words and phrases), and sending the converted digital data instructions to the remote computer system. Both existing implementations suffer from significant disadvantages. Raw audio sent across a cellular network suffers from signal degradation, in turn diminishing the integrity of the audio signal to be converted to digital data instructions and, hence, the accuracy of the ultimate instructions. While converting the audio signal to digital data at the vehicle addresses this problem, it requires expensive computing power in the vehicle, which is logistically and cost prohibitive.
Thus, there is a need for a mobile speech recognition system and method that addresses the disadvantages with the current implementations.
SUMMARY
The present invention provides a voice communication system and method for improving voice recognition processing at a server system that receives voice input from a remotely located user system. The user system includes a microphone configured to receive user voice input, a processor configured to perform front-end voice recognition processing of the received user voice input, and a communication component configured to send the front-end processed user voice input to a destination over a network. The server system includes a communication component configured to receive the sent front-end processed user voice input, and a processor configured to complete voice recognition processing of the sent front-end processed user voice input.
In accordance with further aspects of the invention, the communication component of the user system communicates wirelessly.
In accordance with still further aspects of the invention, the processor of the user system includes a sampling component configured to sample the received user voice input.
The processor of the user system further includes at least one of a noise cancellation component, an echo-cancellation component or an end-pointing component.
SYSTEM AND METHOD FOR TRANSMITTING VOICE INPUT FROM
A REMOTE LOCATION OVER A WIRELESS DATA CHANNEL
INVENTORS
Gilad Odinak Thomas R. McCann :lulien Rivarol Vergin FIELD OF THE INVENTION
This invention relates generally to the field of wireless communication, digital signal processing, and to a system and method for transmitting data derived from a voice input at a remote, mobile location over a wireless data channel.
BACKGROUND OF THE INVENTION
Cellular telephone transmission has long been used in mobile communication.
Traditionally, cellular telephone transmission has been used to facilitate conversations between remote individuals. More recently, these same systems have been modified to facilitate the communication of verbal instructions to remote computer systems using speech recognition programs. In these modified systems, the speaker's verbal instructions are converted to digital data instructions, which in turn are used by a computer system to carry out the desired operation. The ability to communicate verbal instructions "hands-free" carnes obvious advantages, not the least of which include safety advantages, especially when the speaker is concurrently attempting to operate a vehicle.
The traditional implementation of this type of speech transmission and recognition occurs in one of two ways: (1) sending raw audio (i.e., the speaker's verbal instructions) from a receiver in the vehicle, such as a cellular telephone, to the remote computer system, where the verbal instructions are converted to data instructions; or (2) performing extensive automated speech recognition (ASR) in the receiver (e.g., recognizing words and phrases), and sending the converted digital data instructions to the remote computer system. Both existing implementations suffer from significant disadvantages. Raw audio sent across a cellular network suffers from signal degradation, in turn diminishing the integrity of the audio signal to be converted to digital data instructions and, hence, the accuracy of the ultimate instructions. While converting the audio signal to digital data at the vehicle addresses this problem, it requires expensive computing power in the vehicle, which is logistically and cost prohibitive.
Thus, there is a need for a mobile speech recognition system and method that addresses the disadvantages with the current implementations.
SUMMARY
The present invention provides a voice communication system and method for improving voice recognition processing at a server system that receives voice input from a remotely located user system. The user system includes a microphone configured to receive user voice input, a processor configured to perform front-end voice recognition processing of the received user voice input, and a communication component configured to send the front-end processed user voice input to a destination over a network. The server system includes a communication component configured to receive the sent front-end processed user voice input, and a processor configured to complete voice recognition processing of the sent front-end processed user voice input.
In accordance with further aspects of the invention, the communication component of the user system communicates wirelessly.
In accordance with still further aspects of the invention, the processor of the user system includes a sampling component configured to sample the received user voice input.
The processor of the user system further includes at least one of a noise cancellation component, an echo-cancellation component or an end-pointing component.
-2-In accordance with yet other aspects of the invention, the processor of the server may include a component configured to perform a function based on the completed voice recognition processing.
In accordance with still another aspect of the invention, the user system may include removable modules. The modules may include a processing module that includes a sampling component configured to sample the received user voice input. Also, the processing module may include a noise cancellation component, an echo-cancellation component and/or an end-pointing component.
In accordance with still further aspects of the invention, the modules may include at least one of a positioning module, a phone adapter module, or a wireless network communication module.
In accordance with another aspect of the invention, there is provided a voice communication method. The communication method involves receiving an acoustic analog signal at a user system directly from a user, converting the analog signal to a digital signal, canceling noise from the digital signal to form a processed digital signal, then detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker. The method further involves sending the detected user speech to a server over a network, and performing speech recognition processing of the sent detected user speech at the server.
In accordance with another aspect of the invention, there is provided a voice communication method which involves receiving an acoustic analog signal at a user system directly from a user, converting the analog signal to a digital signal, canceling noise from the digital signal to form a processed digital signal, then detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo-cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker. The method further involves sending the detected user speech to a server over a network, performing speech recognition processing of the detected user speech at the server, and performing a function at the server based on the performed speech recognition processing.
In accordance with another aspect of the invention, there is provided a voice communication system comprising a user system in communication with a server system over a network. The user system includes a microphone configured to receive user acoustic
In accordance with still another aspect of the invention, the user system may include removable modules. The modules may include a processing module that includes a sampling component configured to sample the received user voice input. Also, the processing module may include a noise cancellation component, an echo-cancellation component and/or an end-pointing component.
In accordance with still further aspects of the invention, the modules may include at least one of a positioning module, a phone adapter module, or a wireless network communication module.
In accordance with another aspect of the invention, there is provided a voice communication method. The communication method involves receiving an acoustic analog signal at a user system directly from a user, converting the analog signal to a digital signal, canceling noise from the digital signal to form a processed digital signal, then detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker. The method further involves sending the detected user speech to a server over a network, and performing speech recognition processing of the sent detected user speech at the server.
In accordance with another aspect of the invention, there is provided a voice communication method which involves receiving an acoustic analog signal at a user system directly from a user, converting the analog signal to a digital signal, canceling noise from the digital signal to form a processed digital signal, then detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo-cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker. The method further involves sending the detected user speech to a server over a network, performing speech recognition processing of the detected user speech at the server, and performing a function at the server based on the performed speech recognition processing.
In accordance with another aspect of the invention, there is provided a voice communication system comprising a user system in communication with a server system over a network. The user system includes a microphone configured to receive user acoustic
3 analog signals, a processor configured to convert the analog signals to digital signals and process the digital signals based on two or more of noise cancellation, echo-cancellation or end-pointing to form a processed voice signal, wherein the processed voice signal is in a format capable of being outputted over a speaker. The user system further includes a communication component configured to send the processed voice signal to a destination, such as the server system, over the network. The server system includes a communication component configured to receive the sent processed voice signal, and a processor configured to perform speech recognition processing of the sent processed voice signal.
In accordance with another aspect of the invention, there is provided a voice communication system which includes means for receiving user acoustic analog signals at a user system directly from a user, means for converting the analog signals to digital signals, means for processing the digital signals at the user system based on two or more of noise cancellation, echo-cancellation or end-pointing to form a processed voice signal, wherein the processed voice signal is in a format capable of being outputted over a speaker. The system 1 S further includes means for sending the processed voice signal to a server over a network, and means for performing speech recognition processing of the sent processed voice signal at the server.
As will be readily appreciated from the foregoing summary, the invention provides an improved system and method for performing voice recognition processing where the processing is performed remote from the person inputting the voice.
BRIEF DESCRIPTION OF THE DRAWINGS
The preferred and alternative embodiments of the present invention are described in detail below with reference to the following drawings.
FIGURE 1 is a diagram illustrating the general architecture of a mobile speech recognition system that operates in accordance with the present invention;
FIGURE 2 is a flow chart illustrating the operation of the present invention;
FIGURE 3 is a flow chart illustrating the operation of one embodiment of the front-end sound processing aspect of the present invention; and FIGURES 4A-C are block diagrams of data and voice package streams generated by an alternate embodiment of the present invention; and FIGURE 5 is a flow diagram.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
The present invention provides a system and method for transmitting data derived from voice input at a remote, mobile location over a wireless data channel. By way of overview and with reference to FIGURE 1, the mobile speech recognition system 10 of the present invention includes a user system 20 in wireless data communication with a server system 40 via a wireless data channel 60. The invention performs front-end sound processing in user system 20 - short of full speech recognition conversion - and sends the results of the sound processing to server 40 in a digital form via wireless data channel 60.
More specifically, FIGURE 1 illustrates the particular components of the preferred embodiment of system 10 for sending preprocessed voice over the wireless data channel.
User system 20 includes a user interface 22 having a microphone 24 for capturing the user's voice, a display 26, speakers 28, user interface 34 for allowing the user to perform various interactive functions, and preferably a video capture device 32. The user system further includes a voice processor 34 for performing front-end sound processing of the user's voice, and a communications device 36, such as an antenna or other signal carnet, for transmitting and receiving wireless information. User system 20 is preferably implemented in a mobile environment, such as in a car or other vehicle.
In an alternate embodiment, user system 20 includes a processor with memory coupled to multiple controllers (not shown) for receiving data from various vehicle data buses and module receptacles (not shown). The module receptacles are configured to receive various add-on modules, such as modules for coupling other electronic devices, communication modules, processing modules and other modules, such as a phone adapter module, a CPU/memory module, a bluetooth, infrared or wireless network communication module, and a global positioning system (GPS) module. In operation, voice capture and front-end sound processing is accomplished in various modules to the user system. Such modules are replaceable, allowing a user to receive up-to-date sound processing software or other more advanced modules without having to replace any hardware components already installed in the vehicle.
User system 20 is in communication with server system 40. The server system includes a server 42 for housing user system information, as well as processing and responding to requests for information from the user system and information sources 44, which may be integral with or independent from server system 40. The information sources store information subject to requests from the user system. In the preferred embodiment, the server includes a processor, a memory, and a database (not shown). Server 42 may be in communication with information sources 44 via direct access (e.g., hard-wired or point-to-point connection) as well as over Internet 46. Server system 40 further includes a means for sending and receiving information to user system 20, discussed below.
Wireless data channel 60 facilitates communication of instructions and information between user system 20 and server system 40. In a preferred embodiment, the wireless data channel may include a satellite system 62 in combination with a satellite dish 64 along with or in the place of an access point 66, the latter as part of a cellular or other wireless transmission network. In operation, instructions are transmitted from user system 20 via transmission/reception device 34 to either the satellite system or access point, which in turn
In accordance with another aspect of the invention, there is provided a voice communication system which includes means for receiving user acoustic analog signals at a user system directly from a user, means for converting the analog signals to digital signals, means for processing the digital signals at the user system based on two or more of noise cancellation, echo-cancellation or end-pointing to form a processed voice signal, wherein the processed voice signal is in a format capable of being outputted over a speaker. The system 1 S further includes means for sending the processed voice signal to a server over a network, and means for performing speech recognition processing of the sent processed voice signal at the server.
As will be readily appreciated from the foregoing summary, the invention provides an improved system and method for performing voice recognition processing where the processing is performed remote from the person inputting the voice.
BRIEF DESCRIPTION OF THE DRAWINGS
The preferred and alternative embodiments of the present invention are described in detail below with reference to the following drawings.
FIGURE 1 is a diagram illustrating the general architecture of a mobile speech recognition system that operates in accordance with the present invention;
FIGURE 2 is a flow chart illustrating the operation of the present invention;
FIGURE 3 is a flow chart illustrating the operation of one embodiment of the front-end sound processing aspect of the present invention; and FIGURES 4A-C are block diagrams of data and voice package streams generated by an alternate embodiment of the present invention; and FIGURE 5 is a flow diagram.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
The present invention provides a system and method for transmitting data derived from voice input at a remote, mobile location over a wireless data channel. By way of overview and with reference to FIGURE 1, the mobile speech recognition system 10 of the present invention includes a user system 20 in wireless data communication with a server system 40 via a wireless data channel 60. The invention performs front-end sound processing in user system 20 - short of full speech recognition conversion - and sends the results of the sound processing to server 40 in a digital form via wireless data channel 60.
More specifically, FIGURE 1 illustrates the particular components of the preferred embodiment of system 10 for sending preprocessed voice over the wireless data channel.
User system 20 includes a user interface 22 having a microphone 24 for capturing the user's voice, a display 26, speakers 28, user interface 34 for allowing the user to perform various interactive functions, and preferably a video capture device 32. The user system further includes a voice processor 34 for performing front-end sound processing of the user's voice, and a communications device 36, such as an antenna or other signal carnet, for transmitting and receiving wireless information. User system 20 is preferably implemented in a mobile environment, such as in a car or other vehicle.
In an alternate embodiment, user system 20 includes a processor with memory coupled to multiple controllers (not shown) for receiving data from various vehicle data buses and module receptacles (not shown). The module receptacles are configured to receive various add-on modules, such as modules for coupling other electronic devices, communication modules, processing modules and other modules, such as a phone adapter module, a CPU/memory module, a bluetooth, infrared or wireless network communication module, and a global positioning system (GPS) module. In operation, voice capture and front-end sound processing is accomplished in various modules to the user system. Such modules are replaceable, allowing a user to receive up-to-date sound processing software or other more advanced modules without having to replace any hardware components already installed in the vehicle.
User system 20 is in communication with server system 40. The server system includes a server 42 for housing user system information, as well as processing and responding to requests for information from the user system and information sources 44, which may be integral with or independent from server system 40. The information sources store information subject to requests from the user system. In the preferred embodiment, the server includes a processor, a memory, and a database (not shown). Server 42 may be in communication with information sources 44 via direct access (e.g., hard-wired or point-to-point connection) as well as over Internet 46. Server system 40 further includes a means for sending and receiving information to user system 20, discussed below.
Wireless data channel 60 facilitates communication of instructions and information between user system 20 and server system 40. In a preferred embodiment, the wireless data channel may include a satellite system 62 in combination with a satellite dish 64 along with or in the place of an access point 66, the latter as part of a cellular or other wireless transmission network. In operation, instructions are transmitted from user system 20 via transmission/reception device 34 to either the satellite system or access point, which in turn
-4-WING~1 I-1003AP
communicate the instructions to server 42, in the former case via satellite dish 64.
Conversely, information may be communicated from the server to the user system along a reverse direction of the same route.
The system and method of the present invention is better understood with reference to FIGURE 2, which illustrates the operational steps of the mobile speech recognition system.
At block 100, the system receives the user's voice (acoustical audio input) at user system 20.
At block 102, the system performs front-end sound processing on the acoustical audio input using processor 34, including converting the audio input to representative digital data. This front-end sound processing is discussed in more detail below with reference to FIGURE 3.
Once front-end sound processing is complete, at block 104 the system packages the digital data for wireless delivery. At block 106, the system transmits the packaged digital voice information from user system 20 to server system 40 via wireless data channel 60.
At block 108, server system 40 performs complete speech recognition processing on the digital voice information using server 42. As part of this process, the server attempts to match the digital voice information with corresponding digital data entries in the server database, which in turn are associated with a program instructional language.
One of three results follow from this process: ( 1 ) a match is found, in which case the corresponding program instructions are executed at block 110; (2) no match is found, in which case no instructions are executed (or partial instructions are executed) at block 110;
or (3) multiple matches are found. In the last case, the server employs statistical models to decide which of the several matched entries is mast probably the desired entry. This statistical modeling may look at previously matched digital voice information and, alone or in context with associated digital voice information, may combine entries in combination to best form appropriate instructional language (e.g., "the chair is white" is more probable than "the chair his white.").
The server may also employ grammar, which is a subset of rules associated with the program instructional language, which includes the structure for valid sentences at certain points in the dialog from which the digital voice information was derived. Once the most probable match is determined, the resulting program instructions are executed at block 110.
The front-end sound processing of the present invention is better understood with reference to FIGURE 3. Microphone 24 of user system 20 captures the user's voice as acoustical audio input. At block 200, this acoustical audio input is converted to an analog electrical signal. At block 202, the analog signal is sampled and digitized.
In the preferred embodiment, there are thousands ol'samples per second. Each sample is assigned a numerical value, which is typically between 0-255 so that it fits in a byte. In an effort to improve the accuracy of the resulting sample, processor 34 performs noise-cancellation,
communicate the instructions to server 42, in the former case via satellite dish 64.
Conversely, information may be communicated from the server to the user system along a reverse direction of the same route.
The system and method of the present invention is better understood with reference to FIGURE 2, which illustrates the operational steps of the mobile speech recognition system.
At block 100, the system receives the user's voice (acoustical audio input) at user system 20.
At block 102, the system performs front-end sound processing on the acoustical audio input using processor 34, including converting the audio input to representative digital data. This front-end sound processing is discussed in more detail below with reference to FIGURE 3.
Once front-end sound processing is complete, at block 104 the system packages the digital data for wireless delivery. At block 106, the system transmits the packaged digital voice information from user system 20 to server system 40 via wireless data channel 60.
At block 108, server system 40 performs complete speech recognition processing on the digital voice information using server 42. As part of this process, the server attempts to match the digital voice information with corresponding digital data entries in the server database, which in turn are associated with a program instructional language.
One of three results follow from this process: ( 1 ) a match is found, in which case the corresponding program instructions are executed at block 110; (2) no match is found, in which case no instructions are executed (or partial instructions are executed) at block 110;
or (3) multiple matches are found. In the last case, the server employs statistical models to decide which of the several matched entries is mast probably the desired entry. This statistical modeling may look at previously matched digital voice information and, alone or in context with associated digital voice information, may combine entries in combination to best form appropriate instructional language (e.g., "the chair is white" is more probable than "the chair his white.").
The server may also employ grammar, which is a subset of rules associated with the program instructional language, which includes the structure for valid sentences at certain points in the dialog from which the digital voice information was derived. Once the most probable match is determined, the resulting program instructions are executed at block 110.
The front-end sound processing of the present invention is better understood with reference to FIGURE 3. Microphone 24 of user system 20 captures the user's voice as acoustical audio input. At block 200, this acoustical audio input is converted to an analog electrical signal. At block 202, the analog signal is sampled and digitized.
In the preferred embodiment, there are thousands ol'samples per second. Each sample is assigned a numerical value, which is typically between 0-255 so that it fits in a byte. In an effort to improve the accuracy of the resulting sample, processor 34 performs noise-cancellation,
-5-echo-cancellation, and end-pointing routines. Noise-cancellation involves an examination of the amount of energy in the analog signal at various frequencies. Processor 34 detects the background noise accompanying the original signal and subtracts it from the sample. Similar to noise-cancellation, echo-cancellation involves an examination of the amount of energy in the analog signal at various frequencies. With echo-cancellation, the goal is to reduce or eliminate from the sampled signal the vocal system prompt or audio component used by the system to elicit the user's original audio input. By comparing sampled signal components to a data stream representing the system prompt, processor 34 is able to subtract the system prompt and thereby separate the data stream representing the system prompt from the data stream representing the user's audio input. Also, undesired radio signals and other signals that processor 34 knows aren't the user's speech are also removed using the above described technique.
After performing noise- and echo-cancellation, processor 34 looks at the remaining amount of energy at the various frequencies and determines whether it contains actual user audio input, or user speech, as opposed to silence or other non-relevant noise. This is accomplished by examining the energy or amplitude of the audio input and the rate of "zero-crossing." Zero-crossing is where the audio signal changes from positive to negative, or vice versa. When the energy and zero-crossings are at certain predetermined levels, there is a likelihood of occurrence of a speech event.
In an alternate embodiment, processor 34 determines the beginning and end of a spoken phrase. This is performed by the technique known as end-pointing (i.e.
speech detection). The determined beginning and end of a spoken phrase constitutes preprocessed information that is sent to server 40.
At block 204, resulting samples are grouped into sets representing acoustic audio for specified time periods. For example, in the preferred embodiment, a group contains all samples in a ten millisecond time period. The groups may have some overlap with respect to samples taken in succession, i.e., group N contains samples from time T to T+10 milliseconds; group N+1 contains samples from time T+8 to T+18 milliseconds, etc. In this example, if 20,000 samples are included per second, each group contains 200 samples.
At block 206, processor 34 converts grouped sets from time domain to frequency domain. In the preferred embodiment, this is accomplished using Fast Fourier Transform: the output of the transformation is a set of numbers, each set representing the amount of audio energy in a frequency band. The width of the frequency bands may be linear or nonlinear.
For example, a given set of 10 millisecond time periods will have the following frequency domain values: 150Hz-300Hz: 75; 350Hz-400Hz: 100; . . . 1000Hz-1200Hz: 125 . .
.
_6_ WfNG-11-1003AP
2500Hz-3000Hz: 47. In this example, the low frequency bands are SOHz each, while the highest frequency band is SOOHz.
FIGURES 4A-C and 5 illustrate an alternate embodiment of the present invention for metering the deliverable information from user system 20 to server 42. FIGURES
S illustrate packetized data streams 300, 310, 320 that are outputted streams during different modes of operation of user system 20. FIGURE 4A illustrates packet stream 300, which includes packets 304, 306 that are filled with voice information. Packet stream 300 also includes an address ID 302 located at the beginning. In this mode of operation, processor 34 has determined that the entire packet stream 300 should be filled with voice information and therefore will populate the packets with only voice information. This mode of operation occurs, for example, when the user is speaking into microphone 24.
FIGURE 4B illustrates a second mode of operation for the delivery of information from user system 20 to server 42. In this mode of operation, some packets 312 are filled with voice information and other packets 314 are filled with data, such as GPS or other vehicle information. Processor 34 operates in the mode shown in FIGURE 4B in order to deliver user system information along with the sent voice information. Periodically, vehicle information must be sent in order to be processed by the server for various functions.
Often times it is desirable that vehicle information be associated with the voice information transmitted at a given time. In one example, a user may verbally request instructions to the nearest gas station, which answer would necessitate knowledge of the vehicle's present location. The present invention may send the voice instruction concurrent with GPS
information, ultimately providing server 42 with both pieces of information as part of a nearly simultaneous transmission.
FIGURE 4C illustrates a mode of operation that processor 34 executes when no voice is being entered by the user, therefore allowing the entire packet stream 320 to be populated by data as shown in packets 322 and 324.
FIGURE 5 illustrates an example decision process performed by processor 34 when deciding how to populate packets sent from user system 20 to server 42. First, at decision block 350, processor 34 determines if user system 20 is receiving speech from the user. If it is determined that no speech is being received, processor 34 formats the transmission to include only data (see block 352)., as shown by example in FI(JURE 4C. If processor 34 determines that user system 20 is receiving speech, the process continues to decision block 356, where the processor determines if any data needs to be sent with the speech. If no data needs to be sent with the speech as determined by processor 34, the processor populates the packets of the transmission with only voice information (see block 358), as shown by WING-11-10()3AP
example in FIGURE 4A. If processor 34 determines that data needs to be sent with the speech, the processor populates some of the data packets with voice and other data packets with data (see block 360), as shown by example in FIGURE 4B. When data packets and voice packets are distributed in the transmission, processor 34 populates packets with voice in order to optimize speech recognition at server 42.
While the preferred embodiment of the invention has been illustrated and described, as noted above, many changes can be made without departing from the spirit and scope of the invention. For example, the order of some or all of the front-end sound processing steps may be altered, such as the order in which noise-cancellation and echo-cancellation routines occur. In another example, the seduence used to determine whether input includes speech or data or both may be altered. Accordingly, the scope of the invention is not limited by the disclosure of the preferred embodiment.
_g_
After performing noise- and echo-cancellation, processor 34 looks at the remaining amount of energy at the various frequencies and determines whether it contains actual user audio input, or user speech, as opposed to silence or other non-relevant noise. This is accomplished by examining the energy or amplitude of the audio input and the rate of "zero-crossing." Zero-crossing is where the audio signal changes from positive to negative, or vice versa. When the energy and zero-crossings are at certain predetermined levels, there is a likelihood of occurrence of a speech event.
In an alternate embodiment, processor 34 determines the beginning and end of a spoken phrase. This is performed by the technique known as end-pointing (i.e.
speech detection). The determined beginning and end of a spoken phrase constitutes preprocessed information that is sent to server 40.
At block 204, resulting samples are grouped into sets representing acoustic audio for specified time periods. For example, in the preferred embodiment, a group contains all samples in a ten millisecond time period. The groups may have some overlap with respect to samples taken in succession, i.e., group N contains samples from time T to T+10 milliseconds; group N+1 contains samples from time T+8 to T+18 milliseconds, etc. In this example, if 20,000 samples are included per second, each group contains 200 samples.
At block 206, processor 34 converts grouped sets from time domain to frequency domain. In the preferred embodiment, this is accomplished using Fast Fourier Transform: the output of the transformation is a set of numbers, each set representing the amount of audio energy in a frequency band. The width of the frequency bands may be linear or nonlinear.
For example, a given set of 10 millisecond time periods will have the following frequency domain values: 150Hz-300Hz: 75; 350Hz-400Hz: 100; . . . 1000Hz-1200Hz: 125 . .
.
_6_ WfNG-11-1003AP
2500Hz-3000Hz: 47. In this example, the low frequency bands are SOHz each, while the highest frequency band is SOOHz.
FIGURES 4A-C and 5 illustrate an alternate embodiment of the present invention for metering the deliverable information from user system 20 to server 42. FIGURES
S illustrate packetized data streams 300, 310, 320 that are outputted streams during different modes of operation of user system 20. FIGURE 4A illustrates packet stream 300, which includes packets 304, 306 that are filled with voice information. Packet stream 300 also includes an address ID 302 located at the beginning. In this mode of operation, processor 34 has determined that the entire packet stream 300 should be filled with voice information and therefore will populate the packets with only voice information. This mode of operation occurs, for example, when the user is speaking into microphone 24.
FIGURE 4B illustrates a second mode of operation for the delivery of information from user system 20 to server 42. In this mode of operation, some packets 312 are filled with voice information and other packets 314 are filled with data, such as GPS or other vehicle information. Processor 34 operates in the mode shown in FIGURE 4B in order to deliver user system information along with the sent voice information. Periodically, vehicle information must be sent in order to be processed by the server for various functions.
Often times it is desirable that vehicle information be associated with the voice information transmitted at a given time. In one example, a user may verbally request instructions to the nearest gas station, which answer would necessitate knowledge of the vehicle's present location. The present invention may send the voice instruction concurrent with GPS
information, ultimately providing server 42 with both pieces of information as part of a nearly simultaneous transmission.
FIGURE 4C illustrates a mode of operation that processor 34 executes when no voice is being entered by the user, therefore allowing the entire packet stream 320 to be populated by data as shown in packets 322 and 324.
FIGURE 5 illustrates an example decision process performed by processor 34 when deciding how to populate packets sent from user system 20 to server 42. First, at decision block 350, processor 34 determines if user system 20 is receiving speech from the user. If it is determined that no speech is being received, processor 34 formats the transmission to include only data (see block 352)., as shown by example in FI(JURE 4C. If processor 34 determines that user system 20 is receiving speech, the process continues to decision block 356, where the processor determines if any data needs to be sent with the speech. If no data needs to be sent with the speech as determined by processor 34, the processor populates the packets of the transmission with only voice information (see block 358), as shown by WING-11-10()3AP
example in FIGURE 4A. If processor 34 determines that data needs to be sent with the speech, the processor populates some of the data packets with voice and other data packets with data (see block 360), as shown by example in FIGURE 4B. When data packets and voice packets are distributed in the transmission, processor 34 populates packets with voice in order to optimize speech recognition at server 42.
While the preferred embodiment of the invention has been illustrated and described, as noted above, many changes can be made without departing from the spirit and scope of the invention. For example, the order of some or all of the front-end sound processing steps may be altered, such as the order in which noise-cancellation and echo-cancellation routines occur. In another example, the seduence used to determine whether input includes speech or data or both may be altered. Accordingly, the scope of the invention is not limited by the disclosure of the preferred embodiment.
_g_
Claims (25)
PROPERTY OR PRIVILEGE IS CLAIMED ARE DEFINED AS FOLLOWS:
1. A voice communication method comprising:
receiving an acoustic analog signal at a user system directly from a user;
converting the analog signal to a digital signal;
canceling noise from the digital signal to form a processed digital signal;
detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo-cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker;
sending the detected user speech to a server over a network, and performing speech recognition processing of the sent detected user speech at the server.
receiving an acoustic analog signal at a user system directly from a user;
converting the analog signal to a digital signal;
canceling noise from the digital signal to form a processed digital signal;
detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo-cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker;
sending the detected user speech to a server over a network, and performing speech recognition processing of the sent detected user speech at the server.
2. The method of claim 1, wherein sending is wirelessly sending.
3. The method of claim 1 or 2, wherein the user system is implemented in a vehicle.
4. The method of claim 1, 2, or 3, wherein converting the analog signal includes sampling the acoustic analog signal.
5. The method of claim 1, 2, 3, or 4, further comprising performing a function at the server based on the performed speech recognition processing.
6. The method of claim 1, 2, 3, 4 or 5, further comprising receiving user system status - information, and wherein sending the detected user speech to a server over a network comprises sending said user system status information with the detected user speech based on transmission requirements.
7. The method of claim 6, wherein sending the detected user speech to a server over a network includes sending the user system status information and the detected user speech in interspersed distinct transmission packets.
8. The method of claim 6, further comprising sending only the user system status information to the server when no acoustic analog signals are received.
9. A voice communication method comprising:
receiving an acoustic analog signal at a user system directly from a user;
converting the analog signal to a digital signal;
canceling noise from the digital signal to form a processed digital signal;
detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo-cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker;
sending the detected user speech to a server over a network;
performing speech recognition processing of the detected user speech at the server, and performing a function at the server based on the performed speech recognition processing.
receiving an acoustic analog signal at a user system directly from a user;
converting the analog signal to a digital signal;
canceling noise from the digital signal to form a processed digital signal;
detecting user speech in the processed digital signal by evaluating two or more of noise cancellation, echo-cancellation or end-pointing, wherein the detected user speech is in a format capable of being outputted over a speaker;
sending the detected user speech to a server over a network;
performing speech recognition processing of the detected user speech at the server, and performing a function at the server based on the performed speech recognition processing.
10. A voice communication system comprising:
a user system comprising:
a microphone configured to receive user acoustic analog signals;
a processor configured to convert the analog signals to digital signals and process the digital signals based on two or more of noise cancellation, echo-cancellation or end-pointing to form a processed voice signal, wherein the processed voice signal is in a format capable of being outputted over a speaker; and a communication component configured to send the processed voice signal to a destination over a network; and a server system coupled to the network, the server comprising:
a communication component configured to receive the sent processed voice signal ; and a processor configured to perform speech recognition processing of the sent processed voice signal.
a user system comprising:
a microphone configured to receive user acoustic analog signals;
a processor configured to convert the analog signals to digital signals and process the digital signals based on two or more of noise cancellation, echo-cancellation or end-pointing to form a processed voice signal, wherein the processed voice signal is in a format capable of being outputted over a speaker; and a communication component configured to send the processed voice signal to a destination over a network; and a server system coupled to the network, the server comprising:
a communication component configured to receive the sent processed voice signal ; and a processor configured to perform speech recognition processing of the sent processed voice signal.
11. The system of claim 10, wherein the communication component of the user system communicates wirelessly.
12. The system of claim 10 or 11, wherein the user system is implemented in a vehicle.
13. The system of claim 10, 11 or 12, wherein the processor of the user system comprises a sampling component configured to sample the received user acoustic analog signals.
14. The system of any one of claims 10 to 13, wherein the processor of the server comprises a component configured to perform a function based on the performed speech recognition processing.
15. The system of any one of claims 10 to 14, wherein the user system further comprises removable modules.
16. The system of claim 15, wherein:
the modules comprise a processing module; and the processor of the user system comprises a sampling component configured to sample the received user acoustic analog signals.
the modules comprise a processing module; and the processor of the user system comprises a sampling component configured to sample the received user acoustic analog signals.
17. The system of claim 15, wherein the modules comprise at least one of a positioning module, a phone adapter module, or a wireless network communication module.
18. A voice communication system comprising:
a means for receiving user acoustic analog signals at a user system directly from a user;
a means for converting the analog signals to digital signals;
a means for processing the digital signals at the user system based on two or more of noise cancellation, echo-cancellation or end-pointing to form a processed voice signal, wherein the processed voice signal is in a format capable of being outputted over a speaker;
a means for sending the processed voice signal to a server over a network; and a means for performing speech recognition processing of the sent processed voice signal at the server.
a means for receiving user acoustic analog signals at a user system directly from a user;
a means for converting the analog signals to digital signals;
a means for processing the digital signals at the user system based on two or more of noise cancellation, echo-cancellation or end-pointing to form a processed voice signal, wherein the processed voice signal is in a format capable of being outputted over a speaker;
a means for sending the processed voice signal to a server over a network; and a means for performing speech recognition processing of the sent processed voice signal at the server.
19. The system of claim 18, wherein the means for sending is a means for wirelessly sending.
20. The system of claim 18 or 19, wherein the user system is implemented in a vehicle.
21. The system of claim 18, 19, or 20, wherein the means for converting comprises a means for sampling the received user acoustic analog signals.
22. The system of any one of claims 18 to 21, further comprising a means for performing a function at the server based on the performed speech recognition processing.
23. The system of any one of claims 18 to 22, further comprising a means for receiving user system status information, and wherein the means for sending the processed voice signal to a server over a network comprises means for sending the user system status information with the processed voice signal based on transmission requirements.
24. The system of claim 23, wherein the means for sending comprises means for sending the user system status information and the processed voice signal in interspersed distinct transmission packets.
25. The system of claim 23, wherein the means for sending the processed voice signal to a server over a network comprises means for sending only the user system status information when no user acoustic analog signals are received at the means for receiving.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US28037901P | 2001-03-29 | 2001-03-29 | |
US60/280,379 | 2001-03-29 | ||
US10/059,905 US6885735B2 (en) | 2001-03-29 | 2002-01-29 | System and method for transmitting voice input from a remote location over a wireless data channel |
US10/059,905 | 2002-01-29 |
Publications (2)
Publication Number | Publication Date |
---|---|
CA2378535A1 CA2378535A1 (en) | 2002-09-29 |
CA2378535C true CA2378535C (en) | 2005-11-22 |
Family
ID=26739342
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA002378535A Expired - Lifetime CA2378535C (en) | 2001-03-29 | 2002-03-22 | System and method for transmitting voice input from a remote location over a wireless data channel |
Country Status (5)
Country | Link |
---|---|
US (4) | US6885735B2 (en) |
EP (1) | EP1382184A1 (en) |
CA (1) | CA2378535C (en) |
MX (1) | MXPA02002811A (en) |
WO (1) | WO2002080500A1 (en) |
Families Citing this family (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6885735B2 (en) * | 2001-03-29 | 2005-04-26 | Intellisist, Llc | System and method for transmitting voice input from a remote location over a wireless data channel |
US8175886B2 (en) * | 2001-03-29 | 2012-05-08 | Intellisist, Inc. | Determination of signal-processing approach based on signal destination characteristics |
US7941313B2 (en) * | 2001-05-17 | 2011-05-10 | Qualcomm Incorporated | System and method for transmitting speech activity information ahead of speech features in a distributed voice recognition system |
US7203643B2 (en) * | 2001-06-14 | 2007-04-10 | Qualcomm Incorporated | Method and apparatus for transmitting speech activity in distributed voice recognition systems |
US7162414B2 (en) * | 2001-12-07 | 2007-01-09 | Intel Corporation | Method and apparatus to perform speech recognition over a data channel |
JP2003345391A (en) * | 2002-05-23 | 2003-12-03 | Denso Corp | Terminal, voice recognition server, voice recognition system and computer program |
US20050105699A1 (en) * | 2002-10-23 | 2005-05-19 | Fujitsu Limited | Communications terminal, voice spectrum information search server, individual information display system, individual information display method in communications terminal and individual information display program |
EP1563675B1 (en) * | 2002-11-22 | 2007-05-30 | Intellisist, Inc. | System and method for providing multi-party message-based voice communications |
GB0229781D0 (en) * | 2002-12-20 | 2003-01-29 | Courtney Peter | Secure transmission of audio signals |
US8055308B2 (en) * | 2003-09-30 | 2011-11-08 | General Motors Llc | Method and system for responding to digital vehicle requests |
US9821344B2 (en) * | 2004-12-10 | 2017-11-21 | Ikan Holdings Llc | Systems and methods for scanning information from storage area contents |
JP2006201749A (en) * | 2004-12-21 | 2006-08-03 | Matsushita Electric Ind Co Ltd | Device in which selection is activated by voice, and method in which selection is activated by voice |
US20060287863A1 (en) * | 2005-06-16 | 2006-12-21 | International Business Machines Corporation | Speaker identification and voice verification for voice applications |
US20070038868A1 (en) * | 2005-08-15 | 2007-02-15 | Top Digital Co., Ltd. | Voiceprint-lock system for electronic data |
US8417185B2 (en) * | 2005-12-16 | 2013-04-09 | Vocollect, Inc. | Wireless headset and method for robust voice data communication |
US7783488B2 (en) * | 2005-12-19 | 2010-08-24 | Nuance Communications, Inc. | Remote tracing and debugging of automatic speech recognition servers by speech reconstruction from cepstra and pitch information |
US7773767B2 (en) | 2006-02-06 | 2010-08-10 | Vocollect, Inc. | Headset terminal with rear stability strap |
US7885419B2 (en) | 2006-02-06 | 2011-02-08 | Vocollect, Inc. | Headset terminal with speech functionality |
US7848314B2 (en) * | 2006-05-10 | 2010-12-07 | Nuance Communications, Inc. | VOIP barge-in support for half-duplex DSR client on a full-duplex network |
US8635069B2 (en) | 2007-08-16 | 2014-01-21 | Crimson Corporation | Scripting support for data identifiers, voice recognition and speech in a telnet session |
EP2225870A4 (en) * | 2007-12-14 | 2011-08-17 | Promptu Systems Corp | Automatic service vehicle hailing and dispatch system and method |
US7973711B2 (en) * | 2008-05-30 | 2011-07-05 | Sony Ericsson Mobile Communications Ab | Controlling satellite navigation receivers in response to low frequency electromagnetic signals |
USD605629S1 (en) | 2008-09-29 | 2009-12-08 | Vocollect, Inc. | Headset |
US8160287B2 (en) | 2009-05-22 | 2012-04-17 | Vocollect, Inc. | Headset with adjustable headband |
US8438659B2 (en) | 2009-11-05 | 2013-05-07 | Vocollect, Inc. | Portable computing device and headset interface |
US20120253817A1 (en) * | 2011-04-04 | 2012-10-04 | Mitel Networks Corporation | Mobile speech attendant access |
US9786296B2 (en) * | 2013-07-08 | 2017-10-10 | Qualcomm Incorporated | Method and apparatus for assigning keyword model to voice operated function |
US9686692B2 (en) * | 2014-03-20 | 2017-06-20 | Microsoft Technology Licesning, LLC | Wi-Fi tile transfer |
US9986075B2 (en) | 2014-06-04 | 2018-05-29 | Qualcomm Incorporated | Mobile device including a substantially centrally located earpiece |
US10033797B1 (en) | 2014-08-20 | 2018-07-24 | Ivanti, Inc. | Terminal emulation over HTML |
US11100278B2 (en) | 2016-07-28 | 2021-08-24 | Ivanti, Inc. | Systems and methods for presentation of a terminal application screen |
WO2018164165A1 (en) * | 2017-03-10 | 2018-09-13 | 株式会社Bonx | Communication system and api server, headset, and mobile communication terminal used in communication system |
US11862169B2 (en) | 2020-09-11 | 2024-01-02 | Avaya Management L.P. | Multilingual transcription at customer endpoint for optimizing interaction results in a contact center |
Family Cites Families (150)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US405027A (en) * | 1889-06-11 | Soda-water apparatus | ||
JPS5712999B2 (en) * | 1974-04-08 | 1982-03-13 | ||
US4052568A (en) | 1976-04-23 | 1977-10-04 | Communications Satellite Corporation | Digital voice switch |
US4625083A (en) | 1985-04-02 | 1986-11-25 | Poikela Timo J | Voice operated switch |
US4630304A (en) | 1985-07-01 | 1986-12-16 | Motorola, Inc. | Automatic background noise estimator for a noise suppression system |
JPS62137454A (en) * | 1985-12-11 | 1987-06-20 | Toyota Motor Corp | Abnormality judgement device for vehicle speed sensor |
US4918732A (en) * | 1986-01-06 | 1990-04-17 | Motorola, Inc. | Frame comparison method for word recognition in high noise environments |
US4876710A (en) | 1986-10-10 | 1989-10-24 | Motorola, Inc. | Method and apparatus for cordless microphone communication system |
US4827500A (en) * | 1987-01-30 | 1989-05-02 | American Telephone And Telegraph Company, At&T Bell Laboratories | Automatic speech recognition to select among call destinations |
US4878170A (en) | 1987-03-17 | 1989-10-31 | Zeevi Eliahu I | Vehicle navigation system |
US5029233A (en) * | 1987-10-09 | 1991-07-02 | Motorola, Inc. | Radio arrangement having two radios sharing circuitry |
JPH02174448A (en) | 1988-12-27 | 1990-07-05 | Sony Corp | Telephone set |
US5259018A (en) | 1989-03-17 | 1993-11-02 | Technophone Limited | Radio telephone system having a handset adapted to be removably connected and containing a NAM for storing a telephone number identifying the handset |
US5012518A (en) * | 1989-07-26 | 1991-04-30 | Itt Corporation | Low-bit-rate speech coder using LPC data reduction processing |
US5148471A (en) * | 1989-11-20 | 1992-09-15 | Motorola, Inc. | Communications device with voice recognition and movable element control interface |
US5114624A (en) * | 1990-04-17 | 1992-05-19 | Shell Oil Company | Process for the production and recovery of ene reaction products |
US5263181A (en) | 1990-10-18 | 1993-11-16 | Motorola, Inc. | Remote transmitter for triggering a voice-operated radio |
US5357513A (en) * | 1990-12-06 | 1994-10-18 | Hughes Aircraft Company | Transmission power level adjustment in radio telephony |
WO1993001664A1 (en) * | 1991-07-08 | 1993-01-21 | Motorola, Inc. | Remote voice control system |
US5168277A (en) | 1991-11-25 | 1992-12-01 | United Technologies Corporation | Radar tracking system having target position dead reckoning |
US5539810A (en) * | 1992-01-27 | 1996-07-23 | Highwaymaster Communications, Inc. | Data messaging in a communications network |
US6279946B1 (en) * | 1998-06-09 | 2001-08-28 | Automotive Technologies International Inc. | Methods for controlling a system in a vehicle using a transmitting/receiving transducer and/or while compensating for thermal gradients |
US5945985A (en) * | 1992-10-27 | 1999-08-31 | Technology International, Inc. | Information system for interactive access to geographic information |
US5777580A (en) * | 1992-11-18 | 1998-07-07 | Trimble Navigation Limited | Vehicle location system |
US5940503A (en) * | 1993-02-02 | 1999-08-17 | Palett; Anthony P. | Mirror mounted mobile telephone system |
US5459814A (en) * | 1993-03-26 | 1995-10-17 | Hughes Aircraft Company | Voice activity detector for speech signals in variable background noise |
GB2280820A (en) * | 1993-07-29 | 1995-02-08 | Ibm | Distributed system for call processing |
JP4001643B2 (en) * | 1993-10-05 | 2007-10-31 | スナップ−オン・テクノロジイズ・インク | Two-hand open type car maintenance equipment |
US5414624A (en) * | 1993-11-08 | 1995-05-09 | Avid Systems Corporation | Automated vehicle parking system |
JP3319838B2 (en) | 1993-11-12 | 2002-09-03 | 本田技研工業株式会社 | Vehicle traction control device |
ZA948426B (en) * | 1993-12-22 | 1995-06-30 | Qualcomm Inc | Distributed voice recognition system |
US5570087A (en) | 1994-02-18 | 1996-10-29 | Lemelson; Jerome H. | Motor vehicle performance monitor and method |
AU684872B2 (en) * | 1994-03-10 | 1998-01-08 | Cable And Wireless Plc | Communication system |
US5749072A (en) * | 1994-06-03 | 1998-05-05 | Motorola Inc. | Communications device responsive to spoken commands and methods of using same |
GB9417600D0 (en) * | 1994-09-01 | 1994-10-19 | British Telecomm | Navigation information system |
US6019475A (en) * | 1994-09-30 | 2000-02-01 | Donnelly Corporation | Modular rearview mirror assembly including an electronic control module |
SE505156C2 (en) * | 1995-01-30 | 1997-07-07 | Ericsson Telefon Ab L M | Procedure for noise suppression by spectral subtraction |
US5712788A (en) * | 1995-02-09 | 1998-01-27 | Zexel Corporation | Incremental route calculation |
US5761638A (en) * | 1995-03-17 | 1998-06-02 | Us West Inc | Telephone network apparatus and method using echo delay and attenuation |
DE19516647A1 (en) | 1995-05-05 | 1996-11-07 | Bayerische Motoren Werke Ag | Navigation system for motor vehicle |
US5752186A (en) * | 1995-06-07 | 1998-05-12 | Jeman Technologies, Inc. | Access free wireless telephony fulfillment service system |
JPH0990974A (en) * | 1995-09-25 | 1997-04-04 | Nippon Telegr & Teleph Corp <Ntt> | Signal processor |
US6249233B1 (en) * | 1996-07-29 | 2001-06-19 | Baran Advanced Technologies (86) Ltd. | Vehicle parking system |
US5729531A (en) * | 1995-10-18 | 1998-03-17 | Telefonaktiebolaget Lm Ericsson | Bandwidth allocation |
US5794164A (en) * | 1995-11-29 | 1998-08-11 | Microsoft Corporation | Vehicle computer system |
AU1044197A (en) * | 1995-12-07 | 1997-06-27 | Telefonaktiebolaget Lm Ericsson (Publ) | Emergency call handling in a cellular telecommunications system |
US5732074A (en) * | 1996-01-16 | 1998-03-24 | Cellport Labs, Inc. | Mobile portable wireless communication system |
US5835881A (en) | 1996-01-16 | 1998-11-10 | Philips Electronics North America Corporation | Portable system for providing voice driving directions |
EP0880827A1 (en) * | 1996-02-07 | 1998-12-02 | L.S. Research, Inc. | Digital wireless speaker system |
US5719921A (en) * | 1996-02-29 | 1998-02-17 | Nynex Science & Technology | Methods and apparatus for activating telephone services in response to speech |
US5963861A (en) | 1996-04-05 | 1999-10-05 | Lucent Technologies Inc. | Dealer-locator service and apparatus for mobile telecommunications system |
US6516466B1 (en) * | 1996-05-02 | 2003-02-04 | Vincent C. Jackson | Method and apparatus for portable digital entertainment system |
US5870667A (en) * | 1996-06-14 | 1999-02-09 | Mediaone Group, Inc. | Support system architecture for use in a communications network environment enabling automatic provisioning, change in service and maintenance |
IL118898A (en) * | 1996-07-21 | 2004-08-31 | Shlomo Zytman | Parking management system |
JP2830914B2 (en) * | 1996-08-27 | 1998-12-02 | 日本電気株式会社 | Channel acquisition method for wireless packet data communication |
US5835868A (en) | 1996-08-30 | 1998-11-10 | Mcelroy; Alejandro S. | Automated system for immobilizing a vehicle and method |
JPH1089976A (en) | 1996-09-13 | 1998-04-10 | Hitachi Ltd | Information display and navigation system |
US6037880A (en) * | 1996-09-23 | 2000-03-14 | Manion; Jeffrey Charles | Integrated parking meter system |
JPH10108284A (en) * | 1996-09-27 | 1998-04-24 | Sony Corp | Microphone, navigation device, communication equipment, audio device and automobile |
US6157317A (en) | 1996-12-02 | 2000-12-05 | Kline And Walker Llc | Secure communication and control system for monitoring, recording, reporting and/or restricting unauthorized use of vehicle. |
US5960399A (en) * | 1996-12-24 | 1999-09-28 | Gte Internetworking Incorporated | Client/server speech processor/recognizer |
SE510864C2 (en) | 1997-01-10 | 1999-06-28 | Modul System Sweden Ab | Parking Control |
US6240365B1 (en) * | 1997-01-21 | 2001-05-29 | Frank E. Bunn | Automated vehicle tracking and service provision system |
US5876710A (en) * | 1997-02-07 | 1999-03-02 | Emisphere Technologies Inc. | Compounds and compositions for delivering active agents |
US5877723A (en) * | 1997-03-05 | 1999-03-02 | Caterpillar Inc. | System and method for determining an operating point |
US6078886A (en) * | 1997-04-14 | 2000-06-20 | At&T Corporation | System and method for providing remote automatic speech recognition services via a packet network |
GB2325110B (en) * | 1997-05-06 | 2002-10-16 | Ibm | Voice processing system |
US6076057A (en) * | 1997-05-21 | 2000-06-13 | At&T Corp | Unsupervised HMM adaptation based on speech-silence discrimination |
US6006148A (en) | 1997-06-06 | 1999-12-21 | Telxon Corporation | Automated vehicle return system |
US6011786A (en) * | 1997-06-06 | 2000-01-04 | Ericsson Inc. | Systems and methods for control channel communication in cellular radiotelephone systems |
FI972723A0 (en) * | 1997-06-24 | 1997-06-24 | Nokia Mobile Phones Ltd | Mobile communications services |
JPH1141643A (en) * | 1997-07-04 | 1999-02-12 | Internatl Business Mach Corp <Ibm> | Radio information processing terminal and control method therefor |
US6199013B1 (en) * | 1997-07-15 | 2001-03-06 | Navigation Technologies Corp. | Maneuver generation program and method |
US6275231B1 (en) * | 1997-08-01 | 2001-08-14 | American Calcar Inc. | Centralized control and management system for automobiles |
US6326613B1 (en) * | 1998-01-07 | 2001-12-04 | Donnelly Corporation | Vehicle interior mirror assembly adapted for containing a rain sensor |
DE19737256B4 (en) | 1997-08-27 | 2005-02-24 | Robert Bosch Gmbh | Vehicle guidance and guidance system |
US6105060A (en) * | 1997-09-05 | 2000-08-15 | Worldspace, Inc. | System for providing global portable internet access using low earth orbit satellite and satellite direct radio broadcast system |
JP3485239B2 (en) * | 1997-09-10 | 2004-01-13 | 富士重工業株式会社 | Vehicle motion control device |
US6052604A (en) * | 1997-10-03 | 2000-04-18 | Motorola, Inc. | Exchange which controls M SIMs and N transceivers and method therefor |
US6526486B2 (en) * | 1997-10-15 | 2003-02-25 | Nokia Mobile Phones Limited | Method of managing messages in a computer memory |
DE59809476D1 (en) | 1997-11-03 | 2003-10-09 | Volkswagen Ag | Autonomous vehicle and method for controlling an autonomous vehicle |
US6243003B1 (en) * | 1999-08-25 | 2001-06-05 | Donnelly Corporation | Accessory module for vehicle |
US6154662A (en) | 1998-02-27 | 2000-11-28 | Lucent Technologies Inc. | Providing voice dialing service for wireless roamers |
DE19815035B4 (en) | 1998-03-05 | 2006-06-22 | Volkswagen Ag | Method and device for displaying information for a navigation device |
SE513880C2 (en) * | 1998-03-17 | 2000-11-20 | Modul System Sweden Ab | Parking system for vehicles |
JP3458734B2 (en) | 1998-04-09 | 2003-10-20 | トヨタ自動車株式会社 | Vehicle motion control device |
US6484174B1 (en) | 1998-04-20 | 2002-11-19 | Sun Microsystems, Inc. | Method and apparatus for session management and user authentication |
US6206283B1 (en) * | 1998-12-23 | 2001-03-27 | At&T Corp. | Method and apparatus for transferring money via a telephone call |
JPH11355854A (en) | 1998-06-10 | 1999-12-24 | Matsushita Electric Ind Co Ltd | Digital data transmission reception method, base station device used for transmission/reception of digital data and mobile terminal used for transmission/reception of digital data |
US6779196B1 (en) * | 1998-06-22 | 2004-08-17 | Phillip Igbinadolor | Integrated car dubbing system |
AR020608A1 (en) * | 1998-07-17 | 2002-05-22 | United Video Properties Inc | A METHOD AND A PROVISION TO SUPPLY A USER REMOTE ACCESS TO AN INTERACTIVE PROGRAMMING GUIDE BY A REMOTE ACCESS LINK |
US6349329B1 (en) * | 1998-09-29 | 2002-02-19 | Radiowave.Com, Inc. | Coordinating delivery of supplemental materials with radio broadcast material |
AU1445100A (en) * | 1998-10-13 | 2000-05-01 | Hadasit Medical Research Services & Development Company Ltd | Method and system for determining a vector index to represent a plurality of speech parameters in signal processing for identifying an utterance |
US6445317B2 (en) * | 1998-11-20 | 2002-09-03 | Telefonaktiebolaget L M Ericsson (Publ) | Adaptively calibrating analog-to-digital conversion |
US6336090B1 (en) * | 1998-11-30 | 2002-01-01 | Lucent Technologies Inc. | Automatic speech/speaker recognition over digital wireless channels |
US20020111172A1 (en) * | 2001-02-14 | 2002-08-15 | Dewolf Frederik M. | Location based profiling |
US6154658A (en) | 1998-12-14 | 2000-11-28 | Lockheed Martin Corporation | Vehicle information and safety control system |
US6128482A (en) | 1998-12-22 | 2000-10-03 | General Motors Corporation | Providing mobile application services with download of speaker independent voice model |
US6370237B1 (en) * | 1998-12-29 | 2002-04-09 | Alcatel Usa Sourcing, Lp | Voice activated dialing with reduced storage requirements |
US6282430B1 (en) * | 1999-01-01 | 2001-08-28 | Motorola, Inc. | Method for obtaining control information during a communication session in a radio communication system |
US20030060211A1 (en) * | 1999-01-26 | 2003-03-27 | Vincent Chern | Location-based information retrieval system for wireless communication device |
US6487402B1 (en) | 1999-01-26 | 2002-11-26 | Qualcomm, Inc | System and method for providing access to a wireless communication service to a group of subscribers who share a set of modems |
JP4085500B2 (en) * | 1999-01-29 | 2008-05-14 | 株式会社エクォス・リサーチ | Vehicle status grasping device, agent device, and vehicle control device |
US6161071A (en) | 1999-03-12 | 2000-12-12 | Navigation Technologies Corporation | Method and system for an in-vehicle computing architecture |
US6408272B1 (en) * | 1999-04-12 | 2002-06-18 | General Magic, Inc. | Distributed voice user interface |
DE19919216C2 (en) * | 1999-04-29 | 2001-10-18 | Daimler Chrysler Ag | Information system in a vehicle |
US6292781B1 (en) * | 1999-05-28 | 2001-09-18 | Motorola | Method and apparatus for facilitating distributed speech processing in a communication system |
US6374214B1 (en) * | 1999-06-24 | 2002-04-16 | International Business Machines Corp. | Method and apparatus for excluding text phrases during re-dictation in a speech recognition system |
US6212473B1 (en) * | 1999-09-20 | 2001-04-03 | Ford Global Technologies, Inc. | Vehicle navigation system having inferred user preferences |
US6370506B1 (en) * | 1999-10-04 | 2002-04-09 | Ericsson Inc. | Communication devices, methods, and computer program products for transmitting information using voice activated signaling to perform in-call functions |
US6868385B1 (en) * | 1999-10-05 | 2005-03-15 | Yomobile, Inc. | Method and apparatus for the provision of information signals based upon speech recognition |
US6296280B1 (en) | 1999-11-02 | 2001-10-02 | Trw Inc. | Adaptive collapsible steering column |
US9076448B2 (en) * | 1999-11-12 | 2015-07-07 | Nuance Communications, Inc. | Distributed real time speech recognition system |
US6363324B1 (en) * | 1999-11-12 | 2002-03-26 | David M Hildebrant | Vehicle location system |
US20030182113A1 (en) * | 1999-11-22 | 2003-09-25 | Xuedong Huang | Distributed speech recognition for mobile communication devices |
US6405027B1 (en) * | 1999-12-08 | 2002-06-11 | Philips Electronics N.A. Corporation | Group call for a wireless mobile communication device using bluetooth |
US6628928B1 (en) | 1999-12-10 | 2003-09-30 | Ecarmerce Incorporated | Internet-based interactive radio system for use with broadcast radio stations |
US6317684B1 (en) | 1999-12-22 | 2001-11-13 | At&T Wireless Services Inc. | Method and apparatus for navigation using a portable communication device |
GB9930731D0 (en) * | 1999-12-22 | 2000-02-16 | Ibm | Voice processing apparatus |
US6526335B1 (en) * | 2000-01-24 | 2003-02-25 | G. Victor Treyz | Automobile personal computer systems |
US7099835B2 (en) * | 2000-01-31 | 2006-08-29 | Roadside Telematics Corporation | Methods and systems for providing life management and enhancement applications and services for telematics and other electronic medium |
US6415018B1 (en) * | 2000-02-08 | 2002-07-02 | Lucent Technologies Inc. | Telecommunication system and method for handling special number calls having geographic sensitivity |
US6381534B2 (en) * | 2000-02-14 | 2002-04-30 | Fujitsu Limited | Navigation information presenting apparatus and method thereof |
US20010033643A1 (en) | 2000-02-17 | 2001-10-25 | Mulvey Kevin C.W. | Telephone privacy protection system |
US6377210B1 (en) * | 2000-02-25 | 2002-04-23 | Grey Island Systems, Inc. | Automatic mobile object locator apparatus and method |
US7457628B2 (en) * | 2000-02-29 | 2008-11-25 | Smarter Agent, Llc | System and method for providing information based on geographic position |
US6496776B1 (en) | 2000-02-29 | 2002-12-17 | Brad W. Blumberg | Position-based information access device and method |
JP3934845B2 (en) | 2000-03-02 | 2007-06-20 | 株式会社東芝 | Mobile communication terminal |
JP3728177B2 (en) * | 2000-05-24 | 2005-12-21 | キヤノン株式会社 | Audio processing system, apparatus, method, and storage medium |
ES2168071B1 (en) | 2000-07-12 | 2003-07-16 | Barros Alejandro Rodriguez | MODULAR REAR VIEW MIRROR WITH INTERCHANGEABLE MULTIPLE SIGNALS FOR VEHICLES OF 2, 3, 4 OR MORE WHEELS. |
US20020028655A1 (en) * | 2000-07-14 | 2002-03-07 | Rosener Douglas K. | Repeater system |
DE60133902D1 (en) * | 2000-07-28 | 2008-06-19 | Siemens Vdo Automotive Corp | |
US20020140545A1 (en) | 2000-08-18 | 2002-10-03 | Peter Nietupski | Integrated RKE and telematics system |
US6421607B1 (en) * | 2000-09-22 | 2002-07-16 | Motorola, Inc. | System and method for distributed navigation service |
US6810323B1 (en) | 2000-09-25 | 2004-10-26 | Motorola, Inc. | System and method for storing and using information associated with geographic locations of interest to a mobile user |
US7457608B2 (en) * | 2000-10-03 | 2008-11-25 | At&T Mobility Ii Llc | Sharing of wireless telephone services for a personal wireless telephone and a vehicular wireless telephone |
US7224935B2 (en) * | 2000-11-29 | 2007-05-29 | Visteon Global Technologies, Inc. | Telematics application for implementation in conjunction with a satellite broadcast delivery system |
US6823306B2 (en) * | 2000-11-30 | 2004-11-23 | Telesector Resources Group, Inc. | Methods and apparatus for generating, updating and distributing speech recognition models |
US6765497B2 (en) * | 2000-12-18 | 2004-07-20 | Motorola, Inc. | Method for remotely accessing vehicle system information and user information in a vehicle |
US7024359B2 (en) * | 2001-01-31 | 2006-04-04 | Qualcomm Incorporated | Distributed voice recognition system using acoustic feature vector modification |
US6344806B1 (en) * | 2001-02-15 | 2002-02-05 | Yoram Katz | Parking status control system and method |
US7277853B1 (en) * | 2001-03-02 | 2007-10-02 | Mindspeed Technologies, Inc. | System and method for a endpoint detection of speech for improved speech recognition in noisy environments |
US6487494B2 (en) | 2001-03-29 | 2002-11-26 | Wingcast, Llc | System and method for reducing the amount of repetitive data sent by a server to a client for vehicle navigation |
US6885735B2 (en) | 2001-03-29 | 2005-04-26 | Intellisist, Llc | System and method for transmitting voice input from a remote location over a wireless data channel |
US7096269B2 (en) | 2001-03-30 | 2006-08-22 | Hitachi, Ltd. | Path selection methods for storage based remote copy |
US6895238B2 (en) * | 2001-03-30 | 2005-05-17 | Motorola, Inc. | Method for providing entertainment to a portable device |
US6687517B2 (en) | 2001-05-16 | 2004-02-03 | Nokia Corporation | Hands-free operation of mobile terminal using wireless communication link |
US6441790B1 (en) * | 2001-06-14 | 2002-08-27 | Kyocera Wireless Corp. | System and method for providing a quasi-isotropic antenna |
US7203643B2 (en) * | 2001-06-14 | 2007-04-10 | Qualcomm Incorporated | Method and apparatus for transmitting speech activity in distributed voice recognition systems |
US7062444B2 (en) * | 2002-01-24 | 2006-06-13 | Intel Corporation | Architecture for DSR client and server development platform |
-
2002
- 2002-01-29 US US10/059,905 patent/US6885735B2/en not_active Expired - Lifetime
- 2002-03-14 MX MXPA02002811A patent/MXPA02002811A/en active IP Right Grant
- 2002-03-22 CA CA002378535A patent/CA2378535C/en not_active Expired - Lifetime
- 2002-03-27 WO PCT/US2002/009422 patent/WO2002080500A1/en not_active Application Discontinuation
- 2002-03-27 EP EP02725373A patent/EP1382184A1/en not_active Withdrawn
-
2004
- 2004-12-22 US US11/020,596 patent/US7634064B2/en active Active
-
2007
- 2007-10-30 US US11/929,458 patent/US7769143B2/en not_active Expired - Fee Related
-
2010
- 2010-07-02 US US12/829,820 patent/US8379802B2/en active Active
Also Published As
Publication number | Publication date |
---|---|
EP1382184A1 (en) | 2004-01-21 |
US7769143B2 (en) | 2010-08-03 |
US8379802B2 (en) | 2013-02-19 |
CA2378535A1 (en) | 2002-09-29 |
US20100274562A1 (en) | 2010-10-28 |
US20050119895A1 (en) | 2005-06-02 |
MXPA02002811A (en) | 2005-07-13 |
US6885735B2 (en) | 2005-04-26 |
US7634064B2 (en) | 2009-12-15 |
US20080140419A1 (en) | 2008-06-12 |
US20020141547A1 (en) | 2002-10-03 |
WO2002080500A1 (en) | 2002-10-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2378535C (en) | System and method for transmitting voice input from a remote location over a wireless data channel | |
US6937977B2 (en) | Method and apparatus for processing an input speech signal during presentation of an output audio signal | |
JP5425945B2 (en) | Speech recognition technology based on local interrupt detection | |
USRE45066E1 (en) | Method and apparatus for the provision of information signals based upon speech recognition | |
CA2717992C (en) | Speech understanding method and system | |
US7356471B2 (en) | Adjusting sound characteristic of a communication network using test signal prior to providing communication to speech recognition server | |
US6882973B1 (en) | Speech recognition system with barge-in capability | |
CN103124318B (en) | Start the method for public conference calling | |
US6941265B2 (en) | Voice recognition system method and apparatus | |
CN102543077B (en) | Male acoustic model adaptation method based on language-independent female speech data | |
US7328159B2 (en) | Interactive speech recognition apparatus and method with conditioned voice prompts | |
US11516599B2 (en) | Personal hearing device, external acoustic processing device and associated computer program product | |
CN102623006A (en) | Mapping obstruent speech energy to lower frequencies | |
US20090138264A1 (en) | Speech to dtmf generation | |
US20070129037A1 (en) | Mute processing apparatus and method | |
WO2007091462A1 (en) | Voice recognizing apparatus, voice recognizing method and program for recognizing voice | |
US20070133589A1 (en) | Mute processing apparatus and method | |
KR20210031186A (en) | Speech preprocessing using auxiliary data for home automation | |
JP2003046647A (en) | System, method and program for repeating call and recording medium with the program recorded thereon |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request | ||
MKEX | Expiry |
Effective date: 20220322 |