WO2006126183A2 - Method and apparatus for fraud detection - Google Patents

Method and apparatus for fraud detection Download PDF

Info

Publication number
WO2006126183A2
WO2006126183A2 PCT/IL2006/000110 IL2006000110W WO2006126183A2 WO 2006126183 A2 WO2006126183 A2 WO 2006126183A2 IL 2006000110 W IL2006000110 W IL 2006000110W WO 2006126183 A2 WO2006126183 A2 WO 2006126183A2
Authority
WO
WIPO (PCT)
Prior art keywords
voice
fraud
interaction
probability
speaker
Prior art date
Application number
PCT/IL2006/000110
Other languages
French (fr)
Other versions
WO2006126183A3 (en
Inventor
Moshe Waserblat
Barak Eilam
Oren Pereg
Ilan Kor
Original Assignee
Nice Systems Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nice Systems Ltd. filed Critical Nice Systems Ltd.
Publication of WO2006126183A2 publication Critical patent/WO2006126183A2/en
Publication of WO2006126183A3 publication Critical patent/WO2006126183A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/22Arrangements for supervision, monitoring or testing
    • H04M3/36Statistical metering, e.g. recording occasions when traffic exceeds capacity of trunks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q20/00Payment architectures, schemes or protocols
    • G06Q20/38Payment protocols; Details thereof
    • G06Q20/40Authorisation, e.g. identification of payer or payee, verification of customer or shop credentials; Review and approval of payers, e.g. check credit lines or negative lists
    • G06Q20/401Transaction verification
    • G06Q20/4016Transaction verification involving fraud or risk level assessment in transaction processing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • G10L17/26Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/38Graded-service arrangements, i.e. some subscribers prevented from establishing certain connections
    • H04M3/382Graded-service arrangements, i.e. some subscribers prevented from establishing certain connections using authorisation codes or passwords
    • H04M3/385Graded-service arrangements, i.e. some subscribers prevented from establishing certain connections using authorisation codes or passwords using speech signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2201/00Electronic components, circuits, software, systems or apparatus used in telephone systems
    • H04M2201/41Electronic components, circuits, software, systems or apparatus used in telephone systems using speaker recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/60Aspects of automatic or semi-automatic exchanges related to security aspects in telephonic communication systems
    • H04M2203/6027Fraud preventions

Definitions

  • the present invention relates to voice analysis in general and to a method and apparatus for fraud detection in particular.
  • Common identity theft actions include using false identity when consuming phone or wireless services, using a stolen name, date of birth, or social security number to open a new credit card account, and opening a checking account under a stolen identity.
  • Identity thefts account for 3.23 millions cases per year with an average damage of USD 10,200 per victim.
  • Around 60% of all fraud actions involving existing accounts and credit cards take one week to over six months to be discovered. This long time frame, which is mainly due to the long billing cycle, enables perpetrators to commit multiple fraud actions before suspicion arises.
  • a fraud detection method for generating a first fraud or fraud attempt probability, within one or more captured or recorded interactions, the method comprising the step of scoring one or more voices belonging to one or more tested speakers in the captured or recorded interaction against one or more voice prints within one or more entries in a voice print collection, the scoring step generating a probability that the one or more voices in the captured or recorded interaction belong to one or more second speakers associated with the one or more voice prints, said probability represents the probability that the one or more captured or recorded interaction is fraudulent.
  • the probability generated at the scoring step represents the similarity between the voice of the tested speaker and the voice of the at least one second speaker, said second speaker can be a fraudster, a legitimate speaker, or allegedly the tested speaker.
  • the method further comprising a fraud probability generation step, for enhancing the first fraud or fraud attempt probability.
  • the fraud probability generation step employs a rule engine activating one or more rules or using one or more thresholds set by a user of the method.
  • the rules or the thresholds can involve data associated with: the captured or recorded interaction, a speaker thereof, an at least one other interaction of the speaker thereof, the one or more voice prints or the one or more second speakers.
  • the rules or the thresholds can be set dynamically.
  • the method can further comprise: a second scoring step for determining a second scoring result by scoring the one or more voices against a background model; and a normalizing step for normalizing the first scoring result with the second scoring result.
  • the scoring step can comprise scoring the one or more tested voices against two or more voice prints, thus obtaining two or more score results.
  • the method can further comprise a score-test normalization step, for normalizing the at least two score results.
  • one or more voice prints or data associated with the voice prints or with the second speaker is stored in a collection.
  • the method can further comprise a retrieving step for retrieving one or more second interactions, and a construction step for constructing the at one or more voice prints from the one or more voices participating in the one or more second interactions.
  • the method can further comprise a step of retrieving one or more voice prints or data associated with the one or more voice prints.
  • the method can further comprise a filtering step for filtering a collection to retrieve the one or more voice prints such that data associated with the voice prints is related to data associated with the recorded or captured interaction.
  • the filtering can be based on an at least one of: area code of calling number, gender of speaker, age pf speaker, language spoken.
  • the step so of the method can be performed in real-time.
  • the method can further comprising a pre processing step for pre processing the captured or recorded interaction or the second interaction.
  • the pre processing step can comprise one of: compressing; decompressing; segmenting; separating speakers; identifying the agent side; and receiving data related to the captured or recorded interaction or to the second interaction.
  • the additional data can comprise computer telephony integration information.
  • the method can further comprise a selection step for selecting one or more selected interactions to be transferred to the scoring step, and associating the one or more selected interactions with one or more second fraud or fraud attempt probabilities.
  • the selection step can comprise one or more of the following: applying one or more rules by a rule engine; auditing interactions; filtering interactions using one or more filtering engines.
  • the filtering engines can comprise one or more of the following: emotion detection, word spotting, speech to text, interaction analysis, and data risk analysis.
  • the second fraud or fraud attempt probability represents a risk level associated with the one or more selected interactions.
  • the selection step can employs one or more rules or one or more thresholds set by a user of the method.
  • the rules or thresholds can involve data associated with one or more of the following: the captured or recorded interaction; a speaker thereof; other interactions of the speaker thereof; the one or more voice prints or the at least one second speaker.
  • the one or more rules or one or more thresholds can set dynamically.
  • the method can further comprise a weighting step for generating a combined fraud attempt probability from the first and the second fraud attempt probabilities.
  • the method can further comprise an alert-generation step for generating an alert if said first fraud or fraud attempt probability or said second fraud or fraud attempt probability or a combination thereof exceeds a predetermined threshold.
  • the alert can be sent to an agent or another contact associated with the captured or recorded interaction.
  • the method can further comprise a reporting step for generating or updating a report comprising details of the one or more captured or recorded interactions if the first fraud or fraud attempt probability or the second fraud or fraud attempt probability or a combination thereof exceeds a predetermined threshold.
  • the method can further comprise an auditing step for auditing the report.
  • the auditing step can comprise one or more of the following: listening to calls; reviewing interactions; reviewing history of one or more speakers participating in one or more interactions; listening to previous interactions of speakers participating in one or more interactions appearing in the report, or filling forms.
  • the method can further comprise a step of interaction capturing or logging.
  • the method can further comprise a step of a-posteriori analysis for one or more of the following: adapting one or more parameters or one or more rules or one or more thresholds used by the method, trend analysis, pattern of behavior recognition, selecting one or more filtering engines.
  • Another aspect of the disclosed invention relates to a method for constructing a voice print collection, the method comprising the steps of: receiving one or more suspect voice samples; extracting characteristic features from the one or more voice sample, thus generating one or more sequences of feature vectors; constructing a voice print from the one or more sequences of feature vectors; and storing the voice print in a collection.
  • the method can further comprise a storage step for storing data in the collection, said data associated with the at least one voice print, or with the at least one voice sample, or with an at least one speaker participating in the voice sample.
  • the method can further comprise a normalization step for generating score normalization parameters to be used when scoring one or more voice samples against one or more voice prints from the collection
  • Yet another aspect of the disclosed invention relates to a fraud detection apparatus for generating a fraud or fraud attempt probability, within one or more captured or recorded interactions
  • the apparatus comprising: a voice print collection comprising one or more entries, each entry comprising a voice print or data associated with the voice print, said voice print is constructed from a voice known or suspected as belonging to a fraudster; and a scoring component for scoring one or more voices in the captured or recorded interaction against the voice print within the one or more entries, the component generating a probability that the one or more voices in the one or more captured or recorded interaction belong to one or more persons associated with the one or more voice prints, said probability represents the probability that the one or more captured or recorded interaction are fraudulent.
  • Fig. 1 is a schematic block diagram of a typical environment in which a preferred embodiment the present invention is used;
  • Fig. 2 is a schematic flowchart of the fraud detection method, in accordance with the present invention.
  • Fig. 3 is a schematic block diagram of the selection system and the fraud detection system, in accordance with the present invention.
  • Fig. 4 is a schematic flow chart of the steps in building a fraudsters database, in accordance with the present invention.
  • Fig. 5 is a schematic block diagram of the pre processing component of the fraud detection system, in accordance with the present invention.
  • the present invention relates to international patent application serial number PCT/IL03/00300 titled APPARATUS, SYSTEM AND METHOD FOR DISPUTE RESOLUTION, REGULATION COMPLIANCE AND QUALITY MANAGEMENT IN FINANCIAL INSTITUTIONS, filed on April 9, 2003, and to United States patent application titled APPARATUS AND METHOD FOR AUDIO ANALYSIS filed on March 17, 2005, and to United States patent application serial number 10/996,811 titled METHOD AND APPARATUS FOR SPEAKER SPOTTING filed on November 23, 2004, and to international patent application serial number PCT/IL02/00741 titled QUALITY MANAGEMENT AND RECORDING SOLUTIONS FOR WALK-IN ENVIRONMENTS, filed on May 9, 2002, the full contents of which are incorporated herein by reference.
  • the present invention overcomes the disadvantages of the prior art by providing a novel method and apparatus which enhance and add to the capabilities of currently available systems for fraud detection in general and identity theft in particular.
  • Identity theft relates to actions in which the fraudsters is impersonating to be another person, by stating details of the other person, such as ID number, social security number, date of birth or the like, by presenting stolen or false documents, or any other method.
  • the disclosed invention provides an apparatus and method for decreasing the time between the execution of an identity theft and its detection, by identifying interactions executed by known fraudsters.
  • the present invention uses data collected prior to, during, or subsequent to the occurrence of interactions in which the participants' voices are captured and possibly logged.
  • the used data includes procedural data, such as the essence of the transaction, date, time, location if available, number called or number called from in the case of a telephonic interaction, and the like.
  • the interaction is captured and possibly logged, so that the voices of the persons participating in the interaction are available.
  • the interaction or the accompanying technical information are optionally processed in order to collect fraudulent calls, i.e. calls which might indicate a fraud attempt.
  • the processing can be manually executed by a person such as a compliance officer considering the information automatically, by considering rules, meta data associated with the interactions or results of engines processing the interactions, or a combination of automatic and manual processing.
  • the suspected calls are assigned a first fraud probability, or a risk level, and are then passed to a second stage performed by a fraud detection component.
  • the system can be tuned to transfer to the fraud detection component only calls that deal with account opening, calls that deal with address changing, calls in which a substantial degree of emotion was detected, calls with specific CTI events, such as multiple hold events, DNIS, VDN, ANI, number dialed, number dialed from or the like.
  • the fraud detection component performs voice parameterization of the voice, i.e. extracts the characteristics of the voice. If possible, the fraud detection component then preferably scores the voice characteristics against a voice print which was constructed from the voice characteristics of the same alleged customer, for example one of the account owners in a financial organization, or a person who identified himself through PIN, SSN or any other external or internal data as a legitimate speaker, for verification purposes.
  • the result of the scoring represents the probability that the customer's voice sample belongs to the voice print against which it was scored.
  • the scoring preferably includes a normalization stage, at which the score is compared to the result of scoring the voice characteristics against a general voice print.
  • the general voice model is a background model, constructed from multiple relevant voices, such as customers' voices, voices of the same gender, or the like. If the score is below a certain threshold, there is a significant probability that the person involved in the interaction is not the same person for which an earlier interaction was recorded. In such case, the voice characteristics are scored against a reservoir of voice prints of known fraudsters. The resulting score represents the probability that the customer's voice sample belongs to any of the fraudster's voice print.
  • the customer's voice sample is scored against one or more voice prints, which are selected from the fraudster collection as having similar characteristics with the customer's voice sample at hand.
  • the characteristics can include characteristics stemming from the capturing or the recording systems, such as callers details, CTI information and other tagged business data, as well as characteristics obtained from external sources, such as account number, area, age or the like, and characteristics obtained from the voice itself, such as gender.
  • a combination of the first fraud probability and the score between the customer's voice sample and the fraudster's voice print contained in the reservoir exceeds a predetermined threshold a high probability that the speaker in the interaction is a known fraudster, and that the interaction is fraudulent is assumed, and a notification, such as an alert or a report is issued for the interaction and steps are taken to control or contain the damage caused by the fraudster.
  • a supervisor or another person is preferably prompted to audit reported interactions.
  • the system preferably further grades the probabilities resulting from scoring a customer's voice sample against the one or more voice prints in the reservoir.
  • the threshold above which a notification is issued for an interaction depends on the resources the user is willing to dedicate to checking the alerts on one hand, and on the implications of unhandled identity thefts on the other hand.
  • Other factors considered in setting the thresholds include considering the status of the customer, the status of the interaction, the importance or the volume of the interaction, the sensitivity of the interaction to the organization, the relative frequency of interactions of the same order of magnitude in the account as well as variations in the same, the sensitivity of the account or the account holder, or the like. For example, a large transaction of a VIP customer in a financial institution will be assigned lower thresholds than other transactions, in order to increase the catching probability of a fraud attempt.
  • the thresholds are preferably set in a dynamic manner according to interaction metadata and data collected from the interaction itself.
  • the fields, as well as the values, according to which the thresholds are set can be determined dynamically. Testing a voice sample against fraudsters voice prints is extremely fast, therefore the result, i.e. the fraud attempt probability is possibly available shortly after an interaction was executed, or even before it ended, thus limiting further damage by closing the credit line, canceling the account or the like.
  • the thresholds determination rules and values above which calls are transferred to the scoring system are determined by the site's requirements and available resources. Either at the first stage or at the second stage, the voice is preprocessed.
  • Preprocessing optionally includes any one or more of the following: decompression in case the voice is compressed; removal of non-speech segments using an automatic algorithm for speech/non-speech detector or other indications like CTI information such as hold events; separating the speakers; and selecting one of the sides if the two (or more) speakers were captured together, for example the customer side in a telephone call between a customer and a service center.
  • the apparatus and method are efficient, since they enable users to focus on those interactions that pose the greatest threats to their business, for example opening a new account, changing details of an existing account, expanding account credit balance, upgrading account, or the like.
  • the disclosed system and apparatus can be operated in conjunction with a capturing or logging system, or as an independent solution receiving calls from a third party.
  • a typical user of the invention is a single- or multiple-site organization, such as a bank, a credit card operator, a chain of department stores having their own credit cards, a call-center, an emergency center, an insurance claims operator, or any other organization that performs financial transactions with customers or is likely to have to verify incoming callers' identity.
  • the number of sites used in the present invention is not limited.
  • the system typically employs a voice and data capturing and logging component at each site, responsible for collecting voice and optionally additional data relating to customer interactions from all relevant sources.
  • site 1 10 and site N 11 are typical sites of a multi-site organization, for example a bank branch of a national bank.
  • each site comprises the input sources and input capturing mechanisms relevant to the activities supported by the site.
  • the possible sources employed at site 1 10 include telephone 12, which is a means by which a substantial number of transactions are performed daily, many of which are also captured.
  • Another source ' of interaction voices are voice-over-IP systems 16 which are taking growing market segments.
  • An additional source is a recording of a vocal chat performed over the internet 20.
  • Yet other sources of vocal transactions are captured interactions performed at walk-in-canters 24, such as recordings performed at a teller's counter at a bank, and recordings from trading systems 26.
  • Additional sources of vocal include microphones, intercom, vocal input by an external system, or the like.
  • the sources of vocal information are not limited to the abovementioned sources, and any additional source of vocal information 28 which can be captured, can be used.
  • Additional data 30 is captured, comprising for example PABX or CTI data including claimed identity of customer, number called or number called from in the case of telephonic interaction, handling agent, hold periods or transfers, business data such as account number or claim number, or the like.
  • Another example for additional data 30 is screen events, containing information such as the interaction subject as entered by the handling agent, image or video data, and products of processing the same, such as recognized face or the like. All the captured voice and data are logged by logging component 32.
  • the capturing and logging mechanisms are specific for each type of source, and are described for example in international patent application serial number PCT/IL02/00741 titled QUALITY MANAGEMENT AND RECORDING SOLUTIONS FOR WALK-IN ENVIRONMENTS, filed on May 9, 2002, and international patent application serial number PCT/IL03/00300 titled APPARATUS, SYSTEM AND METHOD FOR DISPUTE RESOLUTION, REGULATION COMPLIANCE AND QUALITY MANAGEMENT IN FINANCIAL INSTITUTIONS, filed on April 9, 2003.
  • Logging component 32 provides a device for recording the interactions on a temporary or permanent media, such as a memory, a magnetic tape, a magnetic disc, an optical disc, a laser disc, a mass-storage device, or the like.
  • Each site further comprises a selection system 36 for assigning a fraud probability to each interaction, and selecting those interactions that are more suspected as contain fraud action or fraud attempt than others, i.e., their fraud probability exceeds a predetermined threshold. For those selected interactions, there is often a need to identify at least one of the speakers. For example, in a call center that speaker would usually be the customer. If there is a significant probability that the tested speaker is a known fraudster, that will have a large contribution to the ability to contain or control possible damages.
  • selection system 36 can serve multiple sites and select suspicious calls from all sites connected to it. Selection system 36 is further detailed in association with Fig. 3 below. Once the suspicious calls had been selected from the sites, the calls with their associated fraud probabilities are transferred to fraud detection system 40.
  • each organization comprises one ore more fraud detection system 40. The number of employed selection systems 36 and fraud detection systems 40 determines the capacity of the apparatus.
  • Fraud detection system 40 comprises a fraud detection component 44 that checks each interaction transferred by any of selection systems 36 and issues a combined fraud probability for the interaction, taking into account the results generated by fraud detection component 44 and the fraud probability as assigned by selection system 36.
  • Fraud detection component 44 further comprises a threshold determination component 46 for controlling the tradeoff between false alarms and miss detected in the system. The higher the value the fraud threshold is set to, the fewer false alarms occur, i.e., less innocent cases reported as fraud attempt are likely to occur, while the system is more likely to miss real fraudsters. However, if the threshold is set too low, more fraud attempts are likely to be discovered, together with more false alarms. The preference of a customer is affected by the price of the resources required to handle false alarms, vs.
  • the thresholds are configurable by each customer, and can even be set differently for specific cases.
  • the setting can be dynamic, static, or a combination thereof.
  • the system can be set such that the threshold relevant for opening an account is lower than the threshold relevant for other transactions, thus causing more false alarms but also higher detection rate for opening accounts.
  • the threshold can also be automatically set per transaction or interaction according to the transaction business value, account or called number sensitivity, alerts on the account or the called number, the account holder service level, account holder credit history or any other underlying rule, guideline or policy the user wishes to configure the system according to.
  • a notification component 48 such as a report or user interface generator is responsible for outputting the combined results, either by a report presented or sent to a user of all interactions whose combined probability exceeds a predetermined threshold, an alert generated once the combined fraud probability associated with an interaction exceeds a certain threshold, or the like.
  • the generated alert is sent to the source of the call or an associated authority, such as the agent who held the call or a supervisor.
  • the alert can be presented to the agent while the call is still being held, and prompt him to take precautions.
  • the user for example a supervisor, is provided with the possibility to audit the report, for example by listening to calls appearing in the report, assessing their status, receiving additional information such as the call history of the suspect or other calls relating to the same account, listening to previous interactions of speakers participating in an interaction appearing in the report, filling forms, entering comments, escalating findings or the like.
  • the procedure of defining the audit process is dynamic and can be set by a system administrator, security officer or the like in each organization.
  • All components of the system including capturing components 12, 16, 20, 24, and 28, selection systems 36, fraud detection system 44 and report/user interface component 48 preferably comprise one or more computing platforms, such as a personal computer, a mainframe computer, or any other type of computing platform that is provisioned with a memory device (not shown), a CPU or microprocessor device, and several I/O ports (not shown).
  • each component can be a DSP chip, an ASIC device storing the commands and data necessary to execute the methods of the present invention, or the like.
  • Each component can further include a storage device (not shown), storing the relevant applications.
  • Each application is a set of logically inter-related computer programs and associated data structures that interact to perform one or more specific tasks.
  • Fig. 2 showing a preferred embodiment of the method of detecting a fraud or a fraud attempt captured in an interaction.
  • the described interaction is a telephone conversation between an agent and a customer.
  • the method can be adapted to any other relevant domain with little or no enhancements.
  • all or part of the interactions associated with a customer site such as a call center, a service center, a financial institute or the like, are captured for analysis and preferably recorded for later retrieval.
  • specific interactions are selected as being suspected for capturing or conveying fraud actions or fraud attempts.
  • the selection process optionally involves activating a rule engine, filtering engines or manual selection of calls through the usage of an auditing system.
  • the filtering engines can comprise emotion detection, word spotting, speech to text, interaction analysis, data risk analysis, video analysis, image analysis or the like.
  • the selection process is further detailed in association with Fig. 3 below.
  • the interactions selected at selection step 54 for further analysis are optionally assigned with a fraud probability determined by the tools involved in the selection process.
  • pre processing is performed on the selected interactions, for enhancing their quality for further analysis.
  • the preprocessing is performed prior to, or as a part of selection step 54.
  • the preprocessing step preferably includes decompressing the interaction, speech/non- speech segment detection, segmenting the interaction, separating the speakers within the interaction, and identifying the sides of the interaction, i.e. determining which side is the agent and which is the customer.
  • the preprocessing preferably uses additional data related to the call, such as Computer Telephony Integration (CTI) information.
  • CTI Computer Telephony Integration
  • the preprocessing step is further detailed in association with Fig. 5 below.
  • the voice of the tested speaker is parameterized, by constructing a sequence of feature vectors, wherein each feature vector relates to a certain point in time, from the enhanced voice, wherein, each feature vector comprises a plurality of characteristics of the voice during a specific time frame within the interaction.
  • one or more previously constructed voice prints are selected from a collection, or a reservoir, such that the parameterized voice sample of the tested speaker would be scored against these voice prints.
  • the selected voice prints can include a voice print of the alleged speaker of the current voice sample, who is a legitimate speaker, in order to verify that the alleged customer is indeed the true customer as recorded, for example at the time the account was opened.
  • the selected voice prints can include voice prints, such as voice prints from a fraudster's reservoir, whose associated data is related to data associated with the voice sample of the current interaction, such as voice prints of the same gender, same area code or the like.
  • the voice can be scored against multiple voice prints, including voice prints of the alleged speaker or speakers, voice print of one or more fraudsters, or any combination in any required order of the above.
  • the system allows the user to set the policy rules regarding the data to be considered from each interaction, and the filtering criteria for voice prints in the voice collection database, in a completely dynamic manner.
  • the selection can be set manually according to policy rules or automatically.
  • the information used for selecting the voice prints can relate to any meta data associated with the interaction or data extracted from the interaction itself, such as spotted words, emotion levels or the like. In yet another preferred embodiment, all voice prints in the reservoir are selected.
  • the time sequence feature vector extracted from the voice sample is scored against the one or more selected voice prints selected at step 64.
  • the system assesses the probability that the caller voice sample matches the previous voice print stored in the system, thus verifying that the caller is the same caller identified in previous calls, and thus the fraud attempt probability is significantly reduced.
  • a new voice print is constructed on the fly from retrieved recorded interactions of the alleged caller. The scoring process is further detailed in association with Fig. 3 below.
  • the parameterized voice sample can first be scored against a voice print of the same alleged customer, and only if the score result is low, i.e., there is a high probability that it is not the same speaker, the voice is scored against part or all of the voices in the collection, which belong to known fraudsters.
  • the voice prints in the selection against which the voice is scored are preferably filtered according to data associated with the interaction, such as area code of calling number, gender, age, language or the like.
  • the results of the scoring between the voice sample and the one or more voice prints from the fraudster collection are enhanced.
  • the results are possibly vertically normalized, and graded or otherwise combined at sub-step 72 into a single fraud probability.
  • step 70 The functionality of step 70 is detailed in association with Fig. 3 below.
  • the fraud probability generated at step 70 is weighted or otherwise combined with the fraud probability assigned to the interaction at selection step 54, if selection step 54 was performed, thus generating a combined fraud probability. If the combined fraud probability exceeds a predetermined threshold, an alert is generated at step 78 for the interaction and preferably sent to the agent handling the interaction, a supervisor, a fraud researcher, a compliance officer or any other contact as defined in the notification rules of the system.
  • the notification policy including the thresholds and the action to be taken can be set according to the interaction sensitivity, transaction value, risk grade, fraud confidence level, structural organization, or any other notification policy that the user wishes to set.
  • an alert can be generated even before the interaction is completed, thus giving the agent or supervisor an option to control and reduce possible damages.
  • a report is generated for all interactions assigned with a combined fraud probability exceeding a predetermined threshold, and a supervisor or another person associated with the call center is prompted to audit the report. Auditing the report comprises listening to calls, reviewing interactions, reviewing history of speakers participating in an interaction, listening to previous calls of speakers participating in the interaction, filling forms or the like.
  • the structure of the form and the integrated audit procedure can be set dynamically by a site administrator, a compliance officer or the like.
  • Step 79 is preferably performed after a predetermined time during which the system was working and assigning fraud probabilities to interactions.
  • a- posteriori analysis step 79 the system analyses and tries to find common denominators to calls that were analyzed and assigned a high probability of conveying or capturing fraud attempts, or calls that were reported by a user as capturing fraud attempts, or the like.
  • the purpose of the analysis is fine tuning of the parameters, rules and thresholds used throughout the method.
  • Another product of the a-posteriori analysis step can be analyzing trends, obtaining insights regarding fraud methods, recurring vocabulary, time of calls, prosody patterns, calling scenario, and the like.
  • Fig. 3 is a schematic block diagram of a preferred embodiment of the proposed apparatus.
  • All components of the proposed apparatus can analyze a call in real-time or near real-time, thus providing a fraud probability for a call while the call is still on-going or shortly after.
  • the fraud probability especially if exceeding a predetermined threshold can be transferred back to the agent, to a supervisor, or to any other entity which can practice precautions.
  • An interactions and data capturing or logging component 104 is responsible for capturing part or all of the interactions taking place in one or more sites, for example a service center, a financial institute such as a bank, an emergency center, an insurance organization or the like, and associated data such as CTI information, images, video and data associated thereof or the like.
  • Selection system 106 optionally comprises an auditing component 136, which enables a user, such as a compliance officer to listen to calls and mark selected calls. Auditing component 136 provides the user with appropriate user- interface and tools to select calls for listening, to assign a score to a call or to route a call to further processing. Additionally, auditing component 136 enables a user to fill up a customized form for the auditing process, including for example a conversation topic, a status such as new, confirmed, rejected, in process, or the like.
  • Selection system 106 further comprises a rule engine 108, which marks calls that might be suitable for further examination, according to predetermined rules, such as call subject (if available), number called from, or any other data item or combination of data items associated with the call.
  • engine refers to one or more computerized components, that interact to receive input of predetermined types, perform a specific analysis on the input, and return one or more results.
  • An engine can comprise software components, firmware, i.e., DSP components, hardware components or any combination thereof.
  • Filtering engines 112 comprise one or more engines, for detecting information embedded within the voice itself, which can offer further information related to the risk level associated with the interaction or with the voice.
  • filtering engines 112 preferably comprises a pre processing engine 116, designed to enhance the captured voice and prepare the captured voice for usage by other engines.
  • Pre processing engine 116 is further detailed in association with Fig. 5 below.
  • Filtering engines 112 further comprise a word sporting (WS) engine 120, which searches the voice for occurrences of words or phrases from a predefined list. Engine 120 preferably supplies for each word found, the position within the call at which it was found, and a certainty indicator for the correctness of the finding.
  • WS word sporting
  • WS engine 120 can search for the phrase "open an account", and report that the phrase was located 20 seconds form the beginning of the conversation, with certainty of 80%.
  • an emotion detection engine 124 is designed to locate highly emotional segments within the conversation. Each segment is associated with an emotion level, in the range, for example of 0-100. Highly emotional segment can suggest anger on at least one side of the conversation. Alternatively it can suggest tension due to the customer practicing a fraud attempt.
  • Additional engines 132 can comprise for example a risk analysis engine dete ⁇ riining the risk associated with an interaction based on additional data, such as CTI information, screen information or the like. Additional engines 132 can also comprise a speech to text engine, a text analysis engine, screen analyzed events, video analyzed scene, OCR engines, additional engines or any combination of the above. Additional engines can be based on CTI information such as CTI call flow analysis including number of holds, number of transfers, queue time or the like. Post processing engine 118 can further enhance the results generated by other engines, for example by integrating results supplied by different engines, such as correlating certain spotted words with segments of high emotion and the like.
  • filtering engines 112 can work in parallel, sequentially, be skipped, be activated depending on the results generated by other engines, be activated by a user, or the like. However, it is preferred that pre-processing engine 116, if present, is activated prior to other engines, and post processing engine 138 is activated consequent to the other engines.
  • the processing components, i.e., auditing component 136, rule engine 108 and filtering engines 112 are preferably not mutual exclusive. Rather, they can cooperate and work in sequence. Depending on the engines' input requirement, an engine may use the results of other engines. For example, the user can use auditing component 136 to listen to calls selected by rule engine 108.
  • rule engine 108 can apply rules which involve results of word spotting engine 120, such as "select all calls in which the phrase "open an account” was spotted, and the customer's first name is John".
  • the goal of the user is to set the rules, the activation order, and the thresholds so that the calls being selected are the most suspicious as being fraud attempts.
  • Risk analysis component 139 combines the output results of auditing component 136, rule engine 108 and filtering engines 112 according to predetermined rules, weights and priorities, and assigns to each call a single risk level estimating the probability that the call is a fraud attempt.
  • the calls whose risk factor exceeds a predetermined threshold, and the data acquired via the usage of the components of selection system 36 are transferred to fraud detection component 140.
  • the processing power of selection system 36 and of the fraud detection system 140 has substantial influence on the selection of the components and the thresholds. If, for example the processing power of fraud detection component 140 is increased, the threshold for the fraud probability associated with calls that are transferred to component 140 can be decreased, so that fraud detection component 140 checks more calls.
  • Component 140 optionally comprises a pre processing component 144, substantially similar to pre processing component 116 of filtering engines 112, which is detailed in association with Fig. 5 below. Pre processing component 144 is preferably activated only if pre processing component 116 of filtering engines 112 was not activated upon the calls.
  • the output of preprocessing component 144 or of preprocessing component 116 of filtering engines 112 is a decompressed enhanced-quality sound of only the customer's side of the conversation.
  • Fraud detection system 140 further comprises a speech parameterization construction component 146, which extracts characteristic features from the enhanced sound, and constructs a sequence of feature vectors for the customer, wherein each feature vector relates to a certain point in time. The extracted characteristics are further detailed in association with Fig. 4 below.
  • System 140 comprises a voice print collection 152, which comprises a voice print for every fraudster's voice available to the user. Each such voice print is constructed from one or more combinations of the abovementioned sequence of feature vectors extracted from the fraudsters' voice.
  • System 140 further comprises a selection component 147 for selecting voice prints from voice print collection 152, for the customer's characteristics to be scored against, and a scoring component 148, which scores the characteristics of the tested voice as extracted from each call transferred by call selection system 106 against the voice prints selected by selection component 147 and outputs one or more scoring results.
  • Selection component 147 preferably searches for an earlier voice recording or a previously constructed voice print of the same alleged customer. If a previously built voice print exists, it is used, otherwise if a voice recording can be retrieved, features are extracted and a voice print is constructed from the customer side of the retrieved interaction, thus generating on-the-fly a voice print. Then the feature vector time sequence extracted from the voice is scored against the voice print of the alleged customer.
  • the score exceeds a predetermined threshold, there is a high probability that the customer is indeed who he/she is claiming to be. If an earlier recording of the same alleged customer does not exist, or the scoring against a previous interaction is skipped, or the determined score is low, the feature vector time sequence is scored against all the voice prints of known fraudsters stored in voice print collection 152. If the score of the voice characteristics of the tested voice against any of the voice prints in voice print collection 152 exceeds a predetermined threshold, the implication is that there is substantial probability that the speaker executed a fraud attempt.
  • the feature vector time sequence at hand is scored against part of the voice prints stored in the reservoir, according to one or more filtering criteria associated with the tested voice or the interaction, such as gender, area code, or the like, in order to save processing power and handle more calls.
  • the process of filtering and the portion of the voice print collection the voice sampled will be scored against, is preferably determined by a set of policy rules.
  • the voice print collection preferably comprises metadata associated with the voice print.
  • the user has full and preferably dynamic control over the data types and the rules.
  • the scoring process is detailed in association with Fig. 4 below. Normalizing component 150 is responsible for normalizing or grading the fraud attempts probabilities generated by comparing a customer's voice to multiple entries in the voice print collection, thus enhancing the score generated by scoring component 148.
  • the normalization process is detailed in association with Fig. 4 below.
  • the output of normalization component 150 is a one or more identities matching the speaker's voice, each possibly associated with a degree of confidence.
  • the overall probability is determined by weighting component 151, taking into account the fraud probability assigned to the interaction by selection system 106 according, for example, to applied policy rules that relate to factors such as credit history of the account, the transaction value, the account value, account type, operation type, and any other dynamically configured formula by the user, and the fraud probability as assigned by combining component 150.
  • Weighting component 151 outputs an overall risk level, or combined probability to being a fraud, associated with the interaction.
  • a notification is issued, such as generating an alert , updating a report or the like, according to the user's requirements.
  • the alert is transferred to the agent who handled, or is still handling the call, to a supervisor, or to any other notification contact, in order to issue a response as long as the fraudster might still be available.
  • the transfer to the source of the conversation is most useful when the call is analyzed in real-time, but also when it is analyzed near-real-time or even off-line.
  • the grading mechanisms of selection system 106, of combining component 150, and the weighting component 151 depend on the user's requirement and resources.
  • relevant voice segments and additional data are collected.
  • Relevant voice segments are preferably high quality recordings of audio containing speech by a known or suspected fraudster.
  • the additional data comprises technical, demographic or other data, such as gender, age or age group, zip code, dialect, accent, area code the call was made from, type of equipment the caller was using, or the like.
  • the data is not limited to types of information anticipated during the installation, and can comprise any other piece of data which will become available at a later time. Users are allowed to define their own set of data as a dynamic profile scheme, and set the rules regarding the voice collection filtering criteria accordingly. Hence the user can avoid sharing with a vendor of this invention the type of sensitive data that is to be collected and the rules associated with filtering the voice collection.
  • the voice segments and optionally the additional data are collected from any available source, either internal to the capturing and logging system as in step 176, or from an external source as in step 180.
  • the voice segments are enhanced if necessary, to obtain as good as possible voice quality.
  • the enhancement comprises separating a summed or conference call. Additionally, the enhancement comprises adjusting the calls to the characteristics of the system. A preferred way of adjusting is compressing and decompressing the call with the same compressor/decompressor used by the apparatus, if the call is obtained from a foreign source and not from internal recording or capturing.
  • the enhancement is substantially similar in nature to the method performed by pre processing component 144 of Fig. 3, which is detailed in association with Fig. 5 below.
  • the enhanced speech is transformed into a sequence of feature vectors.
  • the features can include, but are not limited to, any subset of the following: LPC, Cepstrum, MFCC, Pitch and Energy.
  • voice prints are constructed from the feature vectors, based on statistical modeling. Each voice print is associated with the available data, such as the identity of the speaker, if available, or other data, and stored in the collection in step 186.
  • the collection is normalized, in order to mark a voice of a "wolf, i.e., a fraudster's voice which has common characteristics with many other voices in the collection and will therefore introduce inconsistencies to the scoring process.
  • the normalization process generates score normalization parameters used when scoring the characteristics of a voice sample against one or more voice prints form the collection.
  • the collection can be stored using any collection such as a a file directory, a data base, commercially available data base products or the like.
  • the collection is located remotely to the fraud detection component 140 or there within.
  • the steps associated with the disclosed invention such as the selection of the voice prints for an interaction to be scored against, the risk level assessment and the like, are determined and stored.
  • the data items such as field names can be changed dynamically by the organization, and can even refer to fields that will become known only at a later time.
  • the data can be stored together or separately from the voice prints.
  • the collection is maintained on an ongoing basis, including, for example, re-indexing, backups and the like. Steps 172, 182, 184, 186, 187, 188 and 189 are performed at the construction of the collection, or when new voices are added to the collection.
  • a suspect feature vector time sequence extracted from the customer side of a suspect call is checked against one or more voice prints stored in the database (or a voice print constructed on-the-fly from an earlier recording of the alleged customer). This is done by scoring the feature vector time sequence against a certain voice print, and scoring the feature vector time sequence against a general model, for example a general customer model, a general model of speakers of the same gender, or the like, constructed from voices belonging to a relevant population. The ratio between the two scores, or another normalization involving the two scores, provides the conditional probability that the feature vectors indeed belong to a speaker associated with the voice print.
  • a score-test normalization step is preferably used to provide a normalized score of a feature vector time sequence against multiple voice prints.
  • the scoring of a tested voice against multiple voice prints is analyzed. For example, if a tested voice has a higher than average score against multiple voice prints, the voice is deduced to have common characteristics with many voices, and therefore the scoring results are meaningless and do not provide information
  • the tested feature vector time sequence can be scored against all or part of the voices stored in the collection, for example a male voice can be scored against only the voices known to belong to males according to the additional data collected at step 172, in order to save processing power, provide a fast response, and be able to check more interactions.
  • the voice is scored against voice prints in the collection, and once a scoring result exceeds a predetermined threshold, no further scoring is performed.
  • the results are vertically normalized to detect the true maximal similarity between the voice and a voice in the collection.
  • Collaboration between multiple companies, by sharing fraudster databases is a preferred embodiment of the disclosed invention. When this option is used, a suspected call is tested against all or part of the voice prints in multiple databases, thus increasing the probability of locating a fraud attempt in case the fraudster committed an earlier fraud attempt relating to a different company.
  • multiple companies can establish a common fraudster database, containing voice prints of all fraudsters known to all participating companies.
  • the proposed invention can be based on a query portal, web services or any other method of client server oriented approach wherein a query is made against a central, possibly third-party, database of identified voice collection and returns one or more possible matches for a given voice sample along with a risk score, or a confidence level.
  • each organization only manages its interactions and has limited or no access to the voice print collection.
  • each organization has full view or update privileges upon the voice collection database and can add new voice samples to the collections to the database.
  • Each organization can preferably employ policy rules specific to his organization regarding how to filter the voice collection and how to set the risk level and confidence level. Step 190 is performed whenever a suspect call is transferred to the system.
  • a-posteriori analysis is performed at step 192.
  • the system seeks common denominators to interactions in which a high fraud probability is detected, or to interactions that were reported as fraud attempts.
  • the common denominators such as certain voice characteristics, problematic area codes, calls script scenario or others can then be used to fine-tune the system parameters for subsequent usage.
  • Manually or automatically feeding back data into the system such as real fraud attempts, or other details and re-analyzing the data can be used for enhancing and adapting the parameters, rules, thresholds, weights, filtering engines usage, and other factors used throughout the fraud detection process.
  • the a-posteriori analysis can also provide insights and enable trend analysis, patterns of behavior recognition or the like.
  • the pre processing system receives vocal data 204, consisting in the general case of conversations between customers and service provider representatives, or other persons associated with users of the apparatus and method of the present invention.
  • System 200 further receives additional data or meta data 208, such as, but not limited to the identity, gender, age of the agent or the customer, computer telephony integration (CTI) information such as hold periods, transfers and the like, telephone numbers involved, and technical data such as type and audio characteristics of the used equipment, such as echo level, feedback level, channel distortion, background noise, or the like.
  • CTI computer telephony integration
  • the purpose of the meta data is to enhance and catalyze the operation of the engines that receive the interaction following the preprocessing. For example, by cutting hold periods containing music, out of an interaction, the word spotting engine processing this interaction will provide more accurate results in shorter time, since the interaction has been shortened.
  • system 200 comprises a decompressing component, for decompressing the voice data in case it was compressed when captured or recorded.
  • the decompression is associated with the employed compression algorithm, such as G729, G726, G723.1 or the like.
  • Decompressing is preferably adaptable to support multiple compression algorithms.
  • System 200 further comprises a segmentation component, which filters out non-speech parts of the conversation. The filtering can be based upon CTI events such as hold, music, tones, DTMF or the like, as well as on analysis of the voice, including for example energy level analysis or speech detection, thus removing music on the customer side, background noises and the like.
  • system 200 comprises a speaker separation component 220.
  • Speaker separation component 220 preferably searches for at least two areas in the voice segment whose characteristics are" substantially different, assumes that each area belongs to one speaker, and tries to look for additional areas with similar characteristics for each speaker. Therefore, it is important to remove all non-speech areas from the voice segment, because such areas introduce mistakes both in the initial choice of the areas and during the expansion of the voice segments.
  • speaker separation component 220 further comprises an agent identification component 224.
  • Agent identification component 224 uses the audio characteristics of the environment, as extracted from a multiplicity of agent recordings using the standard equipment of the environment, to build a general agent model, which is characteristic to the environment.
  • the system Given an interaction to separate, the system first looks for an area in the interaction whose extracted characteristics are as similar as possible to the general agent model of the environment. Then, the system looks for an area which is as different as possible from the first area.
  • This enhanced mechanism reduces the errors stemming from wrong initial choice of areas, and from erroneous expansion, thus improving the overall accuracy of the speaker segmentation, and yielding a higher quality onesided voice segment known to belong to the customer.
  • the customer voice is then more reliably scored against known suspect voice prints.
  • Other methods of selecting a starting point include but are not limited to using screen events related to agent operations, customer feedback such as answering questions, spotting words such as compliance phrases or the like.
  • the disclosed invention provides an apparatus and method for identifying a fraud attempt and decreasing the time between the execution of an identity theft and its detection, by locating interactions made by known fraudsters.
  • the first stage is to locate suspected calls, using a rule engine, a human operator, filtering engines or a combination of the above.
  • the calls that are determined by the first stage as suspected are assigned an initial fraud probability and are then transferred to a fraud detection component.
  • the fraud detection component first tries to score a parameterized voice sample taken from the interaction against a voice print of allegedly the same speaker. If the score is below a certain threshold, there is a non-negligible probability that the speaker is a fraudster. In such case, the system scores the voice against voice prints in a collection of voice prints that are known to belong to fraudsters.
  • the voice is preprocessed. Preprocessing includes decompression in case the voice is compressed, removal of non-speech segments, separating the two or more speakers and selecting a specific side, if the two sides were captured together.
  • the apparatus and method are efficient, since they enable users to focus on those calls that pose the greatest threats to their business, for example opening a new account, changing details of an existing account and the like.
  • the disclosed system and apparatus can be operated in conjunction with a capturing or logging system, or as an independent solution receiving calls from a third party. It will be appreciated by persons skilled in the art that many embodiments and alternatives exist to the disclosed invention.
  • Various filtering engines can be used, including speech to text engines, text analysis engines, and others. Different voice enhancements can be performed upon the voice in order to improve its quality and the accuracy of later processing performed upon it. During the comparison stage, more, less, or different features can be extracted form the voice, and the voice print can be constructed in various ways.
  • the scoring of a voice sample against a voice print can be performed in additional ways, such as giving more weight to certain features than to other features, or the like.
  • the methods described hereinabove can be generalized to include other interaction types, such as a face-to-face meeting in a walk-in center, an internet transaction or the like.
  • the rules and thresholds for selecting the calls and for combining the results can also be implemented in a variety of ways.
  • system can be used as a plug-in, i.e. an addition to an existing capturing or recording system.
  • the employment of the proposed apparatus does not dictate changes in the environment, but rather an addition for fraud detection purposes.

Abstract

An apparatus and method for detecting a fraud or fraud attempt in a captured interaction. The method comprising a selection step in which interactions suspected as capturing fraud attempts are selected for further analysis, and assigned a first fraud probability, and a fraud detection step in which the voice is scored against one or more voice prints, of the same alleged customer or of known fraudsters. The first fraud or fraud attempt probability is combined with the result of the scoring of the fraud detection step, to generate a total fraud or fraud attempt probability. If the total fraud or fraud attempt probability exceeds a threshold, a notification is issued. The selection, scoring and combination thereof are performed using user-defined rules and thresholds.

Description

METHOD AND APPARATUS FOR FRAUD DETECTION
BACKGROUND OF THE INVENTION
FIELD OF THE INVENTION The present invention relates to voice analysis in general and to a method and apparatus for fraud detection in particular.
DISCUSSION OF THE RELATED ART Acts of fraud are occurring in ever growing numbers and are becoming a serious problem, harming innocent people, services or goods suppliers, public safety personnel, and credit companies. Fraud and fraud attempts, including theft identities occur in all sectors such as financial institutes, public safety services, insurance or the like. In financial markets, significant harm is created by direct financial loss due to lost goods, lost working hours and expenses on one hand, and distress, shame, credit loss and inconvenience on the other hand. In the public safety domain, frauds can harass, but can also danger life as well as property. Over 9 million people became victims of identity theft in the year 2004 in the United States alone. Common identity theft actions include using false identity when consuming phone or wireless services, using a stolen name, date of birth, or social security number to open a new credit card account, and opening a checking account under a stolen identity. Identity thefts account for 3.23 millions cases per year with an average damage of USD 10,200 per victim. Around 60% of all fraud actions involving existing accounts and credit cards take one week to over six months to be discovered. This long time frame, which is mainly due to the long billing cycle, enables perpetrators to commit multiple fraud actions before suspicion arises. Currently available credit card fraud detection and prevention systems are based on: early alert, when the card holder is aware of the identity theft; usage of mutual repository database of suspects or credit card numbers; and noticing transactions which are incompatible with the card holder's profile, in terms of time, location, purchased goods, etc. However, these methods are of little use when no alert of stolen identity is available, such as when a stolen identity is used for opening a new account or obtaining a new credit card.
There is therefore a need for a novel method and apparatus that in addition to the currently available methods and systems will reduce the number of fraud actions in general and identity thefts in particular. The method and system should also be able to shorten the time it takes to expose additional fraud actions and fraud attempts, and preferably to even uncover them before the interaction is over.
SUMMARY OF THE PRESENT INVENTION
It is an object of the present invention to provide a novel method for an apparatus and method for detecting interactions capturing fraud attempts which overcome the disadvantages of the prior art. In accordance with the present invention, there is thus provided a fraud detection method for generating a first fraud or fraud attempt probability, within one or more captured or recorded interactions, the method comprising the step of scoring one or more voices belonging to one or more tested speakers in the captured or recorded interaction against one or more voice prints within one or more entries in a voice print collection, the scoring step generating a probability that the one or more voices in the captured or recorded interaction belong to one or more second speakers associated with the one or more voice prints, said probability represents the probability that the one or more captured or recorded interaction is fraudulent. Within the method, the probability generated at the scoring step represents the similarity between the voice of the tested speaker and the voice of the at least one second speaker, said second speaker can be a fraudster, a legitimate speaker, or allegedly the tested speaker. The method further comprising a fraud probability generation step, for enhancing the first fraud or fraud attempt probability. The fraud probability generation step employs a rule engine activating one or more rules or using one or more thresholds set by a user of the method. The rules or the thresholds can involve data associated with: the captured or recorded interaction, a speaker thereof, an at least one other interaction of the speaker thereof, the one or more voice prints or the one or more second speakers. The rules or the thresholds can be set dynamically. The method can further comprise: a second scoring step for determining a second scoring result by scoring the one or more voices against a background model; and a normalizing step for normalizing the first scoring result with the second scoring result. Within the method, the scoring step can comprise scoring the one or more tested voices against two or more voice prints, thus obtaining two or more score results. When scoring against two or more voice prints, the method can further comprise a score-test normalization step, for normalizing the at least two score results. Within the method, one or more voice prints or data associated with the voice prints or with the second speaker is stored in a collection. The method can further comprise a retrieving step for retrieving one or more second interactions, and a construction step for constructing the at one or more voice prints from the one or more voices participating in the one or more second interactions. The method can further comprise a step of retrieving one or more voice prints or data associated with the one or more voice prints. When retrieving a voice print, the method can further comprise a filtering step for filtering a collection to retrieve the one or more voice prints such that data associated with the voice prints is related to data associated with the recorded or captured interaction. The filtering can be based on an at least one of: area code of calling number, gender of speaker, age pf speaker, language spoken. The step so of the method can be performed in real-time. The method can further comprising a pre processing step for pre processing the captured or recorded interaction or the second interaction. The pre processing step can comprise one of: compressing; decompressing; segmenting; separating speakers; identifying the agent side; and receiving data related to the captured or recorded interaction or to the second interaction. The additional data can comprise computer telephony integration information. The method can further comprise a selection step for selecting one or more selected interactions to be transferred to the scoring step, and associating the one or more selected interactions with one or more second fraud or fraud attempt probabilities. Within the method, the selection step can comprise one or more of the following: applying one or more rules by a rule engine; auditing interactions; filtering interactions using one or more filtering engines. The filtering engines can comprise one or more of the following: emotion detection, word spotting, speech to text, interaction analysis, and data risk analysis. Within the method, the second fraud or fraud attempt probability represents a risk level associated with the one or more selected interactions. The selection step can employs one or more rules or one or more thresholds set by a user of the method. The rules or thresholds can involve data associated with one or more of the following: the captured or recorded interaction; a speaker thereof; other interactions of the speaker thereof; the one or more voice prints or the at least one second speaker. The one or more rules or one or more thresholds can set dynamically. The method can further comprise a weighting step for generating a combined fraud attempt probability from the first and the second fraud attempt probabilities. The method can further comprise an alert-generation step for generating an alert if said first fraud or fraud attempt probability or said second fraud or fraud attempt probability or a combination thereof exceeds a predetermined threshold. The alert can be sent to an agent or another contact associated with the captured or recorded interaction. The method can further comprise a reporting step for generating or updating a report comprising details of the one or more captured or recorded interactions if the first fraud or fraud attempt probability or the second fraud or fraud attempt probability or a combination thereof exceeds a predetermined threshold. The method can further comprise an auditing step for auditing the report. The auditing step can comprise one or more of the following: listening to calls; reviewing interactions; reviewing history of one or more speakers participating in one or more interactions; listening to previous interactions of speakers participating in one or more interactions appearing in the report, or filling forms. The method can further comprise a step of interaction capturing or logging. The method can further comprise a step of a-posteriori analysis for one or more of the following: adapting one or more parameters or one or more rules or one or more thresholds used by the method, trend analysis, pattern of behavior recognition, selecting one or more filtering engines. Another aspect of the disclosed invention relates to a method for constructing a voice print collection, the method comprising the steps of: receiving one or more suspect voice samples; extracting characteristic features from the one or more voice sample, thus generating one or more sequences of feature vectors; constructing a voice print from the one or more sequences of feature vectors; and storing the voice print in a collection. The method can further comprise a storage step for storing data in the collection, said data associated with the at least one voice print, or with the at least one voice sample, or with an at least one speaker participating in the voice sample. The method can further comprise a normalization step for generating score normalization parameters to be used when scoring one or more voice samples against one or more voice prints from the collection
Yet another aspect of the disclosed invention relates to a fraud detection apparatus for generating a fraud or fraud attempt probability, within one or more captured or recorded interactions, the apparatus comprising: a voice print collection comprising one or more entries, each entry comprising a voice print or data associated with the voice print, said voice print is constructed from a voice known or suspected as belonging to a fraudster; and a scoring component for scoring one or more voices in the captured or recorded interaction against the voice print within the one or more entries, the component generating a probability that the one or more voices in the one or more captured or recorded interaction belong to one or more persons associated with the one or more voice prints, said probability represents the probability that the one or more captured or recorded interaction are fraudulent. BRIEF DESCRIPTION OF THE DRAWINGS The present invention will be understood and appreciated more fully from the following detailed description taken in conjunction with the drawings in which: " Fig. 1 is a schematic block diagram of a typical environment in which a preferred embodiment the present invention is used;
Fig. 2 is a schematic flowchart of the fraud detection method, in accordance with the present invention;
Fig. 3 is a schematic block diagram of the selection system and the fraud detection system, in accordance with the present invention;
Fig. 4 is a schematic flow chart of the steps in building a fraudsters database, in accordance with the present invention; and
Fig. 5 is a schematic block diagram of the pre processing component of the fraud detection system, in accordance with the present invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
The present invention relates to international patent application serial number PCT/IL03/00300 titled APPARATUS, SYSTEM AND METHOD FOR DISPUTE RESOLUTION, REGULATION COMPLIANCE AND QUALITY MANAGEMENT IN FINANCIAL INSTITUTIONS, filed on April 9, 2003, and to United States patent application titled APPARATUS AND METHOD FOR AUDIO ANALYSIS filed on March 17, 2005, and to United States patent application serial number 10/996,811 titled METHOD AND APPARATUS FOR SPEAKER SPOTTING filed on November 23, 2004, and to international patent application serial number PCT/IL02/00741 titled QUALITY MANAGEMENT AND RECORDING SOLUTIONS FOR WALK-IN ENVIRONMENTS, filed on May 9, 2002, the full contents of which are incorporated herein by reference.
The present invention overcomes the disadvantages of the prior art by providing a novel method and apparatus which enhance and add to the capabilities of currently available systems for fraud detection in general and identity theft in particular. Identity theft relates to actions in which the fraudsters is impersonating to be another person, by stating details of the other person, such as ID number, social security number, date of birth or the like, by presenting stolen or false documents, or any other method. The disclosed invention provides an apparatus and method for decreasing the time between the execution of an identity theft and its detection, by identifying interactions executed by known fraudsters.
The present invention uses data collected prior to, during, or subsequent to the occurrence of interactions in which the participants' voices are captured and possibly logged. The used data includes procedural data, such as the essence of the transaction, date, time, location if available, number called or number called from in the case of a telephonic interaction, and the like. In addition, the interaction is captured and possibly logged, so that the voices of the persons participating in the interaction are available. At a first stage, the interaction or the accompanying technical information are optionally processed in order to collect fraudulent calls, i.e. calls which might indicate a fraud attempt. The processing can be manually executed by a person such as a compliance officer considering the information automatically, by considering rules, meta data associated with the interactions or results of engines processing the interactions, or a combination of automatic and manual processing. The suspected calls are assigned a first fraud probability, or a risk level, and are then passed to a second stage performed by a fraud detection component. For a non-limiting example, the system can be tuned to transfer to the fraud detection component only calls that deal with account opening, calls that deal with address changing, calls in which a substantial degree of emotion was detected, calls with specific CTI events, such as multiple hold events, DNIS, VDN, ANI, number dialed, number dialed from or the like. Other factors affecting the portion of the interactions volume that will be analyzed can involve words spotted in the interaction, high emotion of either side of the interaction, external metadata of the calls, business data, screen recorded activity, user (agents) manually tagging the calls during the interactions, files tagging by external data, or the like. The fraud detection component performs voice parameterization of the voice, i.e. extracts the characteristics of the voice. If possible, the fraud detection component then preferably scores the voice characteristics against a voice print which was constructed from the voice characteristics of the same alleged customer, for example one of the account owners in a financial organization, or a person who identified himself through PIN, SSN or any other external or internal data as a legitimate speaker, for verification purposes. In the context of the present invention, the result of the scoring represents the probability that the customer's voice sample belongs to the voice print against which it was scored. The scoring preferably includes a normalization stage, at which the score is compared to the result of scoring the voice characteristics against a general voice print. The general voice model is a background model, constructed from multiple relevant voices, such as customers' voices, voices of the same gender, or the like. If the score is below a certain threshold, there is a significant probability that the person involved in the interaction is not the same person for which an earlier interaction was recorded. In such case, the voice characteristics are scored against a reservoir of voice prints of known fraudsters. The resulting score represents the probability that the customer's voice sample belongs to any of the fraudster's voice print. Alternatively, the customer's voice sample is scored against one or more voice prints, which are selected from the fraudster collection as having similar characteristics with the customer's voice sample at hand. The characteristics can include characteristics stemming from the capturing or the recording systems, such as callers details, CTI information and other tagged business data, as well as characteristics obtained from external sources, such as account number, area, age or the like, and characteristics obtained from the voice itself, such as gender. If a combination of the first fraud probability and the score between the customer's voice sample and the fraudster's voice print contained in the reservoir exceeds a predetermined threshold a high probability that the speaker in the interaction is a known fraudster, and that the interaction is fraudulent is assumed, and a notification, such as an alert or a report is issued for the interaction and steps are taken to control or contain the damage caused by the fraudster. A supervisor or another person is preferably prompted to audit reported interactions. During the scoring stage against the collection, the system preferably further grades the probabilities resulting from scoring a customer's voice sample against the one or more voice prints in the reservoir. The threshold above which a notification is issued for an interaction depends on the resources the user is willing to dedicate to checking the alerts on one hand, and on the implications of unhandled identity thefts on the other hand. Other factors considered in setting the thresholds include considering the status of the customer, the status of the interaction, the importance or the volume of the interaction, the sensitivity of the interaction to the organization, the relative frequency of interactions of the same order of magnitude in the account as well as variations in the same, the sensitivity of the account or the account holder, or the like. For example, a large transaction of a VIP customer in a financial institution will be assigned lower thresholds than other transactions, in order to increase the catching probability of a fraud attempt. In general, the more important the interaction or the customer - the lower the thresholds are set, so that fraud attempts are more likely to be caught. The thresholds are preferably set in a dynamic manner according to interaction metadata and data collected from the interaction itself. The fields, as well as the values, according to which the thresholds are set can be determined dynamically. Testing a voice sample against fraudsters voice prints is extremely fast, therefore the result, i.e. the fraud attempt probability is possibly available shortly after an interaction was executed, or even before it ended, thus limiting further damage by closing the credit line, canceling the account or the like. The thresholds determination rules and values above which calls are transferred to the scoring system are determined by the site's requirements and available resources. Either at the first stage or at the second stage, the voice is preprocessed. Preprocessing optionally includes any one or more of the following: decompression in case the voice is compressed; removal of non-speech segments using an automatic algorithm for speech/non-speech detector or other indications like CTI information such as hold events; separating the speakers; and selecting one of the sides if the two (or more) speakers were captured together, for example the customer side in a telephone call between a customer and a service center. The apparatus and method are efficient, since they enable users to focus on those interactions that pose the greatest threats to their business, for example opening a new account, changing details of an existing account, expanding account credit balance, upgrading account, or the like. The disclosed system and apparatus can be operated in conjunction with a capturing or logging system, or as an independent solution receiving calls from a third party. Referring now to Fig. 1, showing a non-limiting environment in which the present invention is used. A typical user of the invention is a single- or multiple-site organization, such as a bank, a credit card operator, a chain of department stores having their own credit cards, a call-center, an emergency center, an insurance claims operator, or any other organization that performs financial transactions with customers or is likely to have to verify incoming callers' identity. The number of sites used in the present invention is not limited. The system typically employs a voice and data capturing and logging component at each site, responsible for collecting voice and optionally additional data relating to customer interactions from all relevant sources. In Fig. I5 site 1 10 and site N 11 are typical sites of a multi-site organization, for example a bank branch of a national bank. The components shown in each site are not necessarily identical, and each site comprises the input sources and input capturing mechanisms relevant to the activities supported by the site. The possible sources employed at site 1 10 include telephone 12, which is a means by which a substantial number of transactions are performed daily, many of which are also captured. Another source ' of interaction voices are voice-over-IP systems 16 which are taking growing market segments. An additional source is a recording of a vocal chat performed over the internet 20. Yet other sources of vocal transactions are captured interactions performed at walk-in-canters 24, such as recordings performed at a teller's counter at a bank, and recordings from trading systems 26. Additional sources of vocal include microphones, intercom, vocal input by an external system, or the like. The sources of vocal information are not limited to the abovementioned sources, and any additional source of vocal information 28 which can be captured, can be used. Additional data 30 is captured, comprising for example PABX or CTI data including claimed identity of customer, number called or number called from in the case of telephonic interaction, handling agent, hold periods or transfers, business data such as account number or claim number, or the like. Another example for additional data 30 is screen events, containing information such as the interaction subject as entered by the handling agent, image or video data, and products of processing the same, such as recognized face or the like. All the captured voice and data are logged by logging component 32. The capturing and logging mechanisms are specific for each type of source, and are described for example in international patent application serial number PCT/IL02/00741 titled QUALITY MANAGEMENT AND RECORDING SOLUTIONS FOR WALK-IN ENVIRONMENTS, filed on May 9, 2002, and international patent application serial number PCT/IL03/00300 titled APPARATUS, SYSTEM AND METHOD FOR DISPUTE RESOLUTION, REGULATION COMPLIANCE AND QUALITY MANAGEMENT IN FINANCIAL INSTITUTIONS, filed on April 9, 2003. Logging component 32 provides a device for recording the interactions on a temporary or permanent media, such as a memory, a magnetic tape, a magnetic disc, an optical disc, a laser disc, a mass-storage device, or the like. Each site further comprises a selection system 36 for assigning a fraud probability to each interaction, and selecting those interactions that are more suspected as contain fraud action or fraud attempt than others, i.e., their fraud probability exceeds a predetermined threshold. For those selected interactions, there is often a need to identify at least one of the speakers. For example, in a call center that speaker would usually be the customer. If there is a significant probability that the tested speaker is a known fraudster, that will have a large contribution to the ability to contain or control possible damages. Alternatively, if the tested speaker is verified to a high degree to be who he or she is claiming to be, that will remove doubts about the credibility of an interaction. Preferably all, or substantial part of the captured and logged data is transferred to selection system 36. In an alternative embodiment, selection system 36 can serve multiple sites and select suspicious calls from all sites connected to it. Selection system 36 is further detailed in association with Fig. 3 below. Once the suspicious calls had been selected from the sites, the calls with their associated fraud probabilities are transferred to fraud detection system 40. Typically, each organization comprises one ore more fraud detection system 40. The number of employed selection systems 36 and fraud detection systems 40 determines the capacity of the apparatus. Fraud detection system 40 comprises a fraud detection component 44 that checks each interaction transferred by any of selection systems 36 and issues a combined fraud probability for the interaction, taking into account the results generated by fraud detection component 44 and the fraud probability as assigned by selection system 36. Fraud detection component 44 further comprises a threshold determination component 46 for controlling the tradeoff between false alarms and miss detected in the system. The higher the value the fraud threshold is set to, the fewer false alarms occur, i.e., less innocent cases reported as fraud attempt are likely to occur, while the system is more likely to miss real fraudsters. However, if the threshold is set too low, more fraud attempts are likely to be discovered, together with more false alarms. The preference of a customer is affected by the price of the resources required to handle false alarms, vs. the expected cost of a miss detect, i.e., a real fraud action. Therefore, the thresholds are configurable by each customer, and can even be set differently for specific cases. The setting can be dynamic, static, or a combination thereof. For example, the system can be set such that the threshold relevant for opening an account is lower than the threshold relevant for other transactions, thus causing more false alarms but also higher detection rate for opening accounts. The threshold can also be automatically set per transaction or interaction according to the transaction business value, account or called number sensitivity, alerts on the account or the called number, the account holder service level, account holder credit history or any other underlying rule, guideline or policy the user wishes to configure the system according to. Setting the rules in a dynamic manner will allow each organization to set its own threshold policy according to meta data or extracted data associated with each transaction or interaction. The threshold setting can also involve data extracted form the interaction itself, such as spotted words, level of emotion or the like. Once each interaction is processed, a policy manager component will set the threshold for this interaction according to policy rules as set by the user. A notification component 48, such as a report or user interface generator is responsible for outputting the combined results, either by a report presented or sent to a user of all interactions whose combined probability exceeds a predetermined threshold, an alert generated once the combined fraud probability associated with an interaction exceeds a certain threshold, or the like. Preferably, the generated alert is sent to the source of the call or an associated authority, such as the agent who held the call or a supervisor. Since all components of the apparatus preferably work in real-time or near-real time, the alert can be presented to the agent while the call is still being held, and prompt him to take precautions. When a fraud attempt report is generated and presented to a user, the user, for example a supervisor, is provided with the possibility to audit the report, for example by listening to calls appearing in the report, assessing their status, receiving additional information such as the call history of the suspect or other calls relating to the same account, listening to previous interactions of speakers participating in an interaction appearing in the report, filling forms, entering comments, escalating findings or the like. The procedure of defining the audit process is dynamic and can be set by a system administrator, security officer or the like in each organization. All components of the system, including capturing components 12, 16, 20, 24, and 28, selection systems 36, fraud detection system 44 and report/user interface component 48 preferably comprise one or more computing platforms, such as a personal computer, a mainframe computer, or any other type of computing platform that is provisioned with a memory device (not shown), a CPU or microprocessor device, and several I/O ports (not shown). Alternatively, each component can be a DSP chip, an ASIC device storing the commands and data necessary to execute the methods of the present invention, or the like. Each component can further include a storage device (not shown), storing the relevant applications. Each application is a set of logically inter-related computer programs and associated data structures that interact to perform one or more specific tasks.
Referring now to Fig. 2, showing a preferred embodiment of the method of detecting a fraud or a fraud attempt captured in an interaction. For clarity reasons, the described interaction is a telephone conversation between an agent and a customer. However, persons skilled in the art will appreciate that the method can be adapted to any other relevant domain with little or no enhancements. At step 50, all or part of the interactions associated with a customer site, such as a call center, a service center, a financial institute or the like, are captured for analysis and preferably recorded for later retrieval. At step 54, specific interactions are selected as being suspected for capturing or conveying fraud actions or fraud attempts. The selection process optionally involves activating a rule engine, filtering engines or manual selection of calls through the usage of an auditing system. The filtering engines can comprise emotion detection, word spotting, speech to text, interaction analysis, data risk analysis, video analysis, image analysis or the like. The selection process is further detailed in association with Fig. 3 below. The interactions selected at selection step 54 for further analysis, are optionally assigned with a fraud probability determined by the tools involved in the selection process. At step 58, pre processing is performed on the selected interactions, for enhancing their quality for further analysis. Optionally, the preprocessing is performed prior to, or as a part of selection step 54. The preprocessing step preferably includes decompressing the interaction, speech/non- speech segment detection, segmenting the interaction, separating the speakers within the interaction, and identifying the sides of the interaction, i.e. determining which side is the agent and which is the customer. The preprocessing preferably uses additional data related to the call, such as Computer Telephony Integration (CTI) information. The preprocessing step is further detailed in association with Fig. 5 below. At step 62, the voice of the tested speaker is parameterized, by constructing a sequence of feature vectors, wherein each feature vector relates to a certain point in time, from the enhanced voice, wherein, each feature vector comprises a plurality of characteristics of the voice during a specific time frame within the interaction. At step 64, one or more previously constructed voice prints are selected from a collection, or a reservoir, such that the parameterized voice sample of the tested speaker would be scored against these voice prints. The selected voice prints can include a voice print of the alleged speaker of the current voice sample, who is a legitimate speaker, in order to verify that the alleged customer is indeed the true customer as recorded, for example at the time the account was opened. Alternatively the selected voice prints can include voice prints, such as voice prints from a fraudster's reservoir, whose associated data is related to data associated with the voice sample of the current interaction, such as voice prints of the same gender, same area code or the like. The voice can be scored against multiple voice prints, including voice prints of the alleged speaker or speakers, voice print of one or more fraudsters, or any combination in any required order of the above. The system allows the user to set the policy rules regarding the data to be considered from each interaction, and the filtering criteria for voice prints in the voice collection database, in a completely dynamic manner. The selection can be set manually according to policy rules or automatically. The information used for selecting the voice prints can relate to any meta data associated with the interaction or data extracted from the interaction itself, such as spotted words, emotion levels or the like. In yet another preferred embodiment, all voice prints in the reservoir are selected. At step 66, the time sequence feature vector extracted from the voice sample is scored against the one or more selected voice prints selected at step 64. When the voice sample is scored against a voice print constructed from an earlier interaction with the same alleged customer, the system assesses the probability that the caller voice sample matches the previous voice print stored in the system, thus verifying that the caller is the same caller identified in previous calls, and thus the fraud attempt probability is significantly reduced. Alternatively, a new voice print is constructed on the fly from retrieved recorded interactions of the alleged caller. The scoring process is further detailed in association with Fig. 3 below. In a preferred embodiment, the parameterized voice sample can first be scored against a voice print of the same alleged customer, and only if the score result is low, i.e., there is a high probability that it is not the same speaker, the voice is scored against part or all of the voices in the collection, which belong to known fraudsters. The voice prints in the selection against which the voice is scored are preferably filtered according to data associated with the interaction, such as area code of calling number, gender, age, language or the like. At step 70, the results of the scoring between the voice sample and the one or more voice prints from the fraudster collection are enhanced. At step 70, the results are possibly vertically normalized, and graded or otherwise combined at sub-step 72 into a single fraud probability. The functionality of step 70 is detailed in association with Fig. 3 below. At step 74, the fraud probability generated at step 70 is weighted or otherwise combined with the fraud probability assigned to the interaction at selection step 54, if selection step 54 was performed, thus generating a combined fraud probability. If the combined fraud probability exceeds a predetermined threshold, an alert is generated at step 78 for the interaction and preferably sent to the agent handling the interaction, a supervisor, a fraud researcher, a compliance officer or any other contact as defined in the notification rules of the system. The notification policy, including the thresholds and the action to be taken can be set according to the interaction sensitivity, transaction value, risk grade, fraud confidence level, structural organization, or any other notification policy that the user wishes to set. Since all steps are preferably performed in real-time, an alert can be generated even before the interaction is completed, thus giving the agent or supervisor an option to control and reduce possible damages. Alternatively, a report is generated for all interactions assigned with a combined fraud probability exceeding a predetermined threshold, and a supervisor or another person associated with the call center is prompted to audit the report. Auditing the report comprises listening to calls, reviewing interactions, reviewing history of speakers participating in an interaction, listening to previous calls of speakers participating in the interaction, filling forms or the like. The structure of the form and the integrated audit procedure can be set dynamically by a site administrator, a compliance officer or the like. Step 79 is preferably performed after a predetermined time during which the system was working and assigning fraud probabilities to interactions, At a- posteriori analysis step 79 the system analyses and tries to find common denominators to calls that were analyzed and assigned a high probability of conveying or capturing fraud attempts, or calls that were reported by a user as capturing fraud attempts, or the like. The purpose of the analysis is fine tuning of the parameters, rules and thresholds used throughout the method. Another product of the a-posteriori analysis step can be analyzing trends, obtaining insights regarding fraud methods, recurring vocabulary, time of calls, prosody patterns, calling scenario, and the like. Referring now to Fig. 3, which is a schematic block diagram of a preferred embodiment of the proposed apparatus. All components of the proposed apparatus can analyze a call in real-time or near real-time, thus providing a fraud probability for a call while the call is still on-going or shortly after. The fraud probability, especially if exceeding a predetermined threshold can be transferred back to the agent, to a supervisor, or to any other entity which can practice precautions. An interactions and data capturing or logging component 104, as detailed in association with sources and capturing mechanisms 12, 16, 20, 24, 28 and 30 and logging unit 32 of Fig. 1, is responsible for capturing part or all of the interactions taking place in one or more sites, for example a service center, a financial institute such as a bank, an emergency center, an insurance organization or the like, and associated data such as CTI information, images, video and data associated thereof or the like. The captured and logged calls and additional information are then processed by a selection system 106, in order to identify calls that should be further investigated, since these will suggest a possibility of a fraud attempt. Selection system 106 optionally comprises an auditing component 136, which enables a user, such as a compliance officer to listen to calls and mark selected calls. Auditing component 136 provides the user with appropriate user- interface and tools to select calls for listening, to assign a score to a call or to route a call to further processing. Additionally, auditing component 136 enables a user to fill up a customized form for the auditing process, including for example a conversation topic, a status such as new, confirmed, rejected, in process, or the like. The user can also integrate the organization validation and risk analysis methods into the auditing process and risk analysis scoring. Selection system 106 further comprises a rule engine 108, which marks calls that might be suitable for further examination, according to predetermined rules, such as call subject (if available), number called from, or any other data item or combination of data items associated with the call. Yet another component of selection system 36 is filtering engines 112. In the context of the disclosed invention, the term engine refers to one or more computerized components, that interact to receive input of predetermined types, perform a specific analysis on the input, and return one or more results. An engine can comprise software components, firmware, i.e., DSP components, hardware components or any combination thereof. Filtering engines 112 comprise one or more engines, for detecting information embedded within the voice itself, which can offer further information related to the risk level associated with the interaction or with the voice. In addition to these engines, filtering engines 112 preferably comprises a pre processing engine 116, designed to enhance the captured voice and prepare the captured voice for usage by other engines. Pre processing engine 116 is further detailed in association with Fig. 5 below. Filtering engines 112 further comprise a word sporting (WS) engine 120, which searches the voice for occurrences of words or phrases from a predefined list. Engine 120 preferably supplies for each word found, the position within the call at which it was found, and a certainty indicator for the correctness of the finding. For example, WS engine 120 can search for the phrase "open an account", and report that the phrase was located 20 seconds form the beginning of the conversation, with certainty of 80%. Yet another available engine is an emotion detection engine 124. Engine 124 is designed to locate highly emotional segments within the conversation. Each segment is associated with an emotion level, in the range, for example of 0-100. Highly emotional segment can suggest anger on at least one side of the conversation. Alternatively it can suggest tension due to the customer practicing a fraud attempt. Yet another available engine is a talk analysis engine 128, which analyses the dynamics of the conversation, providing such details as the percentage of the time each speaker was speaking, the mutual silence periods, number of cross-talk events of each side, and the like. Additional engines 132 can comprise for example a risk analysis engine deteπriining the risk associated with an interaction based on additional data, such as CTI information, screen information or the like. Additional engines 132 can also comprise a speech to text engine, a text analysis engine, screen analyzed events, video analyzed scene, OCR engines, additional engines or any combination of the above. Additional engines can be based on CTI information such as CTI call flow analysis including number of holds, number of transfers, queue time or the like. Post processing engine 118 can further enhance the results generated by other engines, for example by integrating results supplied by different engines, such as correlating certain spotted words with segments of high emotion and the like. It should be noted that the engines within filtering engines 112 can work in parallel, sequentially, be skipped, be activated depending on the results generated by other engines, be activated by a user, or the like. However, it is preferred that pre-processing engine 116, if present, is activated prior to other engines, and post processing engine 138 is activated consequent to the other engines. The processing components, i.e., auditing component 136, rule engine 108 and filtering engines 112 are preferably not mutual exclusive. Rather, they can cooperate and work in sequence. Depending on the engines' input requirement, an engine may use the results of other engines. For example, the user can use auditing component 136 to listen to calls selected by rule engine 108. In another example rule engine 108 can apply rules which involve results of word spotting engine 120, such as "select all calls in which the phrase "open an account" was spotted, and the customer's first name is John". The goal of the user is to set the rules, the activation order, and the thresholds so that the calls being selected are the most suspicious as being fraud attempts. Risk analysis component 139 combines the output results of auditing component 136, rule engine 108 and filtering engines 112 according to predetermined rules, weights and priorities, and assigns to each call a single risk level estimating the probability that the call is a fraud attempt. The calls whose risk factor exceeds a predetermined threshold, and the data acquired via the usage of the components of selection system 36 are transferred to fraud detection component 140. The processing power of selection system 36 and of the fraud detection system 140 has substantial influence on the selection of the components and the thresholds. If, for example the processing power of fraud detection component 140 is increased, the threshold for the fraud probability associated with calls that are transferred to component 140 can be decreased, so that fraud detection component 140 checks more calls. Component 140 optionally comprises a pre processing component 144, substantially similar to pre processing component 116 of filtering engines 112, which is detailed in association with Fig. 5 below. Pre processing component 144 is preferably activated only if pre processing component 116 of filtering engines 112 was not activated upon the calls. The output of preprocessing component 144 or of preprocessing component 116 of filtering engines 112 is a decompressed enhanced-quality sound of only the customer's side of the conversation. Fraud detection system 140 further comprises a speech parameterization construction component 146, which extracts characteristic features from the enhanced sound, and constructs a sequence of feature vectors for the customer, wherein each feature vector relates to a certain point in time. The extracted characteristics are further detailed in association with Fig. 4 below. System 140 comprises a voice print collection 152, which comprises a voice print for every fraudster's voice available to the user. Each such voice print is constructed from one or more combinations of the abovementioned sequence of feature vectors extracted from the fraudsters' voice. System 140 further comprises a selection component 147 for selecting voice prints from voice print collection 152, for the customer's characteristics to be scored against, and a scoring component 148, which scores the characteristics of the tested voice as extracted from each call transferred by call selection system 106 against the voice prints selected by selection component 147 and outputs one or more scoring results. Selection component 147 preferably searches for an earlier voice recording or a previously constructed voice print of the same alleged customer. If a previously built voice print exists, it is used, otherwise if a voice recording can be retrieved, features are extracted and a voice print is constructed from the customer side of the retrieved interaction, thus generating on-the-fly a voice print. Then the feature vector time sequence extracted from the voice is scored against the voice print of the alleged customer. If the score exceeds a predetermined threshold, there is a high probability that the customer is indeed who he/she is claiming to be. If an earlier recording of the same alleged customer does not exist, or the scoring against a previous interaction is skipped, or the determined score is low, the feature vector time sequence is scored against all the voice prints of known fraudsters stored in voice print collection 152. If the score of the voice characteristics of the tested voice against any of the voice prints in voice print collection 152 exceeds a predetermined threshold, the implication is that there is substantial probability that the speaker executed a fraud attempt. As an alternative, the feature vector time sequence at hand is scored against part of the voice prints stored in the reservoir, according to one or more filtering criteria associated with the tested voice or the interaction, such as gender, area code, or the like, in order to save processing power and handle more calls. The process of filtering and the portion of the voice print collection the voice sampled will be scored against, is preferably determined by a set of policy rules. The voice print collection preferably comprises metadata associated with the voice print. The user has full and preferably dynamic control over the data types and the rules. The scoring process is detailed in association with Fig. 4 below. Normalizing component 150 is responsible for normalizing or grading the fraud attempts probabilities generated by comparing a customer's voice to multiple entries in the voice print collection, thus enhancing the score generated by scoring component 148. The normalization process is detailed in association with Fig. 4 below. The output of normalization component 150 is a one or more identities matching the speaker's voice, each possibly associated with a degree of confidence. The overall probability is determined by weighting component 151, taking into account the fraud probability assigned to the interaction by selection system 106 according, for example, to applied policy rules that relate to factors such as credit history of the account, the transaction value, the account value, account type, operation type, and any other dynamically configured formula by the user, and the fraud probability as assigned by combining component 150. Weighting component 151 outputs an overall risk level, or combined probability to being a fraud, associated with the interaction. For those interactions which are assigned a combined probability higher than a predetermined threshold are, a notification is issued, such as generating an alert , updating a report or the like, according to the user's requirements. Alternatively, the alert is transferred to the agent who handled, or is still handling the call, to a supervisor, or to any other notification contact, in order to issue a response as long as the fraudster might still be available. The transfer to the source of the conversation is most useful when the call is analyzed in real-time, but also when it is analyzed near-real-time or even off-line. The grading mechanisms of selection system 106, of combining component 150, and the weighting component 151 depend on the user's requirement and resources. It should be noted that the more calls the system will examine, and the more voices each suspected call is scored against - the higher the chances to catch a fraudster. In a typical environment, only those interactions which are assigned high fraud probability by selection system 106 are transferred to fraud detection component 140. In an alternative embodiment, selection system 106 is not activated, thus making the combined probability equal to the probability assigned by combining component 150. This embodiment requires high processing power, since it involves processing all interactions by fraud detection system 140, rather than a portion of them, as determined by selection system 106. When processing power is unlimited, substantially all interactions are processed by all filtering engines 112, and substantially all interactions are transferred to fraud detection component 140. Referring now to Fig. 4, showing a preferred embodiment of the process of constructing and using voice print collection 152 of Fig. 3. At step 172, relevant voice segments and additional data are collected. Relevant voice segments are preferably high quality recordings of audio containing speech by a known or suspected fraudster. The additional data comprises technical, demographic or other data, such as gender, age or age group, zip code, dialect, accent, area code the call was made from, type of equipment the caller was using, or the like. Furthermore, the data is not limited to types of information anticipated during the installation, and can comprise any other piece of data which will become available at a later time. Users are allowed to define their own set of data as a dynamic profile scheme, and set the rules regarding the voice collection filtering criteria accordingly. Hence the user can avoid sharing with a vendor of this invention the type of sensitive data that is to be collected and the rules associated with filtering the voice collection. This can be achieved, for example, by using XML, otherwise employing values and tags to label the values, or any other method. The voice segments and optionally the additional data are collected from any available source, either internal to the capturing and logging system as in step 176, or from an external source as in step 180. At step 181 the voice segments are enhanced if necessary, to obtain as good as possible voice quality. The enhancement comprises separating a summed or conference call. Additionally, the enhancement comprises adjusting the calls to the characteristics of the system. A preferred way of adjusting is compressing and decompressing the call with the same compressor/decompressor used by the apparatus, if the call is obtained from a foreign source and not from internal recording or capturing. The enhancement is substantially similar in nature to the method performed by pre processing component 144 of Fig. 3, which is detailed in association with Fig. 5 below. At step 182, the enhanced speech is transformed into a sequence of feature vectors. The features can include, but are not limited to, any subset of the following: LPC, Cepstrum, MFCC, Pitch and Energy. At step 184, voice prints are constructed from the feature vectors, based on statistical modeling. Each voice print is associated with the available data, such as the identity of the speaker, if available, or other data, and stored in the collection in step 186. In step 187, the collection is normalized, in order to mark a voice of a "wolf, i.e., a fraudster's voice which has common characteristics with many other voices in the collection and will therefore introduce inconsistencies to the scoring process. The normalization process generates score normalization parameters used when scoring the characteristics of a voice sample against one or more voice prints form the collection. The collection can be stored using any collection such as a a file directory, a data base, commercially available data base products or the like. The collection is located remotely to the fraud detection component 140 or there within. At step 188, the data items, the rules, parameters and thresholds for
OR affecting the steps associated with the disclosed invention, such as the selection of the voice prints for an interaction to be scored against, the risk level assessment and the like, are determined and stored. The data items, such as field names can be changed dynamically by the organization, and can even refer to fields that will become known only at a later time. The data can be stored together or separately from the voice prints. At step 189 the collection is maintained on an ongoing basis, including, for example, re-indexing, backups and the like. Steps 172, 182, 184, 186, 187, 188 and 189 are performed at the construction of the collection, or when new voices are added to the collection. At step 190 a suspect feature vector time sequence extracted from the customer side of a suspect call is checked against one or more voice prints stored in the database (or a voice print constructed on-the-fly from an earlier recording of the alleged customer). This is done by scoring the feature vector time sequence against a certain voice print, and scoring the feature vector time sequence against a general model, for example a general customer model, a general model of speakers of the same gender, or the like, constructed from voices belonging to a relevant population. The ratio between the two scores, or another normalization involving the two scores, provides the conditional probability that the feature vectors indeed belong to a speaker associated with the voice print. A score-test normalization step is preferably used to provide a normalized score of a feature vector time sequence against multiple voice prints. In the score-test normalization, the scoring of a tested voice against multiple voice prints is analyzed. For example, if a tested voice has a higher than average score against multiple voice prints, the voice is deduced to have common characteristics with many voices, and therefore the scoring results are meaningless and do not provide information The tested feature vector time sequence can be scored against all or part of the voices stored in the collection, for example a male voice can be scored against only the voices known to belong to males according to the additional data collected at step 172, in order to save processing power, provide a fast response, and be able to check more interactions. In yet another alternative, the voice is scored against voice prints in the collection, and once a scoring result exceeds a predetermined threshold, no further scoring is performed. In case the suspect voice is a "wolf, i.e., it has common characteristics with multiple voices in the collection, the results are vertically normalized to detect the true maximal similarity between the voice and a voice in the collection. Collaboration between multiple companies, by sharing fraudster databases is a preferred embodiment of the disclosed invention. When this option is used, a suspected call is tested against all or part of the voice prints in multiple databases, thus increasing the probability of locating a fraud attempt in case the fraudster committed an earlier fraud attempt relating to a different company. Alternatively, multiple companies can establish a common fraudster database, containing voice prints of all fraudsters known to all participating companies. In yet another preferred embodiment, the proposed invention can be based on a query portal, web services or any other method of client server oriented approach wherein a query is made against a central, possibly third-party, database of identified voice collection and returns one or more possible matches for a given voice sample along with a risk score, or a confidence level. In this method, each organization only manages its interactions and has limited or no access to the voice print collection. In yet another preferred alternative, each organization has full view or update privileges upon the voice collection database and can add new voice samples to the collections to the database. Each organization can preferably employ policy rules specific to his organization regarding how to filter the voice collection and how to set the risk level and confidence level. Step 190 is performed whenever a suspect call is transferred to the system. Preferably, after a predetermined period of time during which the system is used, a-posteriori analysis is performed at step 192. During the analysis, the system seeks common denominators to interactions in which a high fraud probability is detected, or to interactions that were reported as fraud attempts. The common denominators, such as certain voice characteristics, problematic area codes, calls script scenario or others can then be used to fine-tune the system parameters for subsequent usage. Manually or automatically feeding back data into the system, such as real fraud attempts, or other details and re-analyzing the data can be used for enhancing and adapting the parameters, rules, thresholds, weights, filtering engines usage, and other factors used throughout the fraud detection process. The a-posteriori analysis can also provide insights and enable trend analysis, patterns of behavior recognition or the like.
Reference is now made to Fig. 5, showing the main components of the preprocessing components, 144 or 116 of Fig. 3. The pre processing system, generally referenced as 200, receives vocal data 204, consisting in the general case of conversations between customers and service provider representatives, or other persons associated with users of the apparatus and method of the present invention. System 200 further receives additional data or meta data 208, such as, but not limited to the identity, gender, age of the agent or the customer, computer telephony integration (CTI) information such as hold periods, transfers and the like, telephone numbers involved, and technical data such as type and audio characteristics of the used equipment, such as echo level, feedback level, channel distortion, background noise, or the like. The purpose of the meta data is to enhance and catalyze the operation of the engines that receive the interaction following the preprocessing. For example, by cutting hold periods containing music, out of an interaction, the word spotting engine processing this interaction will provide more accurate results in shorter time, since the interaction has been shortened.
In a preferred embodiment of the disclosed invention, system 200 comprises a decompressing component, for decompressing the voice data in case it was compressed when captured or recorded. The decompression is associated with the employed compression algorithm, such as G729, G726, G723.1 or the like. Decompressing is preferably adaptable to support multiple compression algorithms. System 200 further comprises a segmentation component, which filters out non-speech parts of the conversation. The filtering can be based upon CTI events such as hold, music, tones, DTMF or the like, as well as on analysis of the voice, including for example energy level analysis or speech detection, thus removing music on the customer side, background noises and the like. In the general case, where the two participants of a conversation are recorded together, there is a need to separate the sound tracks, so features characteristic of one side can be reliably extracted from the sound segment. To that end, system 200 comprises a speaker separation component 220. Speaker separation component 220 preferably searches for at least two areas in the voice segment whose characteristics are" substantially different, assumes that each area belongs to one speaker, and tries to look for additional areas with similar characteristics for each speaker. Therefore, it is important to remove all non-speech areas from the voice segment, because such areas introduce mistakes both in the initial choice of the areas and during the expansion of the voice segments. In addition, the separation is highly sensitive to the initial choice of the starting points within the voice segment, where the system looks for the two sides of the conversation, meaning that a poor choice of starting points can cause separation which is mostly erroneous, containing for example music as one side, the same speaker on both sides or other errors. Additional difficulty arises since even if the separation is performed well, it is difficult to tell the agent side from the customer side. Therefore, speaker separation component 220 further comprises an agent identification component 224. Agent identification component 224 uses the audio characteristics of the environment, as extracted from a multiplicity of agent recordings using the standard equipment of the environment, to build a general agent model, which is characteristic to the environment. Given an interaction to separate, the system first looks for an area in the interaction whose extracted characteristics are as similar as possible to the general agent model of the environment. Then, the system looks for an area which is as different as possible from the first area. This enhanced mechanism reduces the errors stemming from wrong initial choice of areas, and from erroneous expansion, thus improving the overall accuracy of the speaker segmentation, and yielding a higher quality onesided voice segment known to belong to the customer. The customer voice is then more reliably scored against known suspect voice prints. Other methods of selecting a starting point include but are not limited to using screen events related to agent operations, customer feedback such as answering questions, spotting words such as compliance phrases or the like.
The disclosed invention provides an apparatus and method for identifying a fraud attempt and decreasing the time between the execution of an identity theft and its detection, by locating interactions made by known fraudsters. The first stage is to locate suspected calls, using a rule engine, a human operator, filtering engines or a combination of the above. The calls that are determined by the first stage as suspected are assigned an initial fraud probability and are then transferred to a fraud detection component. The fraud detection component first tries to score a parameterized voice sample taken from the interaction against a voice print of allegedly the same speaker. If the score is below a certain threshold, there is a non-negligible probability that the speaker is a fraudster. In such case, the system scores the voice against voice prints in a collection of voice prints that are known to belong to fraudsters. If a combination of the initial fraud probability, and the similarity between the voice in the interaction and a voice in the reservoir exceeds a predetermined threshold it is reported, and steps are taken to control or contain the damage caused by the fraudster. In some business types, such as insurance, the customer side is more likely to be checked for fraudsters, while in others, such as trading floors the agent side is to be checked as well. Either at the selection stage or at the fraud detection stage, the voice is preprocessed. Preprocessing includes decompression in case the voice is compressed, removal of non-speech segments, separating the two or more speakers and selecting a specific side, if the two sides were captured together. The apparatus and method are efficient, since they enable users to focus on those calls that pose the greatest threats to their business, for example opening a new account, changing details of an existing account and the like. The disclosed system and apparatus can be operated in conjunction with a capturing or logging system, or as an independent solution receiving calls from a third party. It will be appreciated by persons skilled in the art that many embodiments and alternatives exist to the disclosed invention. Various filtering engines can be used, including speech to text engines, text analysis engines, and others. Different voice enhancements can be performed upon the voice in order to improve its quality and the accuracy of later processing performed upon it. During the comparison stage, more, less, or different features can be extracted form the voice, and the voice print can be constructed in various ways. In addition, the scoring of a voice sample against a voice print can be performed in additional ways, such as giving more weight to certain features than to other features, or the like. Additionally, the methods described hereinabove can be generalized to include other interaction types, such as a face-to-face meeting in a walk-in center, an internet transaction or the like. The rules and thresholds for selecting the calls and for combining the results can also be implemented in a variety of ways.
It will also be appreciated by persons skilled in the art that the system can be used as a plug-in, i.e. an addition to an existing capturing or recording system. The employment of the proposed apparatus does not dictate changes in the environment, but rather an addition for fraud detection purposes.
It will also be appreciated by persons skilled in the art that the present invention is not limited to what has been particularly shown and described hereinabove. Rather the scope of the present invention is defined only by the claims which follow.

Claims

CLAIMSWhat is claimed is:
1. A fraud detection method for generating a first fraud or fraud attempt probability, within an at least one captured or recorded interaction, the method comprising the step of scoring an at least one voice belonging to an at least one tested speaker in the at least one captured or recorded interaction against an at least one voice print within an at least one entry in a voice print collection, the scoring step generating a probability that the at least one voice in the captured or recorded interaction belongs to an at least one second speaker associated with the at least one voice print, said probability represents the probability that the at least one captured or recorded interaction is fraudulent.
2. The method of claim 1 wherein the probability generated at the scoring step represents the similarity between the voice of the tested speaker and the voice of the at least one second speaker.
3. The method of claim 1 wherein the at least one second speaker is at least one of the group consisting of: a fraudster; a legitimate speaker; or allegedly the tested speaker.
4. The method of claim 1 further comprising a fraud probability generation step, for enhancing the first fraud or fraud attempt probability.
5. The method of claim 4 wherein the fraud probability generation step employs a rule engine activating an at least one rule or using an at least one threshold set by a user of the method.
6. The method of claim 5 wherein said at least one rule or at least one threshold involves data associated with: the captured or recorded interaction; a speaker thereof; an at least one other interaction of the speaker thereof; the at least one voice print; or the at least one second speaker.
7. The method of claim 5 wherein the at least one rule or the at least one threshold is set dynamically.
8. The method of claim 1 further comprising: a second scoring step for determining a second scoring result by scoring the at least one voice against a background model; and a normalizing step for normalizing the first scoring result with the second scoring result.
9. The method of claim 1 wherein the scoring step comprises scoring the tested voice against at least two voice prints, thus obtaining at least two score results.
10. The method of claim 9 further comprising a score-test normalization step, for normalizing the at least two score results.
11. The method of claim 1 wherein the at least one voice print or data associated with the at least one voice print or with the second speaker is stored in a collection.
12. The method of claim 1 further comprising: a retrieving step for retrieving an at least one second interaction; and a construction step for constructing the at least one voice print from an at least one voice participating in the at least one second interaction.
13. The method of claim 1 further comprising a step of retrieving an at least one voice print or retrieving data associated with the at least one voice print.
14. The method of claim 13 further comprising a filtering step for filtering a collection to retrieve the at least one voice print such that data associated with the at least one voice print is related to data associated with the at least one recorded or captured interaction.
15. The method of claim 14 wherein the filtering is based on an at least one of the group consisting of: dynamic data associated with the at least one voice recorded or captured interaction, dynamic data associated with the at least one voice print, predetermined data associated with the at least one voice recorded or captured interaction, predetermined data associated with the at least one voice print, area code of calling number, gender of speaker, age pf speaker, language spoken.
16. The method of claim 1 wherein the steps are performed in real-time.
17. The method according to claim 1 further comprising a pre processing step for pre processing the captured or recorded interaction.
18. The method of claim 17 wherein the pre processing step comprises at least one of the group consisting of: compressing; decompressing; segmenting; separating speakers; identifying the agent side; and receiving data related to the at least one captured or recorded interaction.
19. The method of claim 18 wherein the additional data comprises computer telephony integration information.
20. The method according to claim 12 further comprising a pre processing step for pre processing the at least one second interaction.
21. The method of claim 1 further comprising a selection step for selecting an at least one selected interaction to be transferred to the scoring step, and associating the at least one selected interaction with a second fraud or fraud attempt probability.
22. The method of claim 21 wherein the selection step comprises at least one of the group consisting of: applying an at least one rule by a rule engine; auditing interactions; filtering interactions using an at least one filtering engine.
23. The method of claim 22 wherein the at least one filtering engine comprises one or more of the group consisting of: emotion detection, word spotting, speech to text, interaction analysis, and data risk analysis.
24. The method of claim 21 wherein the second fraud or fraud attempt probability represents a risk level associated with the at least one selected interaction.
25. The method of claim 21 wherein the selection step employs an at least one rule or an at least one threshold set by a user of the method.
26. The method of claim 25 wherein said rule or threshold involves data associated with at least one of the group consisting of: the captured or recorded interaction; a speaker thereof; other interactions of the speaker thereof; the at least one voice print or the at least one second speaker.
27. The method of claim 25 wherein the at least one rule or the at least one threshold is set dynamically.
28. The method of claim 21 further comprising a weighting step for generating a combined fraud attempt probability from the first and the second fraud attempt probabilities.
29. The method according to claim 1 further comprising an alert-generation step for generating an alert if said first fraud or fraud attempt probability exceeds a predetermined threshold.
30. The method of claim 29 wherein the alert is sent to an agent or another contact associated with the captured or recorded interaction.
31. The method according to claim 21 further comprising an alert-generation step for generating an alert if said first fraud or fraud attempt probability or said second fraud or fraud attempt probability or a combination thereof exceeds a predetermined threshold.
32. The method according to claim 1 further comprising a reporting step for generating or updating a report comprising details of the at least one captured or recorded interaction if said first fraud or fraud attempt probability exceeds a predetermined threshold.
33. The method of claim 32 further comprising an auditing step for auditing the report.
34. The method of claim 33 wherein auditing the report comprises at least one of the group consisting of: listening to calls; reviewing interactions; reviewing history of at least one speaker participating in an at least one interaction; listening to previous interactions of speakers participating in an at least one interaction appearing in the report, or filling forms.
35. The method according to claim 21 further comprising a reporting step for generating or updating a report comprising details of the at least one captured or recorded interaction if said first fraud or fraud attempt probability or said second fraud or fraud attempt probability or a combination thereof exceeds a predetermined threshold.
36. The method of claim 1 further comprising a step of interaction capturing or logging.
37. The method of claim 1 further comprising a step of a-posteriori analysis for at least one of the group consisting of: adapting an at least one parameter or an at least one rule or an at least one threshold used by the method, trend analysis, pattern of behavior recognition, selecting an at least one filtering engine.
38. A method for constructing a voice print collection, the method comprising the steps of: receiving at least one suspect voice sample; extracting characteristic features from the at least one voice sample, thus generating an at least one sequence of feature vectors; constructing an at least one voice print from the at least one sequence of feature vectors; and storing the at least one voice print in a collection.
39. The method of claim 38 further comprising a storage step for storing data in the collection, said data associated with the at least one voice print, or with the at least one voice sample, or with an at least one speaker participating in the voice sample.
40. The method of claim 38 further comprising a normalization step for generating score normalization parameters to be used when scoring an at least one voice sample against an at least one voice print from the collection.
41. A fraud detection apparatus for generating a fraud or fraud attempt probability, within an at least one captured or recorded interaction, the apparatus comprising: a voice print collection comprising at least one entry, each entry comprising a voice print or data associated with the voice print, said voice print is constructed from a voice known or suspected as belonging to a fraudster; and a scoring component for scoring an at least one voice in the at least one captured or recorded interaction against the voice print within the at least one entry, the component generating a probability that the at least one voice in the at least one captured or recorded interaction belongs to an at least one person associated with the voice print, said probability represents the probability that the at least one captured or recorded interaction is fraudulent.
PCT/IL2006/000110 2005-05-27 2006-01-26 Method and apparatus for fraud detection WO2006126183A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11/139,812 US7386105B2 (en) 2005-05-27 2005-05-27 Method and apparatus for fraud detection
US11/139,812 2005-05-27

Publications (2)

Publication Number Publication Date
WO2006126183A2 true WO2006126183A2 (en) 2006-11-30
WO2006126183A3 WO2006126183A3 (en) 2007-10-18

Family

ID=37452434

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IL2006/000110 WO2006126183A2 (en) 2005-05-27 2006-01-26 Method and apparatus for fraud detection

Country Status (2)

Country Link
US (2) US7386105B2 (en)
WO (1) WO2006126183A2 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2437477A1 (en) * 2010-09-30 2012-04-04 British Telecommunications public limited company Fraud detection
WO2012042200A1 (en) * 2010-09-30 2012-04-05 British Telecommunications Public Limited Company Speech comparison
US8666841B1 (en) 2007-10-09 2014-03-04 Convergys Information Management Group, Inc. Fraud detection engine and method of using the same
EP2806425A3 (en) * 2013-05-22 2014-12-24 Nuance Communications, Inc. System and method for speaker verification
EP2830014A1 (en) * 2013-07-23 2015-01-28 Amadeus S.A.S. Secure channel payment processing system and method
US9406298B2 (en) 2013-02-07 2016-08-02 Nuance Communications, Inc. Method and apparatus for efficient i-vector extraction
US9865266B2 (en) 2013-02-25 2018-01-09 Nuance Communications, Inc. Method and apparatus for automated speaker parameters adaptation in a deployed speaker verification system
CN111599365A (en) * 2020-04-08 2020-08-28 云知声智能科技股份有限公司 Adaptive threshold generation system and method for voiceprint recognition system
WO2022164410A1 (en) * 2021-01-26 2022-08-04 Turkcell Teknoloji Arastirma Ve Gelistirme Anonim Sirketi A system for detecting fraudulent cases in data stream
US11734754B1 (en) * 2019-01-30 2023-08-22 United Services Automobile Association (Usaa) Intelligent loan recommendation agent
US11935125B1 (en) * 2016-09-07 2024-03-19 United Services Automobile Association (Usaa) Digital imagery, audio, and meta-data

Families Citing this family (180)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6621834B1 (en) * 1999-11-05 2003-09-16 Raindance Communications, Inc. System and method for voice transmission over network protocols
WO2003021569A1 (en) * 2001-08-29 2003-03-13 Tokyo Electron Limited Facility monitor
EP1529315A4 (en) * 2002-07-22 2006-10-18 Measurement Spec Inc Handheld device having ultrasonic transducer for axial transmission of acoustic signals
US7333798B2 (en) 2002-08-08 2008-02-19 Value Added Communications, Inc. Telecommunication call management and monitoring system
US8509736B2 (en) 2002-08-08 2013-08-13 Global Tel*Link Corp. Telecommunication call management and monitoring system with voiceprint verification
US7783021B2 (en) 2005-01-28 2010-08-24 Value-Added Communications, Inc. Digital telecommunications call management and monitoring system
EP1712931A1 (en) * 2005-04-14 2006-10-18 Qinetiq Limited Method and apparatus for detecting a target in a scene
US8924285B2 (en) * 2005-04-21 2014-12-30 Verint Americas Inc. Building whitelists comprising voiceprints not associated with fraud and screening calls using a combination of a whitelist and blacklist
US8930261B2 (en) * 2005-04-21 2015-01-06 Verint Americas Inc. Method and system for generating a fraud risk score using telephony channel based audio and non-audio data
US20070280436A1 (en) * 2006-04-14 2007-12-06 Anthony Rajakumar Method and System to Seed a Voice Database
US8639757B1 (en) 2011-08-12 2014-01-28 Sprint Communications Company L.P. User localization using friend location information
US8510215B2 (en) * 2005-04-21 2013-08-13 Victrio, Inc. Method and system for enrolling a voiceprint in a fraudster database
US20120053939A9 (en) * 2005-04-21 2012-03-01 Victrio Speaker verification-based fraud system for combined automated risk score with agent review and associated user interface
US20120253805A1 (en) * 2005-04-21 2012-10-04 Anthony Rajakumar Systems, methods, and media for determining fraud risk from audio signals
US8793131B2 (en) 2005-04-21 2014-07-29 Verint Americas Inc. Systems, methods, and media for determining fraud patterns and creating fraud behavioral models
US20060248019A1 (en) * 2005-04-21 2006-11-02 Anthony Rajakumar Method and system to detect fraud using voice data
US9571652B1 (en) 2005-04-21 2017-02-14 Verint Americas Inc. Enhanced diarization systems, media and methods of use
US20150381801A1 (en) * 2005-04-21 2015-12-31 Verint Americas Inc. Systems, methods, and media for disambiguating call data to determine fraud
US8073691B2 (en) * 2005-04-21 2011-12-06 Victrio, Inc. Method and system for screening using voice data and metadata
US9113001B2 (en) 2005-04-21 2015-08-18 Verint Americas Inc. Systems, methods, and media for disambiguating call data to determine fraud
US8094803B2 (en) 2005-05-18 2012-01-10 Mattersight Corporation Method and system for analyzing separated voice data of a telephonic communication between a customer and a contact center by applying a psychological behavioral model thereto
US8094790B2 (en) 2005-05-18 2012-01-10 Mattersight Corporation Method and software for training a customer service representative by analysis of a telephonic interaction between a customer and a contact center
US7788266B2 (en) 2005-08-26 2010-08-31 Veveo, Inc. Method and system for processing ambiguous, multi-term search queries
US20070168295A1 (en) * 2006-01-17 2007-07-19 Chao-Hung Wu Verification method for personal credit purchases
US9002313B2 (en) * 2006-02-22 2015-04-07 Federal Signal Corporation Fully integrated light bar
US20070194906A1 (en) * 2006-02-22 2007-08-23 Federal Signal Corporation All hazard residential warning system
US9346397B2 (en) 2006-02-22 2016-05-24 Federal Signal Corporation Self-powered light bar
WO2007120985A2 (en) * 2006-02-22 2007-10-25 Federal Signal Corporation Public safety warning network
US7476013B2 (en) * 2006-03-31 2009-01-13 Federal Signal Corporation Light bar and method for making
US7746794B2 (en) * 2006-02-22 2010-06-29 Federal Signal Corporation Integrated municipal management console
JP2007233590A (en) * 2006-02-28 2007-09-13 Oki Electric Ind Co Ltd Portable electronic device, personal identification system, personal identification method, and manufacturing method of semiconductor device
US8494152B1 (en) * 2006-02-28 2013-07-23 Allstate Insurance Company Systems and methods for automated call-handling and processing
JP4607797B2 (en) * 2006-03-06 2011-01-05 株式会社東芝 Behavior discrimination device, method and program
EP2911071A1 (en) * 2006-04-20 2015-08-26 Veveo, Inc. User interface methods and systems for selecting and presenting content based on user navigation and selection actions associated with the content
US8000242B2 (en) * 2006-07-06 2011-08-16 Alcatel Lucent Reducing packet loss for a packet data service during congestion in a transport network
US8423831B2 (en) * 2006-07-11 2013-04-16 Oracle America, Inc. System and method for performing auditing and correction
US7873156B1 (en) * 2006-09-29 2011-01-18 Verint Americas Inc. Systems and methods for analyzing contact center interactions
US7822605B2 (en) * 2006-10-19 2010-10-26 Nice Systems Ltd. Method and apparatus for large population speaker identification in telephone interactions
US20080152111A1 (en) * 2006-12-26 2008-06-26 Motorola, Inc. Handling of incoming calls
US20110022395A1 (en) * 2007-02-15 2011-01-27 Noise Free Wireless Inc. Machine for Emotion Detection (MED) in a communications device
US20080201158A1 (en) 2007-02-15 2008-08-21 Johnson Mark D System and method for visitation management in a controlled-access environment
US8542802B2 (en) 2007-02-15 2013-09-24 Global Tel*Link Corporation System and method for three-way call detection
US7599475B2 (en) * 2007-03-12 2009-10-06 Nice Systems, Ltd. Method and apparatus for generic analytics
US8718262B2 (en) 2007-03-30 2014-05-06 Mattersight Corporation Method and system for automatically routing a telephonic communication base on analytic attributes associated with prior telephonic communication
US8023639B2 (en) 2007-03-30 2011-09-20 Mattersight Corporation Method and system determining the complexity of a telephonic communication received by a contact center
US11257080B2 (en) 2007-05-04 2022-02-22 Michael Sasha John Fraud deterrence for secure transactions
US8078515B2 (en) 2007-05-04 2011-12-13 Michael Sasha John Systems and methods for facilitating electronic transactions and deterring fraud
US10853855B2 (en) * 2007-05-20 2020-12-01 Michael Sasha John Systems and methods for automatic and transparent client authentication and online transaction verification
WO2008138029A1 (en) * 2007-05-11 2008-11-20 Fmt Worldwide Pty Ltd A detection filter
US20090025084A1 (en) * 2007-05-11 2009-01-22 Fraud Management Technologies Pty Ltd Fraud detection filter
US7899762B2 (en) * 2007-05-17 2011-03-01 International Business Machines Corporation Apparatus, method, and computer program product providing improved identification of suspect entries in transaction data
US8831941B2 (en) * 2007-05-29 2014-09-09 At&T Intellectual Property Ii, L.P. System and method for tracking fraudulent electronic transactions using voiceprints of uncommon words
US10419611B2 (en) 2007-09-28 2019-09-17 Mattersight Corporation System and methods for determining trends in electronic communications
US8050988B2 (en) * 2007-10-24 2011-11-01 Thomson Reuters Global Resources Method and system of generating audit procedures and forms
US8036980B2 (en) * 2007-10-24 2011-10-11 Thomson Reuters Global Resources Method and system of generating audit procedures and forms
US8943539B2 (en) 2007-11-21 2015-01-27 Rovi Guides, Inc. Enabling a friend to remotely modify user data
WO2009089621A1 (en) * 2008-01-14 2009-07-23 Algo Communication Products Ltd. Methods and systems for searching audio records
US8180112B2 (en) * 2008-01-21 2012-05-15 Eastman Kodak Company Enabling persistent recognition of individuals in images
JP5024154B2 (en) * 2008-03-27 2012-09-12 富士通株式会社 Association apparatus, association method, and computer program
US8250637B2 (en) * 2008-04-29 2012-08-21 International Business Machines Corporation Determining the degree of relevance of duplicate alerts in an entity resolution system
US8326788B2 (en) * 2008-04-29 2012-12-04 International Business Machines Corporation Determining the degree of relevance of alerts in an entity resolution system
US20090276222A1 (en) * 2008-05-01 2009-11-05 Cybertech Information International, Inc. Method and system for incorporating one or more inputs and defining and applying a plurality of rules for the different stages of speech and video analytics systems
US8145482B2 (en) * 2008-05-25 2012-03-27 Ezra Daya Enhancing analysis of test key phrases from acoustic sources with key phrase training models
US8886663B2 (en) * 2008-09-20 2014-11-11 Securus Technologies, Inc. Multi-party conversation analyzer and logger
US8533843B2 (en) * 2008-10-13 2013-09-10 Hewlett-Packard Development Company, L. P. Device, method, and program product for determining an overall business service vulnerability score
JP5042194B2 (en) 2008-10-27 2012-10-03 インターナショナル・ビジネス・マシーンズ・コーポレーション Apparatus and method for updating speaker template
DK2364495T3 (en) * 2008-12-10 2017-01-16 Agnitio S L Method of verifying the identity of a speaking and associated computer-readable medium and computer
US8379819B2 (en) * 2008-12-24 2013-02-19 Avaya Inc Indexing recordings of telephony sessions
US9225838B2 (en) 2009-02-12 2015-12-29 Value-Added Communications, Inc. System and method for detecting three-way call circumvention attempts
US20100235908A1 (en) * 2009-03-13 2010-09-16 Silver Tail Systems System and Method for Detection of a Change in Behavior in the Use of a Website Through Vector Analysis
US20100235909A1 (en) * 2009-03-13 2010-09-16 Silver Tail Systems System and Method for Detection of a Change in Behavior in the Use of a Website Through Vector Velocity Analysis
US20100274653A1 (en) * 2009-04-28 2010-10-28 Ayman Hammad Notification social networking
US8422641B2 (en) * 2009-06-15 2013-04-16 Calabrio, Inc. Distributed record server architecture for recording call sessions over a VoIP network
US20100328035A1 (en) * 2009-06-29 2010-12-30 International Business Machines Corporation Security with speaker verification
US8130915B2 (en) * 2009-08-26 2012-03-06 International Business Machines Corporation Verification of user presence during an interactive voice response system session
US9953083B2 (en) * 2010-02-16 2018-04-24 Excalibur Ip, Llc System and method for determining an authority rank for real time searching
US8715178B2 (en) * 2010-02-18 2014-05-06 Bank Of America Corporation Wearable badge with sensor
US8715179B2 (en) * 2010-02-18 2014-05-06 Bank Of America Corporation Call center quality management tool
US9138186B2 (en) * 2010-02-18 2015-09-22 Bank Of America Corporation Systems for inducing change in a performance characteristic
US8396192B2 (en) * 2010-03-03 2013-03-12 Calabrio, Inc. Desktop recording architecture for recording call sessions over a telephony network
WO2011116514A1 (en) * 2010-03-23 2011-09-29 Nokia Corporation Method and apparatus for determining a user age range
US20110235520A1 (en) * 2010-03-26 2011-09-29 Calabrio, Inc. Live monitoring of call sessions over an ip telephony network
US8306814B2 (en) * 2010-05-11 2012-11-06 Nice-Systems Ltd. Method for speaker source classification
EP2405365B1 (en) * 2010-07-09 2013-06-19 Sony Ericsson Mobile Communications AB Method and device for mnemonic contact image association
US8498998B2 (en) * 2010-10-11 2013-07-30 International Business Machines Corporation Grouping identity records to generate candidate lists to use in an entity and relationship resolution process
US8666861B2 (en) * 2010-10-21 2014-03-04 Visa International Service Association Software and methods for risk and fraud mitigation
KR101496632B1 (en) * 2010-11-04 2015-03-03 한국전자통신연구원 System for safe contents service for youths and method therefor
US20140025385A1 (en) * 2010-12-30 2014-01-23 Nokia Corporation Method, Apparatus and Computer Program Product for Emotion Detection
US20120203698A1 (en) * 2011-02-07 2012-08-09 Dustin Duncan Method and System for Fraud Detection and Notification
US20120284026A1 (en) * 2011-05-06 2012-11-08 Nexidia Inc. Speaker verification system
US8954317B1 (en) 2011-07-01 2015-02-10 West Corporation Method and apparatus of processing user text input information
US20130024358A1 (en) * 2011-07-21 2013-01-24 Bank Of America Corporation Filtering transactions to prevent false positive fraud alerts
US20130024375A1 (en) * 2011-07-21 2013-01-24 Bank Of America Corporation Multi-stage filtering for fraud detection
US8620646B2 (en) * 2011-08-08 2013-12-31 The Intellisis Corporation System and method for tracking sound pitch across an audio signal using harmonic envelope
US10424011B2 (en) 2011-11-02 2019-09-24 Gain Credit Holdings, Inc Systems and methods for shared lending risk
US8862461B2 (en) * 2011-11-30 2014-10-14 Match.Com, Lp Fraud detection using text analysis
US8654948B2 (en) * 2011-12-22 2014-02-18 Cox Communications, Inc. Systems and methods of detecting communications fraud
US20130185180A1 (en) * 2012-01-18 2013-07-18 Bank Of America Corporation Determining the investigation priority of potential suspicious events within a financial institution
US9390445B2 (en) 2012-03-05 2016-07-12 Visa International Service Association Authentication using biometric technology through a consumer device
US9716691B2 (en) * 2012-06-07 2017-07-25 Early Warning Services, Llc Enhanced 2CHK authentication security with query transactions
US9026853B2 (en) * 2012-07-31 2015-05-05 Hewlett-Packard Development Company, L.P. Enhancing test scripts
US9368116B2 (en) 2012-09-07 2016-06-14 Verint Systems Ltd. Speaker separation in diarization
US20140074466A1 (en) 2012-09-10 2014-03-13 Google Inc. Answering questions using environmental context
US8484017B1 (en) * 2012-09-10 2013-07-09 Google Inc. Identifying media content
US9286899B1 (en) * 2012-09-21 2016-03-15 Amazon Technologies, Inc. User authentication for devices using voice input or audio signatures
US9837078B2 (en) 2012-11-09 2017-12-05 Mattersight Corporation Methods and apparatus for identifying fraudulent callers
US10134400B2 (en) 2012-11-21 2018-11-20 Verint Systems Ltd. Diarization using acoustic labeling
SG11201504186UA (en) 2012-12-19 2015-07-30 Visa Int Service Ass System and method for voice authentication
US8694315B1 (en) * 2013-02-05 2014-04-08 Visa International Service Association System and method for authentication using speaker verification techniques and fraud model
US9143506B2 (en) 2013-02-13 2015-09-22 Daniel Duncan Systems and methods for identifying biometric information as trusted and authenticating persons using trusted biometric information
US8572398B1 (en) 2013-02-13 2013-10-29 Daniel Duncan Systems and methods for identifying biometric information as trusted and authenticating persons using trusted biometric information
US8914645B2 (en) 2013-02-13 2014-12-16 Daniel Duncan Systems and methods for identifying biometric information as trusted and authenticating persons using trusted biometric information
US9620123B2 (en) * 2013-05-02 2017-04-11 Nice Ltd. Seamless authentication and enrollment
US9460722B2 (en) 2013-07-17 2016-10-04 Verint Systems Ltd. Blind diarization of recorded calls with arbitrary number of speakers
US9984706B2 (en) 2013-08-01 2018-05-29 Verint Systems Ltd. Voice activity detection using a soft decision mechanism
US10014006B1 (en) 2013-09-10 2018-07-03 Ampersand, Inc. Method of determining whether a phone call is answered by a human or by an automated device
US9053711B1 (en) * 2013-09-10 2015-06-09 Ampersand, Inc. Method of matching a digitized stream of audio signals to a known audio recording
US9129152B2 (en) * 2013-11-14 2015-09-08 Adobe Systems Incorporated Exemplar-based feature weighting
US20150142446A1 (en) * 2013-11-21 2015-05-21 Global Analytics, Inc. Credit Risk Decision Management System And Method Using Voice Analytics
US9472194B2 (en) * 2014-03-21 2016-10-18 Wells Fargo Bank, N.A. Enhanced fraud detection
US10237399B1 (en) 2014-04-01 2019-03-19 Securus Technologies, Inc. Identical conversation detection method and apparatus
US9621713B1 (en) 2014-04-01 2017-04-11 Securus Technologies, Inc. Identical conversation detection method and apparatus
US20150363735A1 (en) * 2014-06-13 2015-12-17 Vivint, Inc. Tracking customer interactions for a business
WO2016015687A1 (en) * 2014-07-31 2016-02-04 腾讯科技(深圳)有限公司 Voiceprint verification method and device
US9509705B2 (en) * 2014-08-07 2016-11-29 Wells Fargo Bank, N.A. Automated secondary linking for fraud detection systems
US10120542B2 (en) 2014-10-08 2018-11-06 International Business Machines Corporation Reproducing state of source environment when image was screen captured on a different computing device using resource location, resource navigation and positional metadata embedded in image
US9232052B1 (en) * 2014-11-21 2016-01-05 Marchex, Inc. Analyzing voice characteristics to detect fraudulent call activity and take corrective action without using recording, transcription or caller ID
US9922048B1 (en) 2014-12-01 2018-03-20 Securus Technologies, Inc. Automated background check via facial recognition
US9875743B2 (en) 2015-01-26 2018-01-23 Verint Systems Ltd. Acoustic signature building for a speaker from multiple sessions
JP6892435B2 (en) * 2015-08-11 2021-06-23 カタリナ マーケティング コーポレーション Systems and methods for self-configuring networked media hub devices
EP3506613A1 (en) * 2015-10-14 2019-07-03 Pindrop Security, Inc. Call detail record analysis to identify fraudulent activity and fraud detection in interactive voice response systems
US10949850B1 (en) * 2015-12-30 2021-03-16 Wells Fargo Bank, N.A. Systems and methods for using location services to detect fraud
US10572961B2 (en) 2016-03-15 2020-02-25 Global Tel*Link Corporation Detection and prevention of inmate to inmate message relay
US9609121B1 (en) 2016-04-07 2017-03-28 Global Tel*Link Corporation System and method for third party monitoring of voice and video calls
US10607220B2 (en) 2016-08-25 2020-03-31 Mastercard International Incorporated Systems and methods for consolidated message processing
USD781311S1 (en) 2016-08-30 2017-03-14 Tinder, Inc. Display screen or portion thereof with a graphical user interface
USD854025S1 (en) 2016-08-30 2019-07-16 Match Group, Llc Display screen or portion thereof with a graphical user interface of an electronic device
USD781882S1 (en) 2016-08-30 2017-03-21 Tinder, Inc. Display screen or portion thereof with a graphical user interface of an electronic device
USD780775S1 (en) 2016-08-30 2017-03-07 Tinder, Inc. Display screen or portion thereof with a graphical user interface of an electronic device
USD852809S1 (en) 2016-08-30 2019-07-02 Match Group, Llc Display screen or portion thereof with a graphical user interface of an electronic device
US11263275B1 (en) * 2017-04-03 2022-03-01 Massachusetts Mutual Life Insurance Company Systems, devices, and methods for parallelized data structure processing
US10027797B1 (en) 2017-05-10 2018-07-17 Global Tel*Link Corporation Alarm control for inmate call monitoring
CN110637321A (en) * 2017-05-16 2019-12-31 维萨国际服务协会 Dynamic claims submission system
US10225396B2 (en) 2017-05-18 2019-03-05 Global Tel*Link Corporation Third party monitoring of a activity within a monitoring platform
US10860786B2 (en) 2017-06-01 2020-12-08 Global Tel*Link Corporation System and method for analyzing and investigating communication data from a controlled environment
US9930088B1 (en) 2017-06-22 2018-03-27 Global Tel*Link Corporation Utilizing VoIP codec negotiation during a controlled environment call
US10623581B2 (en) * 2017-07-25 2020-04-14 Vail Systems, Inc. Adaptive, multi-modal fraud detection system
US10896673B1 (en) * 2017-09-21 2021-01-19 Wells Fargo Bank, N.A. Authentication of impaired voices
US10769647B1 (en) * 2017-12-21 2020-09-08 Wells Fargo Bank, N.A. Divergent trend detection and mitigation computing system
CN109995834A (en) * 2017-12-30 2019-07-09 中国移动通信集团贵州有限公司 Massive dataflow processing method, calculates equipment and storage medium at device
US11275854B2 (en) 2018-02-01 2022-03-15 Nuance Communications, Inc. Conversation print system and method
US10003688B1 (en) 2018-02-08 2018-06-19 Capital One Services, Llc Systems and methods for cluster-based voice verification
US11038886B1 (en) 2018-02-08 2021-06-15 Wells Fargo Bank, N.A. Compliance management system
US10694040B1 (en) 2018-02-26 2020-06-23 Wells Fargo Bank, N.A. Centralized event log generation and analysis for contact centers
US10388286B1 (en) * 2018-03-20 2019-08-20 Capital One Services, Llc Systems and methods of sound-based fraud protection
US11843719B1 (en) * 2018-03-30 2023-12-12 8X8, Inc. Analysis of customer interaction metrics from digital voice data in a data-communication server system
CN108735222B (en) * 2018-05-03 2019-05-17 广州国音科技有限公司 A kind of vocal print identification method and system based on Application on Voiceprint Recognition
US11538128B2 (en) 2018-05-14 2022-12-27 Verint Americas Inc. User interface for fraud alert management
US10313511B1 (en) * 2018-06-05 2019-06-04 Wells Fargo Bank, N.A. Customer self-help control system for contact centers
US10721350B1 (en) 2018-08-21 2020-07-21 Wells Fargo Bank, N.A. Fraud detection in contact centers using deep learning model
WO2020051695A1 (en) * 2018-09-11 2020-03-19 Mastercard Technologies Canada ULC Optimized execution of fraud detection rules
US11023687B2 (en) * 2018-10-08 2021-06-01 Verint Americas Inc. System and method for sentiment analysis of chat ghost typing
US10956932B2 (en) * 2018-10-18 2021-03-23 Adobe Inc. System for providing proof and attestation services for claim verification
US10484532B1 (en) * 2018-10-23 2019-11-19 Capital One Services, Llc System and method detecting fraud using machine-learning and recorded voice clips
US10887452B2 (en) 2018-10-25 2021-01-05 Verint Americas Inc. System architecture for fraud detection
EP3987743A1 (en) 2019-06-20 2022-04-27 Verint Americas Inc. Systems and methods for authentication and fraud detection
US11144536B2 (en) * 2019-06-26 2021-10-12 Nice Ltd. Systems and methods for real-time analytics detection for a transaction utilizing synchronously updated statistical aggregation data
US11470194B2 (en) 2019-08-19 2022-10-11 Pindrop Security, Inc. Caller verification via carrier metadata
US11741560B2 (en) * 2019-09-09 2023-08-29 Deckard Technologies, Inc. Detecting and validating improper homeowner exemptions through data mining, natural language processing, and machine learning
US10958779B1 (en) * 2019-09-23 2021-03-23 Capital One Services, Llc Machine learning dataset generation using a natural language processing technique
US11076043B2 (en) * 2019-10-04 2021-07-27 Red Box Recorders Limited Systems and methods of voiceprint generation and use in enforcing compliance policies
US11868453B2 (en) 2019-11-07 2024-01-09 Verint Americas Inc. Systems and methods for customer authentication based on audio-of-interest
US11295310B2 (en) 2020-02-04 2022-04-05 Visa International Service Association Method, system, and computer program product for fraud detection
US11146678B2 (en) * 2020-02-24 2021-10-12 International Business Machines Corporation Determining the context of calls
US11170800B2 (en) * 2020-02-27 2021-11-09 Microsoft Technology Licensing, Llc Adjusting user experience for multiuser sessions based on vocal-characteristic models
US11363138B2 (en) 2020-10-23 2022-06-14 Nuance Communications, Inc. Fraud detection system and method
CN112735431B (en) * 2020-12-29 2023-12-22 三星电子(中国)研发中心 Model training method and device and artificial intelligent dialogue recognition method and device
US11582336B1 (en) * 2021-08-04 2023-02-14 Nice Ltd. System and method for gender based authentication of a caller
CN114299921B (en) * 2021-12-07 2022-11-18 浙江大学 Voiceprint security scoring method and system for voice command
US20230196368A1 (en) * 2021-12-17 2023-06-22 SOURCE Ltd. System and method for providing context-based fraud detection

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6480825B1 (en) * 1997-01-31 2002-11-12 T-Netix, Inc. System and method for detecting a recorded voice
US6526871B1 (en) * 2001-08-24 2003-03-04 Federal-Mogul World Wide, Inc. Monobloc piston for diesel engines
US20030101348A1 (en) * 2001-07-12 2003-05-29 Russo Anthony P. Method and system for determining confidence in a digital transaction
US20040225499A1 (en) * 2001-07-03 2004-11-11 Wang Sandy Chai-Jen Multi-platform capable inference engine and universal grammar language adapter for intelligent voice application execution

Family Cites Families (81)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4145715A (en) 1976-12-22 1979-03-20 Electronic Management Support, Inc. Surveillance system
US4527151A (en) 1982-05-03 1985-07-02 Sri International Method and apparatus for intrusion detection
US4821118A (en) 1986-10-09 1989-04-11 Advanced Identification Systems, Inc. Video image system for personal identification
US5353618A (en) * 1989-08-24 1994-10-11 Armco Steel Company, L.P. Apparatus and method for forming a tubular frame member
GB9000105D0 (en) 1990-01-03 1990-03-07 Racal Recorders Ltd Recording system
US5051827A (en) 1990-01-29 1991-09-24 The Grass Valley Group, Inc. Television signal encoder/decoder configuration control
US5091780A (en) 1990-05-09 1992-02-25 Carnegie-Mellon University A trainable security system emthod for the same
EP0484076B1 (en) 1990-10-29 1996-12-18 Kabushiki Kaisha Toshiba Video camera having focusing and image-processing function
DE69124777T2 (en) 1990-11-30 1997-06-26 Canon Kk Device for the detection of the motion vector
GB2259212B (en) 1991-08-27 1995-03-29 Sony Broadcast & Communication Standards conversion of digital video signals
GB2268354B (en) 1992-06-25 1995-10-25 Sony Broadcast & Communication Time base conversion
JP2604863Y2 (en) * 1993-11-10 2000-06-05 旭光学工業株式会社 Automatic focusing camera with red-eye prevention function
US5519446A (en) 1993-11-13 1996-05-21 Goldstar Co., Ltd. Apparatus and method for converting an HDTV signal to a non-HDTV signal
US5491511A (en) 1994-02-04 1996-02-13 Odle; James A. Multimedia capture and audit system for a video surveillance network
SG43099A1 (en) * 1994-03-02 1997-10-17 British Telecomm Charging for usage of a telecommunications network
IL113434A0 (en) 1994-04-25 1995-07-31 Katz Barry Surveillance system and method for asynchronously recording digital data with respect to video data
US6028626A (en) 1995-01-03 2000-02-22 Arc Incorporated Abnormality detection and surveillance system
MY132441A (en) 1995-01-17 2007-10-31 Sarnoff Corp Method and apparatus for detecting object movement within an image sequence
US5751346A (en) 1995-02-10 1998-05-12 Dozier Financial Corporation Image retention and information security system
US5796439A (en) 1995-12-21 1998-08-18 Siemens Medical Systems, Inc. Video format conversion process and apparatus
US5742349A (en) 1996-05-07 1998-04-21 Chrontel, Inc. Memory efficient video graphics subsystem with vertical filtering and scan rate conversion
US6081606A (en) 1996-06-17 2000-06-27 Sarnoff Corporation Apparatus and a method for detecting motion within an image sequence
US7304662B1 (en) 1996-07-10 2007-12-04 Visilinx Inc. Video surveillance system and method
US5895453A (en) 1996-08-27 1999-04-20 Sts Systems, Ltd. Method and system for the detection, management and prevention of losses in retail and other environments
US5790096A (en) 1996-09-03 1998-08-04 Allus Technology Corporation Automated flat panel display control system for accomodating broad range of video types and formats
GB9620082D0 (en) 1996-09-26 1996-11-13 Eyretel Ltd Signal monitoring apparatus
US6031573A (en) 1996-10-31 2000-02-29 Sensormatic Electronics Corporation Intelligent video information management system performing multiple functions in parallel
WO1998022936A1 (en) * 1996-11-22 1998-05-28 T-Netix, Inc. Subword-based speaker verification using multiple classifier fusion, with channel, fusion, model, and threshold adaptation
US6037991A (en) 1996-11-26 2000-03-14 Motorola, Inc. Method and apparatus for communicating video information in a communication system
EP0858066A1 (en) 1997-02-03 1998-08-12 Koninklijke Philips Electronics N.V. Method and device for converting the digital image rate
WO1998044441A1 (en) * 1997-03-31 1998-10-08 Bellsouth Intellectual Property Corporation A system and method for associating services information with selected elements of an organization
US5897616A (en) * 1997-06-11 1999-04-27 International Business Machines Corporation Apparatus and methods for speaker verification/identification/classification employing non-acoustic and/or acoustic models and databases
US6295367B1 (en) 1997-06-19 2001-09-25 Emtera Corporation System and method for tracking movement of objects in a scene using correspondence graphs
US6092197A (en) 1997-12-31 2000-07-18 The Customer Logic Company, Llc System and method for the secure discovery, exploitation and publication of information
US6014647A (en) 1997-07-08 2000-01-11 Nizzari; Marcia M. Customer interaction tracking
US6097429A (en) 1997-08-01 2000-08-01 Esco Electronics Corporation Site control unit for video security system
US6108711A (en) 1998-09-11 2000-08-22 Genesys Telecommunications Laboratories, Inc. Operating system having external media layer, workflow layer, internal media layer, and knowledge base for routing media events between transactions
US6111610A (en) 1997-12-11 2000-08-29 Faroudja Laboratories, Inc. Displaying film-originated video on high frame rate monitors without motions discontinuities
US6704409B1 (en) 1997-12-31 2004-03-09 Aspect Communications Corporation Method and apparatus for processing real-time transactions and non-real-time transactions
US6327343B1 (en) 1998-01-16 2001-12-04 International Business Machines Corporation System and methods for automatic call and data transfer processing
US6167395A (en) 1998-09-11 2000-12-26 Genesys Telecommunications Laboratories, Inc Method and apparatus for creating specialized multimedia threads in a multimedia communication center
US6212178B1 (en) 1998-09-11 2001-04-03 Genesys Telecommunication Laboratories, Inc. Method and apparatus for selectively presenting media-options to clients of a multimedia call center
US6170011B1 (en) 1998-09-11 2001-01-02 Genesys Telecommunications Laboratories, Inc. Method and apparatus for determining and initiating interaction directionality within a multimedia communication center
US6138139A (en) 1998-10-29 2000-10-24 Genesys Telecommunications Laboraties, Inc. Method and apparatus for supporting diverse interaction paths within a multimedia communication center
US6070142A (en) 1998-04-17 2000-05-30 Andersen Consulting Llp Virtual customer sales and service center and method
US6134530A (en) 1998-04-17 2000-10-17 Andersen Consulting Llp Rule based routing system and method for a virtual sales and service center
US20010043697A1 (en) 1998-05-11 2001-11-22 Patrick M. Cox Monitoring of and remote access to call center activity
US6604108B1 (en) 1998-06-05 2003-08-05 Metasolutions, Inc. Information mart system and information mart browser
US6628835B1 (en) 1998-08-31 2003-09-30 Texas Instruments Incorporated Method and system for defining and recognizing complex events in a video sequence
US6570608B1 (en) 1998-09-30 2003-05-27 Texas Instruments Incorporated System and method for detecting interactions of people and vehicles
US6549613B1 (en) 1998-11-05 2003-04-15 Ulysses Holding Llc Method and apparatus for intercept of wireline communications
US6330025B1 (en) 1999-05-10 2001-12-11 Nice Systems Ltd. Digital video logging system
WO2000073996A1 (en) 1999-05-28 2000-12-07 Glebe Systems Pty Ltd Method and apparatus for tracking a moving object
US7103806B1 (en) 1999-06-04 2006-09-05 Microsoft Corporation System for performing context-sensitive decisions about ideal communication modalities considering information about channel reliability
GB2352948B (en) 1999-07-13 2004-03-31 Racal Recorders Ltd Voice activity monitoring apparatus and methods
US6427137B2 (en) 1999-08-31 2002-07-30 Accenture Llp System, method and article of manufacture for a voice analysis system that detects nervousness for preventing fraud
US6275806B1 (en) 1999-08-31 2001-08-14 Andersen Consulting, Llp System method and article of manufacture for detecting emotion in voice signals by utilizing statistics for voice signal parameters
US20010052081A1 (en) 2000-04-07 2001-12-13 Mckibben Bernard R. Communication network with a service agent element and method for providing surveillance services
JP2001357484A (en) 2000-06-14 2001-12-26 Kddi Corp Road abnormality detector
US6981000B2 (en) 2000-06-30 2005-12-27 Lg Electronics Inc. Customer relationship management system and operation method thereof
US20020059283A1 (en) 2000-10-20 2002-05-16 Enteractllc Method and system for managing customer relations
US20020054211A1 (en) 2000-11-06 2002-05-09 Edelson Steven D. Surveillance video camera enhancement system
US6441734B1 (en) 2000-12-12 2002-08-27 Koninklijke Philips Electronics N.V. Intruder detection through trajectory analysis in monitoring and surveillance systems
US20020087385A1 (en) 2000-12-28 2002-07-04 Vincent Perry G. System and method for suggesting interaction strategies to a customer service representative
US20050234727A1 (en) * 2001-07-03 2005-10-20 Leo Chiu Method and apparatus for adapting a voice extensible markup language-enabled voice system for natural speech recognition and system response
US7953219B2 (en) 2001-07-19 2011-05-31 Nice Systems, Ltd. Method apparatus and system for capturing and analyzing interaction based content
GB0118921D0 (en) 2001-08-02 2001-09-26 Eyretel Telecommunications interaction analysis
US6912272B2 (en) 2001-09-21 2005-06-28 Talkflow Systems, Llc Method and apparatus for managing communications and for creating communication routing rules
US20030128099A1 (en) 2001-09-26 2003-07-10 Cockerham John M. System and method for securing a defined perimeter using multi-layered biometric electronic processing
US6559769B2 (en) 2001-10-01 2003-05-06 Eric Anthony Early warning real-time security system
AU2003207979A1 (en) 2002-02-06 2003-09-02 Nice Systems Ltd. Method and apparatus for video frame sequence-based object tracking
EP1472869A4 (en) 2002-02-06 2008-07-30 Nice Systems Ltd System and method for video content analysis-based detection, surveillance and alarm management
US7386113B2 (en) 2002-02-25 2008-06-10 Genesys Telecommunications Laboratories, Inc. System and method for integrated resource scheduling and agent work management
AU2003263957A1 (en) 2002-08-16 2004-03-03 Nuasis Corporation Contact center architecture
US7076427B2 (en) 2002-10-18 2006-07-11 Ser Solutions, Inc. Methods and apparatus for audio data monitoring and evaluation using speech recognition
US20040098295A1 (en) 2002-11-15 2004-05-20 Iex Corporation Method and system for scheduling workload
EP1450321A1 (en) * 2003-02-21 2004-08-25 Swisscom Mobile AG Method and system for detecting possible fraud in paying transactions
WO2004091250A1 (en) 2003-04-09 2004-10-21 Telefonaktiebolaget Lm Ericsson (Publ) Lawful interception of multimedia calls
US7447909B2 (en) 2003-06-05 2008-11-04 Nortel Networks Limited Method and system for lawful interception of packet switched network services
DE10358333A1 (en) 2003-12-12 2005-07-14 Siemens Ag Telecommunication monitoring procedure uses speech and voice characteristic recognition to select communications from target user groups
WO2006045102A2 (en) 2004-10-20 2006-04-27 Seven Networks, Inc. Method and apparatus for intercepting events in a communication system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6480825B1 (en) * 1997-01-31 2002-11-12 T-Netix, Inc. System and method for detecting a recorded voice
US20040225499A1 (en) * 2001-07-03 2004-11-11 Wang Sandy Chai-Jen Multi-platform capable inference engine and universal grammar language adapter for intelligent voice application execution
US20030101348A1 (en) * 2001-07-12 2003-05-29 Russo Anthony P. Method and system for determining confidence in a digital transaction
US6526871B1 (en) * 2001-08-24 2003-03-04 Federal-Mogul World Wide, Inc. Monobloc piston for diesel engines

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8666841B1 (en) 2007-10-09 2014-03-04 Convergys Information Management Group, Inc. Fraud detection engine and method of using the same
WO2012042200A1 (en) * 2010-09-30 2012-04-05 British Telecommunications Public Limited Company Speech comparison
EP2437477A1 (en) * 2010-09-30 2012-04-04 British Telecommunications public limited company Fraud detection
US9406298B2 (en) 2013-02-07 2016-08-02 Nuance Communications, Inc. Method and apparatus for efficient i-vector extraction
US9865266B2 (en) 2013-02-25 2018-01-09 Nuance Communications, Inc. Method and apparatus for automated speaker parameters adaptation in a deployed speaker verification system
EP2806425A3 (en) * 2013-05-22 2014-12-24 Nuance Communications, Inc. System and method for speaker verification
US9258425B2 (en) 2013-05-22 2016-02-09 Nuance Communications, Inc. Method and system for speaker verification
EP2830014A1 (en) * 2013-07-23 2015-01-28 Amadeus S.A.S. Secure channel payment processing system and method
US11935125B1 (en) * 2016-09-07 2024-03-19 United Services Automobile Association (Usaa) Digital imagery, audio, and meta-data
US11734754B1 (en) * 2019-01-30 2023-08-22 United Services Automobile Association (Usaa) Intelligent loan recommendation agent
CN111599365A (en) * 2020-04-08 2020-08-28 云知声智能科技股份有限公司 Adaptive threshold generation system and method for voiceprint recognition system
CN111599365B (en) * 2020-04-08 2023-05-05 云知声智能科技股份有限公司 Adaptive threshold generation system and method for voiceprint recognition system
WO2022164410A1 (en) * 2021-01-26 2022-08-04 Turkcell Teknoloji Arastirma Ve Gelistirme Anonim Sirketi A system for detecting fraudulent cases in data stream

Also Published As

Publication number Publication date
US7801288B2 (en) 2010-09-21
US7386105B2 (en) 2008-06-10
US20080154609A1 (en) 2008-06-26
US20060285665A1 (en) 2006-12-21
WO2006126183A3 (en) 2007-10-18

Similar Documents

Publication Publication Date Title
US7386105B2 (en) Method and apparatus for fraud detection
US10685657B2 (en) Biometrics platform
US8219404B2 (en) Method and apparatus for recognizing a speaker in lawful interception systems
US8078463B2 (en) Method and apparatus for speaker spotting
US7822605B2 (en) Method and apparatus for large population speaker identification in telephone interactions
US8145562B2 (en) Apparatus and method for fraud prevention
US10142461B2 (en) Multi-party conversation analyzer and logger
US9240013B2 (en) Evaluation of voice communications
US8005675B2 (en) Apparatus and method for audio analysis
US9113001B2 (en) Systems, methods, and media for disambiguating call data to determine fraud
US8571853B2 (en) Method and system for laughter detection
EP2622832B1 (en) Speech comparison
US20150381801A1 (en) Systems, methods, and media for disambiguating call data to determine fraud
US20110082874A1 (en) Multi-party conversation analyzer & logger
US20120155663A1 (en) Fast speaker hunting in lawful interception systems
WO2008096336A2 (en) Method and system for laughter detection
EP1662483A1 (en) Method and apparatus for speaker spotting
CN116665708A (en) Illegal service operation detection system and method thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

NENP Non-entry into the national phase

Ref country code: RU

WWW Wipo information: withdrawn in national office

Country of ref document: RU

122 Ep: pct application non-entry in european phase

Ref document number: 06701522

Country of ref document: EP

Kind code of ref document: A2