US20100010895A1 - Prediction of a degree of relevance between query rewrites and a search query - Google Patents

Prediction of a degree of relevance between query rewrites and a search query Download PDF

Info

Publication number
US20100010895A1
US20100010895A1 US12/169,218 US16921808A US2010010895A1 US 20100010895 A1 US20100010895 A1 US 20100010895A1 US 16921808 A US16921808 A US 16921808A US 2010010895 A1 US2010010895 A1 US 2010010895A1
Authority
US
United States
Prior art keywords
relevance
advertisements
query
degree
search query
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/169,218
Inventor
Evgeniy Gabrilovich
Donald Metzler
Vanja Josifovski
Andrei Broder
Vassilis Plachouras
Vanessa Murdock
Massimiliano Ciaramita
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yahoo Inc
Original Assignee
Yahoo Inc until 2017
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yahoo Inc until 2017 filed Critical Yahoo Inc until 2017
Priority to US12/169,218 priority Critical patent/US20100010895A1/en
Assigned to YAHOO! INC. reassignment YAHOO! INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JOSIFOVSKI, VANJA, MURDOCK, VANESSA, CIARAMITA, MASSIMILIANO, PLACHOURAS, VASSILIS, BRODER, ANDREI, METZLER, DONALD, GABRILOVICH, EVGENIY
Priority to PCT/US2009/048681 priority patent/WO2010005801A2/en
Publication of US20100010895A1 publication Critical patent/US20100010895A1/en
Assigned to YAHOO HOLDINGS, INC. reassignment YAHOO HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YAHOO! INC.
Assigned to OATH INC. reassignment OATH INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YAHOO HOLDINGS, INC.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0241Advertisements
    • G06Q30/0251Targeted advertisements
    • G06Q30/0255Targeted advertisements based on user history
    • G06Q30/0256User search
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising

Definitions

  • Online advertisement service providers such as Yahoo! Inc., serve advertisements for placement on a webpage based on bid phrases associated with advertisements and keywords within search queries received at a sponsored search web server.
  • ad providers may rely on query rewrites to provide broader search coverage.
  • a query rewrite corresponds to a set of terms that may relate to the original search query to varying degrees. When query rewrites are utilized, advertisements associated with keywords within the query rewrites may be served as well.
  • the relatedness or relevance between a search query and a query rewrite may vary. That is, some query rewrites may be more relevant to the original search query than others. For example, the rewrite “automobile” may be more related or relevant to the search query “car” than the rewrite “travel.” Serving advertisements based on rewrites that are not relevant to a search query both frustrates advertisers, whose advertisements are not being displayed to interested potential customers, and users who are viewing advertisements that are not relevant to a submitted search query.
  • FIG. 1 is a diagram of a system for predicting a degree of relevance between query rewrites and a search query
  • FIG. 2 is a flow diagram describing an operation of the system in FIG. 1 in a first embodiment
  • FIG. 3 is a flow diagram describing an operation of the system in FIG. 1 in a second embodiment
  • FIG. 4 is a flow diagram for predicting a degree of relevance between a search query and advertisements associated with a query rewrite
  • FIG. 5 is a flow chart for generating a prediction model to predict a degree of relevance between advertisements and search queries.
  • FIG. 6 illustrates a general computer system, which may represent a sponsored search web server, terminal, or any of the other computing devices referenced herein.
  • the present disclosure is directed to systems and methods for predicting a degree of relevance between query rewrites and a search query. Determining a degree of relevance between a query rewrite and a search query before serving the advertisements based on the query rewrite allows an ad provider to improve the accuracy of the advertisements it serves. By improving the accuracy of served advertisements, advertiser satisfaction with the ad provider is increased because the advertisements of the advertiser are being displayed to interested customers. Additionally, improving the accuracy of served advertisements increases user satisfaction because the users are being shown advertisements for products or services in which the user may actually be interested.
  • FIG. 1 is a diagram of a system 100 for predicting a degree of relevance between query rewrites and a search query.
  • the system 100 includes a sponsored search web server 105 in communication with a query rewrite database 110 , an advertisement database 115 , and a relevance module 155 . Also shown is a terminal 120 that communicates with the system.
  • the sponsored search web server 105 may include suitable logic, code, and/or circuitry that may enable generating web pages, including sponsored search web pages with a search result list and a list of advertisements.
  • the search result list and list of advertisements may be associated with a search query 125 communicated from the terminal 120 .
  • the sponsored search web server 105 may correspond to an Intel® based computer running applications such as Apache® or Microsoft Internet Information Server®, which may be utilized to generate the web pages.
  • the sponsored search web server 105 may be implemented using any conventional computer or other data processing device.
  • the sponsored search web server 105 may further be implemented using a specialized data processing device which has been particularly adapted to perform the functions of a sponsored search web server 105 . These functions may include communicating with a user operating an Internet browser running on a terminal 125 .
  • the sponsored search web server 105 may also be adapted to communicate with other networked equipment and to retrieve information from various databases, such as a query rewrite database 110 , and/or an advertisement database 115 .
  • the terminal 120 may include suitable logic, code, and/or circuitry that may enable communicating information over a network connection, such as an Internet connection.
  • a network connection such as an Internet connection.
  • the terminal 120 may correspond to an Intel® based computer running a Windows® operating system with a browser, such as Internet Explorer®.
  • the terminal 120 may be adapted to communicate a search query 125 to the sponsored search web server 115 and to display web pages communicated from a web server, such as a search result list generated by a sponsored search web server 105 .
  • the query rewrite database 110 may include information for relating a query terms 130 from a search query 125 specified by a user at the terminal 125 to rewrites 135 .
  • the query rewrite database 110 may also include information corresponding to a relevance attribute 140 for specifying the degree to which a query term 130 and a rewrite 135 relate to one another.
  • a search query 125 with the query term 130 “camera” may be related to the rewrites 135 “digital camera”, “photography”, and “film”, as shown in FIG. 1 .
  • the rewrite 135 “digital camera” is more related or relevant to the query term 130 “camera” than the rewrite 135 “film.”
  • the relevance attribute 140 for “digital camera” may be higher than the relevance attribute 140 for “film.”
  • the advertisement database 115 may include information for associating terms 145 with a plurality of advertisements 150 .
  • the terms 145 may correspond to terms in a search query 125 specified by a user at the terminal 120 and/or rewrites 135 stored in the query rewrite database 110 that are associated with search queries 125 .
  • Advertisements 150 may have been previously associated with the terms 145 via, for example, a bidding process where advertisers bid on keywords or terms 145 .
  • the information communicated from the advertisement database 115 may include data defining text, images, video, audio or other information, such as links to another computer database include the advertisement data.
  • the relevance module may include suitable logic, code, and/or circuitry that may enable predicting the relevance between a query term and a query rewrite and also for predicting the relevance between a query term and an advertisement.
  • the relevance module 155 may reside within the sponsored search web server 105 or in another computer (not shown) in communication with the sponsored search web server and/or the query rewrites database 110 and advertisement database 115 .
  • the relevance module may be utilized to specify the relevance attribute 140 associated with a query term 130 and a rewrite 135 located in the query rewrite database 110 .
  • FIG. 2 is a flow diagram describing an operation of the system 100 ( FIG. 1 ) in a first embodiment.
  • the system 100 may receive a search query.
  • a user at a terminal 120 may navigate to a sponsored search web page hosted by the sponsored search web server 105 and specify a search query 125 , such as “camera.”
  • relevant rewrites may be located.
  • the sponsored search web server 105 may search through a query rewrite database 110 to locate query rewrites related or relevant to the search query “camera” specified by the user. In this case, the rewrites “digital camera”, “photography”, and “film” may be located.
  • advertisements associated with the relevant rewrites may be served or delivered.
  • the sponsored search web server 105 may serve or deliver advertisements specified in the advertisement database 115 and associated with the rewrites “digital camera”, “photography”, and “film” to the user at the terminal 120 as part of a sponsored search result web page.
  • the number of rewrites utilized may be limited to those that have the highest relevance. At least one advantage of this approach is that relevant rewrites are utilized. This helps ensure that the advertisements presented to the user at the terminal 120 are better targeted.
  • FIG. 3 is a flow diagram describing an operation of the system 100 ( FIG. 1 ) in a second embodiment.
  • the system 100 may receive a search query and at block 305 , relevant rewrites may be located as described above with reference to FIG. 2 .
  • relevant advertisements associated with the relevant rewrites may be retrieved and delivered to the user as part of a sponsored search result web page. In doing so, a determination may be made as to whether an advertisement associated with a rewrite is relevant to the original search query. Once the determination is made, the relevant advertisements may be served or delivered to the user at the terminal 120 . This approach improves the targeting of the advertisements further because the advertisements served are the relevant advertisements of the relevant rewrites rather than the non-relevant advertisements of the relevant rewrites.
  • FIG. 4 is a flow diagram for predicting a degree of relevance between a search query and advertisements associated with a query rewrite.
  • a search query may be received.
  • a user at a terminal 120 may specify a search query 125 via a sponsored search web page hosted by a sponsored search web server 105 .
  • all the rewrites associated with the search query 125 may be retrieved.
  • the rewrites may have been previously associated with the search query 125 by human operators or via statistical processes for associating rewrites with the search queries. For example, the choice of key words selected by advertisers for an advertisement may be utilized to generate the rewrites.
  • a plurality of advertisements associated with each rewrite may be retrieved.
  • the plurality of advertisements may have been previously associated with the rewrites by human operators or automatically. For example, an advertiser may have bid on key words within the rewrite. In doing so, the advertiser's advertisements may become associated with the rewrite.
  • the relevance between each advertisement of the plurality of advertisements and the received search query may be determined by extracting a set of features indicative of the relatedness of the advertisement and the search query and passing the extracted features through a prediction module for predicting the relevance.
  • the prediction module corresponds to a parameterized set of features belonging to advertisements and search queries of known relatedness to one another.
  • the relatedness or relevance between a new advertisement and new search query may be determined by comparing the features extracted from the new advertisement and new search query to the features extracted from advertisements and search queries of known relatedness to one another.
  • the overall relevance between the rewrite and the received search query may be determined based on the relevance between the plurality of advertisements associated with the rewrite and the original search query.
  • the relevance between the rewrite and the received search query may correspond to the average relevance between all the advertisements associated with the rewrite and the search query.
  • the value corresponding to the relevance may be stored in a database, such as the query rewrite database 110 shown in FIG. 1 .
  • FIG. 5 is a flow chart for generating a prediction model to predict a degree of relevance between advertisements and search queries.
  • a training set may be constructed by presenting a plurality of advertisements and search queries to a human operator and receiving an indication from the human operator at block 505 as to whether the presented plurality of advertisements are relevant to the search queries.
  • the human operator may indicate that the plurality of advertisements is relevant to a query or is not relevant to the query.
  • the human operator may indicate a degree of relevance between the plurality of advertisements and the query on a scale, such as zero to ten.
  • a system such as the system 100 shown in FIG. 1 may implicitly determine a degree of relevance between the plurality of advertisements and the queries based on click-through information available in sources such as search logs. For example, if Internet users typically click on an advertisement when displayed in response to a given search query, the system 100 may infer that the advertisement is relevant to the search query.
  • a set of features may be extracted from the advertisements and search queries via the relevance module 155 shown in FIG. 1 .
  • a feature typically measures the relatedness or a degree of relevance between the advertisements and search query, measures an overall quality of the advertisements, or measures a relationship between the advertisements themselves.
  • the set of features may include information regarding an advertisement and/or search query with respect to word overlap, cosine similarity, translation, pointwise mutual information, chi-squared, bid price, score coefficient of variation, and topical cohesiveness, each of which is described below.
  • Word overlap is a feature that measures a degree to which terms, also known as keywords or bid phrases, associated with the plurality of advertisements overlap with terms in the content of the search query.
  • the relevance module may create a word overlap score based on whether all the terms associated with the advertisement are present in the content of the search query, whether none of the terms associated with the advertisement are present in the content of the search query, or a proportion of the terms associated with the advertisement that are present in the content of the search query.
  • the word overlap score of each advertisement is then aggregated to calculate a word overlap score of the plurality of advertisements and the content of the search query.
  • the relevance module may calculate four values associated with the feature using the equations:
  • X min ⁇ ( P , A ) min ⁇ ⁇ A ⁇ X ⁇ ( P , A )
  • X max ⁇ ( P , A ) max ⁇ ⁇ A ⁇ X ⁇ ( P , A )
  • A is the plurality of advertisements
  • P is the search query
  • SCORE(P,A) is an ad score returned by an ad provider for an advertisement with respect to terms from the search query.
  • An ad score is typically a measure of the degree of relevance between an advertisement and a keyword.
  • X min (P,A) results in a minimum feature value associated with an advertisement of the plurality of advertisements and search query content.
  • a plurality of advertisements may include a first advertisement, a second advertisement, a third advertisement, a fourth advertisement, and a fifth advertisement.
  • the first advertisement is associated with a word overlap score of 1
  • the second advertisement is associated with a word overlap score of 2
  • the third advertisement is associated with a word overlap score of 3
  • the fourth advertisement is associated with a word overlap score of 4
  • the fifth advertisement is associated with a word overlap score of 5.
  • the X min (P,A) of the word overlap feature for the plurality of advertisements is 1 because 1 is the lowest word overlap score associated with one of the advertisements of the plurality of advertisements.
  • X max (P,A) results in a maximum feature value associated with an advertisement of the plurality of advertisements and search query content.
  • the X max (P,A) of the word overlap feature of the plurality of advertisements is 5 because 5 is the greatest word overlap score associated with one of the advertisements of the plurality of advertisements.
  • X mean (P,A) results in a mean of the feature values associated with the advertisements of the plurality of advertisements and search query content.
  • X mean (P,A) of the word overlap feature is 3 because 3 is the average of the word overlap scores associated with the advertisements of the plurality of advertisements.
  • X wmean (P,A) results in a mean of the feature values associated with the advertisements of the plurality of advertisements and search query content that has been weighted based on an ad score associated with each advertisement of the plurality of advertisements.
  • the first advertisement is associated with an ad score of 1
  • the second advertisement is associated with an ad score of 2
  • the third advertisement is associated with an ad score of 3
  • the fourth advertisement is associated with an ad score of 4
  • the fifth advertisement is associated with an ad score of 5
  • X wmean (P,A) of the word overlap feature is calculated to be 3.67.
  • Cosine similarity is a feature that measures a degree to which terms associated with the plurality of advertisements overlap with terms in the content of the search query, with a score that has been weighted based on a number of times a term appears in both the plurality of advertisements and the content of the search query.
  • the cosine similarity feature may be calculated using the equation:
  • sim ⁇ ( P , A ) ⁇ t ⁇ P ⁇ A ⁇ w Pt ⁇ w At ⁇ t ⁇ P ⁇ w Pt 2 ⁇ ⁇ t ⁇ A ⁇ w At 2
  • w Pt weight with respect to search query and term
  • w At weight with respect to advertisement and term
  • tf idf the term frequency-inverse document frequency
  • the tf idf weight w Pt of term t in the search query may be computed using the equation:
  • tf is term frequency
  • N is the total number of advertisements in the plurality of advertisements
  • n t is the number of advertisements in the plurality of advertisements in which term t occurs.
  • the weight w At of term t in the plurality of advertisements may be computed in the same way.
  • Translation is a feature that measures a degree of topical relationship between the plurality of advertisements and the content of the search query.
  • the relevance module generally computes a probability that two terms (in the same language) are associated with each other, such that one term appears in the plurality of advertisements and the other term appears in the search query content.
  • the translation feature indicates a degree of topical relationship between a plurality of advertisements and search query content even though the same term does not appear in both the plurality of advertisements and the content of the search query, as required by features such as word overlap and cosine similarity. For example, if the plurality of advertisements includes the term “old cars” and the content of the search query includes the term “antique automobiles,” the translation feature would indicate that the plurality of advertisements and the content of the search query are related due to the relationship between the terms “old cars” and “antique automobiles.”
  • the relevance module may build translation tables such as those described in Y. Al-Onaizan, J. Curin, M. Jahr, K. Knight, J. Lafferty, D. Melamed, F. J. Och, D. Purdy, N. A. Smith, and D. Yarowsky, Statistical Machine Translation, Final Report, JHU workshop, 1999; P. F. Brown, J. Cocke, S. A. Della Pietra, V. J. Della Pietra, F. Jelineck, J. D. Lafferty, R. L. Mercer, and P.
  • the translation tables provide a distribution of a probability of a first term translating to a second term, given an alignment between two sentences, and other information such as how likely a term is to have many other translations, the relative distance between two terms in their respective sentences, and the appearance of words in common classes of words.
  • the relevance module may compute a probability that two terms (in the same language) are associated with each other, such that one term appears in the plurality of advertisements and the other term appears in the search query content.
  • the relevance module concatenates the plurality of advertisements to form a meta-document, also known as a “source.”
  • the relevance module also concatenates the search query content to form a second meta-document, also known as a “target.”
  • the “source” and “target” are known collectively as a “parallel corpus.”
  • the relevance module determines a number of times a term in the source is associated with a term in the target, and normalizes the total number of times the term was found in the source.
  • the relevance module then computes an alignment between the source and the target by assuming that a pair of terms with a highest probability are aligned with each other, and then aligning the remaining terms in each of the source and target sentence pairs accordingly. It should be appreciated that each term in the source may be aligned with one term in the target, but that each term in the target may be aligned with any number of terms in the source, because the relevance module iterates over source terms and looks at each term one time.
  • the relevance module then re-estimates a number of times a source term is associated with a target term, given the alignment described above.
  • the above-described blocks of estimating probabilities, adjusting the alignment to maximize the probabilities, and re-estimating the probabilities are repeated until the probabilities do not change, or change only a very small amount.
  • the relevance module may improve the alignment by limiting a number of words a term in the target is allowed to translate to; by preventing words at the beginning of the source sentence from translating to words at the ends of the target sentence; and/or by grouping words together that are similar in meaning or semantic context and aligning words that appear in the same group.
  • the relevance module may calculate a translation score of the plurality of advertisements and the content of the search query based on factors such as an average of the translation properties of all terms in the content of the search query translating to all terms in a title and description of a candidate advertisement, or a proportion of terms in the content of a search query that have a translation in a title or description of an advertisement.
  • Pointwise mutual information and chi-squared are features that measure a degree of relevance between the plurality of advertisements and the content of the search query based on a co-occurrence of terms. For example, if an advertisement includes both the term “automobile” and the term “car”, and the content of a search query includes both the term “automobile” and the term “car”, because the terms “automobile” and “car” are related and appear in both the advertisement and the search query content, pointwise mutual information and chi-squared information will indicate that the advertisement and the search query content are related.
  • pointwise mutual information may be calculated using the equation:
  • t 1 is a term from the search query content
  • t 2 is a term from an advertisement
  • P(t) is a probability that term t appears anywhere on the Internet
  • P(t 1 ,t 2 ) is a probability that terms t 1 and t 2 occur in the same search query.
  • P(t) may be calculated by dividing the number of search queries that occur on the Internet where term t is present divided by the total number of search queries that occur on the Internet.
  • P(t 1 ,t 2 ) may be calculated by dividing the number of search queries that occur on the Internet where terms t 1 and t 2 are present divided by the total number of search query that occur on the Internet. It will be appreciated that a number of search queries that occur on the Internet may be approximated based on a number of search queries indexed by a commercial search engine.
  • the relevance module forms pairs of terms t 1 and t 2 for the pointwise mutual information calculation by extracting a top number of terms, such as the top 50 terms, based on the tf idf weight of the terms in a search query.
  • chi-squared may be calculated using the equation:
  • X 2 ⁇ L ⁇ ⁇ ( o 11 ⁇ o 12 - o 12 ⁇ o 21 ) 2 ( o 11 + o 12 ) ⁇ ( o 11 + o 21 ) ⁇ ( o 12 + o 22 ) ⁇ ( o 21 + o 22 )
  • o 11 stands for the number of search queries available on the Internet that contain both terms t 1 and t 2
  • o 12 stands for the number of search queries on the Internet in which t 2 occurs but t 1 does not occur.
  • is a number of search queries appearing in one or more search logs
  • o 11 stands for the number of search queries in the search logs that contain both terms t 1 and t 2
  • o 12 stands for the number of search queries in the search logs in which t 2 occurs but t 1 does not occur.
  • the relevance module computes the chi-squared statistic (X 2 ) for each advertisement and the search query content, and counts the number of pairs of terms for which the chi-squared statistic is above a threshold, such as 95%. It will be appreciated that if the chi-squared statistic for a pair of terms is above the threshold, the pair of terms is related. Therefore, the more pairs of terms between the plurality of advertisements and the search query content that are related, the more likely it is that the plurality of advertisements and the search query content are related.
  • Bid price is a feature that may indicate an overall quality of a plurality of advertisements. For example, if the advertisements of the plurality of advertisements are associated with a large bid price for a term obtained from the content of the search query, the fact that an advertiser is wiling to pay a large amount for an action associated with their advertisement is likely an indication that an advertisement is of a high quality. Therefore, the plurality of advertisements is likely of a high overall quality.
  • the plurality of advertisements is likely of a low overall quality.
  • Coefficient of variation is a feature that measures a degree of variance of ad scores between the advertisements of the plurality of advertisements.
  • an ad score is a value that represents a degree of relevance between an advertisement and a keyword.
  • the relevance module typically uses coefficient of variation information instead of a standard deviation or variance information because coefficient of variation information is normalized with respect to a mean of the ad score.
  • the relevance module may calculate a coefficient of variation using the equation:
  • ⁇ SCORE is a standard deviation of the ad scores of the advertisements in the plurality of advertisements and ⁇ SCORE is a mean of the ad scores of the advertisements in the plurality of advertisements.
  • Topical cohesiveness is a feature that measures how topically related the advertisements of the plurality of advertisements are to each other. For example, if a term “cheap hotels” is obtained from the content of a search query and the bid phrases associated with the plurality of advertisements are “cheap cars,” “hotel discounts,” and “swimming pools,” then the plurality of advertisements have a low topical cohesiveness since they relate to very different topics. However, if the term “cheap hotels” is obtained from the content of the search query and the bid phrases associated with the plurality of advertisements are “hotel discounts,” “inexpensive hotels,” and “vacation hotels,” then the results are more topically cohesive and more likely to be satisfying to an Internet user.
  • the advertisements of the plurality of advertisements will also be topically related. Conversely, if the plurality of advertisements is of a low quality, the advertisements of the plurality of advertisements are typically not topically related.
  • the topical cohesive feature is typically used in conjunction with other features, such as the word overlap, cosine similarity, pointwise mutual information, and chi-squared features described above, to determine a degree of relevance between advertisements and the content of a search query or a search query.
  • the relevance module may build a relevance model over terms and/or semantic classes. With respect to terms, the relevance module may first build a statistical model using the equation:
  • ⁇ w ⁇ ⁇ ⁇ A ⁇ P ⁇ ( w
  • A) is a likelihood that term w is present in an advertisement, as explained below
  • WP) is a likelihood of an advertisement given the search query (WP), as explained below
  • ⁇ w is shorthand for P(w
  • WP), may be estimated using the equation:
  • SCORE(WP,A) is an ad score for an advertisement given a search query.
  • ⁇ w is estimated using the equations described above, it is often referred to in information retrieval literature as a relevance model.
  • the relevance module may generate a number of semantic classes associated with the advertisement and a score associated with the advertisement and the semantic class.
  • a semantic class is a topical classification that an advertisement may relate to. Examples of semantic classes include topics such as entertainment, automobile, and sports. Further, each semantic class may include subclasses, such as golf or tennis for the semantic class sports. It will be appreciated that this hierarchy may continue such that each subclass includes further subclasses.
  • each advertisement includes the term “cars”
  • the entropy of the relevance model would be 0, since the relevance model would be peaked around the term “cars” since P(cars
  • model) 1 and P(other words
  • model) 0.
  • a first advertisement includes the term “cat”
  • a second advertisement includes the term “dog”
  • a third advertisement includes the term “rabbit”
  • a fourth advertisement includes the term “turtle”
  • a fifth advertisement includes the term “fish”
  • the entropy of the relevance model would be very large, since the distribution is spread across five different terms, instead of just one.
  • the relevance module may calculate an entropy score using the equation:
  • H ⁇ ( ⁇ ) - ⁇ w ⁇ V ⁇ ⁇ w ⁇ log ⁇ ⁇ ⁇ w
  • the method After extracting the set of features from the plurality of advertisements and the content of the search query at block 510 , the method loops to block 500 and the above-described process is repeated for another plurality of advertisements and another search query. This process is repeated until at block 515 the relevance module generates a prediction model that may be utilized to predict whether a set of candidate advertisements is relevant to the content of a set of search queries based on the indications of relevance received from one or more human operators received at block 505 and the set of features extracted at block 510 . In one implementation, the relevance module generates the prediction model using machine-learning algorithms.
  • FIG. 6 illustrates a general computer system, which may represent a sponsored search web server 105 , terminal 120 , or any of the other computing devices referenced herein.
  • the computer system 600 may include a set of instructions 645 that may be executed to cause the computer system 600 to perform any one or more of the methods or computer based functions disclosed herein.
  • the computer system 600 may operate as a standalone device or may be connected, e.g., using a network, to other computer systems or peripheral devices.
  • the computer system 600 may include a processor 605 , such as, a central processing unit (CPU), a graphics processing unit (GPU), or both.
  • the processor 605 may be a component in a variety of systems.
  • the processor 605 may be part of a standard personal computer or a workstation.
  • the processor 605 may be one or more general processors, digital signal processors, application specific integrated circuits, field programmable gate arrays, servers, networks, digital circuits, analog circuits, combinations thereof, or other now known or later developed devices for analyzing and processing data.
  • the processor 605 may implement a software program, such as code generated manually (i.e., programmed).
  • the computer system 600 may include a memory 610 that can communicate via a bus 620 .
  • the advertisement database 115 and the query rewrite database may be stored in the memory.
  • the memory 610 may be a main memory, a static memory, or a dynamic memory.
  • the memory 610 may include, but may not be limited to computer readable storage media such as various types of volatile and non-volatile storage media, including but not limited to random access memory, read-only memory, programmable read-only memory, electrically programmable read-only memory, electrically erasable read-only memory, flash memory, magnetic tape or disk, optical media and the like.
  • the memory 610 may include a cache or random access memory for the processor 605 .
  • the memory 610 may be separate from the processor 605 , such as a cache memory of a processor, the system memory, or other memory.
  • the memory 610 may be an external storage device or database for storing data. Examples may include a hard drive, compact disc (“CD”), digital video disc (“DVD”), memory card, memory stick, floppy disc, universal serial bus (“USB”) memory device, or any other device operative to store data.
  • the memory 610 may be operable to store instructions 645 executable by the processor 605 .
  • the functions, acts or tasks illustrated in the figures or described herein may be performed by the programmed processor 605 executing the instructions 645 stored in the memory 610 .
  • the computer system 600 may further include a display 630 , such as a liquid crystal display (LCD), an organic light emitting diode (OLED), a flat panel display, a solid state display, a cathode ray tube (CRT), a projector, a printer or other now known or later developed display device for outputting determined information.
  • a display 630 such as a liquid crystal display (LCD), an organic light emitting diode (OLED), a flat panel display, a solid state display, a cathode ray tube (CRT), a projector, a printer or other now known or later developed display device for outputting determined information.
  • the display 630 may act as an interface for the user to see the functioning of the processor 605 , or specifically as an interface with the software stored in the memory 610 or in the drive unit 615 .
  • the computer system 600 may include an input device 630 configured to allow a user to interact with any of the components of system 600 .
  • the input device 625 may be a number pad, a keyboard, or a cursor control device, such as a mouse, or a joystick, touch screen display, remote control or any other device operative to interact with the system 600 .
  • the computer system 600 may also include a disk or optical drive unit 615 .
  • the disk drive unit 615 may include a computer-readable medium 640 in which one or more sets of instructions 645 , e.g. software, can be embedded. Further, the instructions 645 may perform one or more of the methods or logic as described herein.
  • the instructions 645 may reside completely, or at least partially, within the memory 610 and/or within the processor 605 during execution by the computer system 600 .
  • the memory 610 and the processor 605 also may include computer-readable media as discussed above.
  • the present disclosure contemplates a computer-readable medium 640 that includes instructions 645 or receives and executes instructions 645 responsive to a propagated signal; so that a device connected to a network 650 may communicate voice, video, audio, images or any other data over the network 650 .
  • the instructions 645 may be implemented with hardware, software and/or firmware, or any combination thereof. Further, the instructions 645 may be transmitted or received over the network 650 via a communication interface 635 .
  • the communication interface 635 may be a part of the processor 605 or may be a separate component.
  • the communication interface 635 may be created in software or may be a physical connection in hardware.
  • the network 650 may include wired networks, wireless networks, or combinations thereof. Information related to business organizations may be provided via the network 650 .
  • the wireless network may be a cellular telephone network, an 802.11, 802.16, 802.20, or WiMax network.
  • the network 650 may be a public network, such as the Internet, a private network, such as an intranet, or combinations thereof, and may utilize a variety of networking protocols now available or later developed including, but not limited to TCP/IP based networking protocols.
  • the computer-readable medium 640 may be a single medium, or the computer-readable medium 640 may be a single medium or multiple media, such as a centralized or distributed database, and/or associated caches and servers that store one or more sets of instructions.
  • the term “computer-readable medium” may also include any medium that may be capable of storing, encoding or carrying a set of instructions for execution by a processor or that may cause a computer system to perform any one or more of the methods or operations disclosed herein.
  • the computer-readable medium 640 may include a solid-state memory such as a memory card or other package that houses one or more non-volatile read-only memories.
  • the computer-readable medium 640 also may be a random access memory or other volatile re-writable memory.
  • the computer-readable medium 640 may include a magneto-optical or optical medium, such as a disk or tapes or other storage device to capture carrier wave signals such as a signal communicated over a transmission medium.
  • a digital file attachment to an e-mail or other self-contained information archive or set of archives may be considered a distribution medium that may be a tangible storage medium. Accordingly, the disclosure may be considered to include any one or more of a computer-readable medium or a distribution medium and other equivalents and successor media, in which data or instructions may be stored.
  • dedicated hardware implementations such as application specific integrated circuits, programmable logic arrays and other hardware devices, may be constructed to implement one or more of the methods described herein.
  • Applications that may include the apparatus and systems of various embodiments may broadly include a variety of electronic and computer systems.
  • One or more embodiments described herein may implement functions using two or more specific interconnected hardware modules or devices with related control and data signals that may be communicated between and through the modules, or as portions of an application-specific integrated circuit. Accordingly, the present system may encompass software, firmware, and hardware implementations.
  • the embodiments disclosed herein provide an approach for predicting a degree of relevance between query rewrites and a search query.
  • a relevance model to predict a degree of relevance between the query rewrites and search query before serving advertisements, an ad provider is able to more accurately serve relevant advertisements.

Abstract

A predictor for determining a degree of relevance between a query rewrite and a search query is provided. The predictor may receive a search query from a user via a terminal and identify a set of candidate query rewrites associated with the search query. The predictor may then extract a set of features from advertisements associated with the query rewrites and the search query and determine a degree of relevance between the advertisements and the search query based on a prediction model. The predictor may then determine the degree of relevance between the rewrites and the search query based on the determined degree of relevance between the advertisements and the search query.

Description

    BACKGROUND
  • Online advertisement service providers (ad providers), such as Yahoo! Inc., serve advertisements for placement on a webpage based on bid phrases associated with advertisements and keywords within search queries received at a sponsored search web server. In some instances, ad providers may rely on query rewrites to provide broader search coverage. A query rewrite corresponds to a set of terms that may relate to the original search query to varying degrees. When query rewrites are utilized, advertisements associated with keywords within the query rewrites may be served as well.
  • However, as noted above, the relatedness or relevance between a search query and a query rewrite may vary. That is, some query rewrites may be more relevant to the original search query than others. For example, the rewrite “automobile” may be more related or relevant to the search query “car” than the rewrite “travel.” Serving advertisements based on rewrites that are not relevant to a search query both frustrates advertisers, whose advertisements are not being displayed to interested potential customers, and users who are viewing advertisements that are not relevant to a submitted search query.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a diagram of a system for predicting a degree of relevance between query rewrites and a search query;
  • FIG. 2 is a flow diagram describing an operation of the system in FIG. 1 in a first embodiment;
  • FIG. 3 is a flow diagram describing an operation of the system in FIG. 1 in a second embodiment;
  • FIG. 4 is a flow diagram for predicting a degree of relevance between a search query and advertisements associated with a query rewrite;
  • FIG. 5 is a flow chart for generating a prediction model to predict a degree of relevance between advertisements and search queries; and
  • FIG. 6 illustrates a general computer system, which may represent a sponsored search web server, terminal, or any of the other computing devices referenced herein.
  • DETAILED DESCRIPTION OF THE DRAWINGS
  • The present disclosure is directed to systems and methods for predicting a degree of relevance between query rewrites and a search query. Determining a degree of relevance between a query rewrite and a search query before serving the advertisements based on the query rewrite allows an ad provider to improve the accuracy of the advertisements it serves. By improving the accuracy of served advertisements, advertiser satisfaction with the ad provider is increased because the advertisements of the advertiser are being displayed to interested customers. Additionally, improving the accuracy of served advertisements increases user satisfaction because the users are being shown advertisements for products or services in which the user may actually be interested.
  • FIG. 1 is a diagram of a system 100 for predicting a degree of relevance between query rewrites and a search query. The system 100 includes a sponsored search web server 105 in communication with a query rewrite database 110, an advertisement database 115, and a relevance module 155. Also shown is a terminal 120 that communicates with the system.
  • The sponsored search web server 105 may include suitable logic, code, and/or circuitry that may enable generating web pages, including sponsored search web pages with a search result list and a list of advertisements. The search result list and list of advertisements may be associated with a search query 125 communicated from the terminal 120. The sponsored search web server 105 may correspond to an Intel® based computer running applications such as Apache® or Microsoft Internet Information Server®, which may be utilized to generate the web pages. The sponsored search web server 105 may be implemented using any conventional computer or other data processing device. The sponsored search web server 105 may further be implemented using a specialized data processing device which has been particularly adapted to perform the functions of a sponsored search web server 105. These functions may include communicating with a user operating an Internet browser running on a terminal 125. The sponsored search web server 105 may also be adapted to communicate with other networked equipment and to retrieve information from various databases, such as a query rewrite database 110, and/or an advertisement database 115.
  • The terminal 120 may include suitable logic, code, and/or circuitry that may enable communicating information over a network connection, such as an Internet connection. For example, the terminal 120 may correspond to an Intel® based computer running a Windows® operating system with a browser, such as Internet Explorer®. The terminal 120 may be adapted to communicate a search query 125 to the sponsored search web server 115 and to display web pages communicated from a web server, such as a search result list generated by a sponsored search web server 105.
  • The query rewrite database 110 may include information for relating a query terms 130 from a search query 125 specified by a user at the terminal 125 to rewrites 135. The query rewrite database 110 may also include information corresponding to a relevance attribute 140 for specifying the degree to which a query term 130 and a rewrite 135 relate to one another. For example, a search query 125 with the query term 130 “camera” may be related to the rewrites 135 “digital camera”, “photography”, and “film”, as shown in FIG. 1. It may be the case that the rewrite 135 “digital camera” is more related or relevant to the query term 130 “camera” than the rewrite 135 “film.” In this case, the relevance attribute 140 for “digital camera” may be higher than the relevance attribute 140 for “film.”
  • The advertisement database 115 may include information for associating terms 145 with a plurality of advertisements 150. The terms 145 may correspond to terms in a search query 125 specified by a user at the terminal 120 and/or rewrites 135 stored in the query rewrite database 110 that are associated with search queries 125. Advertisements 150 may have been previously associated with the terms 145 via, for example, a bidding process where advertisers bid on keywords or terms 145. The information communicated from the advertisement database 115 may include data defining text, images, video, audio or other information, such as links to another computer database include the advertisement data.
  • The relevance module may include suitable logic, code, and/or circuitry that may enable predicting the relevance between a query term and a query rewrite and also for predicting the relevance between a query term and an advertisement. The relevance module 155 may reside within the sponsored search web server 105 or in another computer (not shown) in communication with the sponsored search web server and/or the query rewrites database 110 and advertisement database 115. In this regard, the relevance module may be utilized to specify the relevance attribute 140 associated with a query term 130 and a rewrite 135 located in the query rewrite database 110.
  • FIG. 2 is a flow diagram describing an operation of the system 100 (FIG. 1) in a first embodiment. At block 200, the system 100 may receive a search query. For example, with reference to FIG. 1 a user at a terminal 120 may navigate to a sponsored search web page hosted by the sponsored search web server 105 and specify a search query 125, such as “camera.” At block 205, relevant rewrites may be located. For example, the sponsored search web server 105 may search through a query rewrite database 110 to locate query rewrites related or relevant to the search query “camera” specified by the user. In this case, the rewrites “digital camera”, “photography”, and “film” may be located. At block 210, advertisements associated with the relevant rewrites may be served or delivered. For example, the sponsored search web server 105 may serve or deliver advertisements specified in the advertisement database 115 and associated with the rewrites “digital camera”, “photography”, and “film” to the user at the terminal 120 as part of a sponsored search result web page. In some instances where advertising space may be limited, the number of rewrites utilized may be limited to those that have the highest relevance. At least one advantage of this approach is that relevant rewrites are utilized. This helps ensure that the advertisements presented to the user at the terminal 120 are better targeted.
  • FIG. 3 is a flow diagram describing an operation of the system 100 (FIG. 1) in a second embodiment. At block 300, the system 100 may receive a search query and at block 305, relevant rewrites may be located as described above with reference to FIG. 2. At block 310, relevant advertisements associated with the relevant rewrites may be retrieved and delivered to the user as part of a sponsored search result web page. In doing so, a determination may be made as to whether an advertisement associated with a rewrite is relevant to the original search query. Once the determination is made, the relevant advertisements may be served or delivered to the user at the terminal 120. This approach improves the targeting of the advertisements further because the advertisements served are the relevant advertisements of the relevant rewrites rather than the non-relevant advertisements of the relevant rewrites.
  • FIG. 4 is a flow diagram for predicting a degree of relevance between a search query and advertisements associated with a query rewrite. At block 400, a search query may be received. For example, with reference to FIG. 1, a user at a terminal 120 may specify a search query 125 via a sponsored search web page hosted by a sponsored search web server 105. At block 405, all the rewrites associated with the search query 125 may be retrieved. The rewrites may have been previously associated with the search query 125 by human operators or via statistical processes for associating rewrites with the search queries. For example, the choice of key words selected by advertisers for an advertisement may be utilized to generate the rewrites.
  • At block 410, a plurality of advertisements associated with each rewrite may be retrieved. The plurality of advertisements may have been previously associated with the rewrites by human operators or automatically. For example, an advertiser may have bid on key words within the rewrite. In doing so, the advertiser's advertisements may become associated with the rewrite.
  • At block 415, the relevance between each advertisement of the plurality of advertisements and the received search query may be determined by extracting a set of features indicative of the relatedness of the advertisement and the search query and passing the extracted features through a prediction module for predicting the relevance. The prediction module corresponds to a parameterized set of features belonging to advertisements and search queries of known relatedness to one another. The relatedness or relevance between a new advertisement and new search query may be determined by comparing the features extracted from the new advertisement and new search query to the features extracted from advertisements and search queries of known relatedness to one another. At block 420, the overall relevance between the rewrite and the received search query may be determined based on the relevance between the plurality of advertisements associated with the rewrite and the original search query. For example, the relevance between the rewrite and the received search query may correspond to the average relevance between all the advertisements associated with the rewrite and the search query. After determining, the relevance between the rewrite and the received query, the value corresponding to the relevance may be stored in a database, such as the query rewrite database 110 shown in FIG. 1.
  • FIG. 5 is a flow chart for generating a prediction model to predict a degree of relevance between advertisements and search queries. At block 500, a training set may be constructed by presenting a plurality of advertisements and search queries to a human operator and receiving an indication from the human operator at block 505 as to whether the presented plurality of advertisements are relevant to the search queries. In some implementations, the human operator may indicate that the plurality of advertisements is relevant to a query or is not relevant to the query. However, in other implementations the human operator may indicate a degree of relevance between the plurality of advertisements and the query on a scale, such as zero to ten.
  • In other implementations, rather than presenting a human operator with a plurality of advertisements and query at block 500 and receiving an indication of relevance at block 505, a system, such as the system 100 shown in FIG. 1 may implicitly determine a degree of relevance between the plurality of advertisements and the queries based on click-through information available in sources such as search logs. For example, if Internet users typically click on an advertisement when displayed in response to a given search query, the system 100 may infer that the advertisement is relevant to the search query.
  • At block 510, a set of features may be extracted from the advertisements and search queries via the relevance module 155 shown in FIG. 1. A feature typically measures the relatedness or a degree of relevance between the advertisements and search query, measures an overall quality of the advertisements, or measures a relationship between the advertisements themselves. In one implementation, the set of features may include information regarding an advertisement and/or search query with respect to word overlap, cosine similarity, translation, pointwise mutual information, chi-squared, bid price, score coefficient of variation, and topical cohesiveness, each of which is described below.
  • Word overlap is a feature that measures a degree to which terms, also known as keywords or bid phrases, associated with the plurality of advertisements overlap with terms in the content of the search query. For each advertisements of the plurality of advertisements, the relevance module may create a word overlap score based on whether all the terms associated with the advertisement are present in the content of the search query, whether none of the terms associated with the advertisement are present in the content of the search query, or a proportion of the terms associated with the advertisement that are present in the content of the search query. The word overlap score of each advertisement is then aggregated to calculate a word overlap score of the plurality of advertisements and the content of the search query.
  • In some implementations, for a feature X measuring a degree of relevance between advertisements and search query content such as the word overlap feature, the relevance module may calculate four values associated with the feature using the equations:
  • X min ( P , A ) = min α A X ( P , A ) X max ( P , A ) = max α A X ( P , A ) X mean ( P , A ) = α A X ( P , A ) A X wmean ( P , A ) = α A SCORE ( P , A ) · X ( P , A ) α A SCORE ( Q , A ) I
  • where A is the plurality of advertisements, P is the search query, and SCORE(P,A) is an ad score returned by an ad provider for an advertisement with respect to terms from the search query. An ad score is typically a measure of the degree of relevance between an advertisement and a keyword.
  • Xmin(P,A) results in a minimum feature value associated with an advertisement of the plurality of advertisements and search query content. For example, a plurality of advertisements may include a first advertisement, a second advertisement, a third advertisement, a fourth advertisement, and a fifth advertisement. The first advertisement is associated with a word overlap score of 1, the second advertisement is associated with a word overlap score of 2 the third advertisement is associated with a word overlap score of 3, the fourth advertisement is associated with a word overlap score of 4, and the fifth advertisement is associated with a word overlap score of 5. Accordingly, the Xmin(P,A) of the word overlap feature for the plurality of advertisements is 1 because 1 is the lowest word overlap score associated with one of the advertisements of the plurality of advertisements.
  • Xmax(P,A) results in a maximum feature value associated with an advertisement of the plurality of advertisements and search query content. Continuing with the example above, the Xmax(P,A) of the word overlap feature of the plurality of advertisements is 5 because 5 is the greatest word overlap score associated with one of the advertisements of the plurality of advertisements.
  • Xmean(P,A) results in a mean of the feature values associated with the advertisements of the plurality of advertisements and search query content. Continuing with the example above, Xmean(P,A) of the word overlap feature is 3 because 3 is the average of the word overlap scores associated with the advertisements of the plurality of advertisements.
  • Xwmean(P,A) results in a mean of the feature values associated with the advertisements of the plurality of advertisements and search query content that has been weighted based on an ad score associated with each advertisement of the plurality of advertisements. Continuing with the example above, if the first advertisement is associated with an ad score of 1, the second advertisement is associated with an ad score of 2, the third advertisement is associated with an ad score of 3, the fourth advertisement is associated with an ad score of 4, and the fifth advertisement is associated with an ad score of 5, Xwmean(P,A) of the word overlap feature is calculated to be 3.67.
  • Cosine similarity is a feature that measures a degree to which terms associated with the plurality of advertisements overlap with terms in the content of the search query, with a score that has been weighted based on a number of times a term appears in both the plurality of advertisements and the content of the search query. In one implementation, the cosine similarity feature may be calculated using the equation:
  • sim ( P , A ) = t P A w Pt w At t P w Pt 2 t A w At 2
  • where wPt (weight with respect to search query and term) and wAt (weight with respect to advertisement and term) are the term frequency-inverse document frequency (tfidf) weights of the term t in the search query and advertisement, respectively. The tfidf weighs of terms result in terms that appear a significant number of times in the plurality of advertisements and/or the search query content being given a large weight, and terms that rarely appear in the plurality of advertisements and/or the search query content also being given a large weight. For a further discussion of tfidf weights, see G. Salton and M McGill, An Introduction to Modern Information Retrieval, McGraw-Hill, 1983, ISBN 0070544840.
  • The tfidf weight wPt of term t in the search query may be computed using the equation:
  • w Pt = tf · log 2 ( N + 1 n t + 0.5 )
  • where tf is term frequency, N is the total number of advertisements in the plurality of advertisements, and nt is the number of advertisements in the plurality of advertisements in which term t occurs. The weight wAt of term t in the plurality of advertisements may be computed in the same way.
  • Translation is a feature that measures a degree of topical relationship between the plurality of advertisements and the content of the search query. As explained in more detail below, to calculate a translation score, the relevance module generally computes a probability that two terms (in the same language) are associated with each other, such that one term appears in the plurality of advertisements and the other term appears in the search query content.
  • The translation feature indicates a degree of topical relationship between a plurality of advertisements and search query content even though the same term does not appear in both the plurality of advertisements and the content of the search query, as required by features such as word overlap and cosine similarity. For example, if the plurality of advertisements includes the term “old cars” and the content of the search query includes the term “antique automobiles,” the translation feature would indicate that the plurality of advertisements and the content of the search query are related due to the relationship between the terms “old cars” and “antique automobiles.”
  • It will be appreciated that when an advertisement is translated into terms to be matched with terms from the search query content, some information regarding the full meaning of the advertisement is lost. To capture the difference between terms and a full advertisement, the relevance module may build translation tables such as those described in Y. Al-Onaizan, J. Curin, M. Jahr, K. Knight, J. Lafferty, D. Melamed, F. J. Och, D. Purdy, N. A. Smith, and D. Yarowsky, Statistical Machine Translation, Final Report, JHU workshop, 1999; P. F. Brown, J. Cocke, S. A. Della Pietra, V. J. Della Pietra, F. Jelineck, J. D. Lafferty, R. L. Mercer, and P. S. Roossin, A Statistical Approach to Machine Translation, Computational Linguistics, 16(2):79-85, 1990; and P. F. Brown, S. A. Della Pietra, V. J. Della Pietra, and R. L. Mercer, The Mathematics of Statistical Machine Translation: Parameter Estimation, Computational Linguistics 19(2):263-311, 1993.
  • The translation tables provide a distribution of a probability of a first term translating to a second term, given an alignment between two sentences, and other information such as how likely a term is to have many other translations, the relative distance between two terms in their respective sentences, and the appearance of words in common classes of words.
  • As stated above, to calculate a translation score, the relevance module may compute a probability that two terms (in the same language) are associated with each other, such that one term appears in the plurality of advertisements and the other term appears in the search query content. To compute the probability, the relevance module concatenates the plurality of advertisements to form a meta-document, also known as a “source.” The relevance module also concatenates the search query content to form a second meta-document, also known as a “target.” The “source” and “target” are known collectively as a “parallel corpus.”
  • The relevance module determines a number of times a term in the source is associated with a term in the target, and normalizes the total number of times the term was found in the source. The relevance module then computes an alignment between the source and the target by assuming that a pair of terms with a highest probability are aligned with each other, and then aligning the remaining terms in each of the source and target sentence pairs accordingly. It should be appreciated that each term in the source may be aligned with one term in the target, but that each term in the target may be aligned with any number of terms in the source, because the relevance module iterates over source terms and looks at each term one time.
  • The relevance module then re-estimates a number of times a source term is associated with a target term, given the alignment described above. The above-described blocks of estimating probabilities, adjusting the alignment to maximize the probabilities, and re-estimating the probabilities are repeated until the probabilities do not change, or change only a very small amount.
  • In some implementations, the relevance module may improve the alignment by limiting a number of words a term in the target is allowed to translate to; by preventing words at the beginning of the source sentence from translating to words at the ends of the target sentence; and/or by grouping words together that are similar in meaning or semantic context and aligning words that appear in the same group.
  • The relevance module may calculate a translation score of the plurality of advertisements and the content of the search query based on factors such as an average of the translation properties of all terms in the content of the search query translating to all terms in a title and description of a candidate advertisement, or a proportion of terms in the content of a search query that have a translation in a title or description of an advertisement.
  • Pointwise mutual information and chi-squared are features that measure a degree of relevance between the plurality of advertisements and the content of the search query based on a co-occurrence of terms. For example, if an advertisement includes both the term “automobile” and the term “car”, and the content of a search query includes both the term “automobile” and the term “car”, because the terms “automobile” and “car” are related and appear in both the advertisement and the search query content, pointwise mutual information and chi-squared information will indicate that the advertisement and the search query content are related.
  • In one implementation, pointwise mutual information may be calculated using the equation:
  • PMI ( t 1 , t 2 ) = log 2 P ( t 1 , t 2 ) P ( t 1 ) P ( t 2 )
  • where t1 is a term from the search query content, t2 is a term from an advertisement, P(t) is a probability that term t appears anywhere on the Internet, and P(t1,t2) is a probability that terms t1 and t2 occur in the same search query. In some implementations P(t) may be calculated by dividing the number of search queries that occur on the Internet where term t is present divided by the total number of search queries that occur on the Internet. Similarly, P(t1,t2) may be calculated by dividing the number of search queries that occur on the Internet where terms t1 and t2 are present divided by the total number of search query that occur on the Internet. It will be appreciated that a number of search queries that occur on the Internet may be approximated based on a number of search queries indexed by a commercial search engine.
  • In some implementations, the relevance module forms pairs of terms t1 and t2 for the pointwise mutual information calculation by extracting a top number of terms, such as the top 50 terms, based on the tfidf weight of the terms in a search query.
  • In one implementation, chi-squared may be calculated using the equation:
  • X 2 = L ( o 11 o 12 - o 12 o 21 ) 2 ( o 11 + o 12 ) ( o 11 + o 21 ) ( o 12 + o 22 ) ( o 21 + o 22 )
  • where |L| is a number of documents available on the Internet (which may be approximated based on a number of search queries indexed by a commercial search engine) and oij are defined in Table 1.
  • TABLE 1
    t1
    Figure US20100010895A1-20100114-P00001
    t1
    t2 O11 O12
    Figure US20100010895A1-20100114-P00001
    t2
    O21 O22

    For example, o11 stands for the number of search queries available on the Internet that contain both terms t1 and t2, and o12 stands for the number of search queries on the Internet in which t2 occurs but t1 does not occur. When a relevance module calculates pointwise mutual information with respect to search queries rather than search query content, |L| is a number of search queries appearing in one or more search logs, o11 stands for the number of search queries in the search logs that contain both terms t1 and t2, and o12 stands for the number of search queries in the search logs in which t2 occurs but t1 does not occur. For a further discussion on a chi-squared statistical property, see Greenwood, P. E., Nikulin, M. S., A Guide to Chi-Squared Testing, Wiley, New York, 1996, ISBN 047155779X.
  • The relevance module computes the chi-squared statistic (X2) for each advertisement and the search query content, and counts the number of pairs of terms for which the chi-squared statistic is above a threshold, such as 95%. It will be appreciated that if the chi-squared statistic for a pair of terms is above the threshold, the pair of terms is related. Therefore, the more pairs of terms between the plurality of advertisements and the search query content that are related, the more likely it is that the plurality of advertisements and the search query content are related.
  • While the features described above such as word overlap, cosine similarity, translation, pointwise mutual information, and chi-squared measure a degree of relevance between the plurality of advertisements and search query content, it will be appreciated that the features described below such as bid price, coefficient of variation, and topical cohesiveness measure how related the advertisements of the plurality of advertisements are to each other.
  • Bid price is a feature that may indicate an overall quality of a plurality of advertisements. For example, if the advertisements of the plurality of advertisements are associated with a large bid price for a term obtained from the content of the search query, the fact that an advertiser is wiling to pay a large amount for an action associated with their advertisement is likely an indication that an advertisement is of a high quality. Therefore, the plurality of advertisements is likely of a high overall quality.
  • Conversely, if a number of advertisements of the plurality of advertisements are associated with a small bid price for a term obtained from the content of the search query, the fact that an advertiser is only willing to pay a small amount for an action associated with their advertisement is likely an indication that an advertisement is of a low quality. Therefore, the plurality of advertisements is likely of a low overall quality.
  • Coefficient of variation is a feature that measures a degree of variance of ad scores between the advertisements of the plurality of advertisements. As described above, an ad score is a value that represents a degree of relevance between an advertisement and a keyword. The relevance module typically uses coefficient of variation information instead of a standard deviation or variance information because coefficient of variation information is normalized with respect to a mean of the ad score.
  • In one implementation, the relevance module may calculate a coefficient of variation using the equation:
  • COV = σ SCORE μ SCORE
  • where σSCORE is a standard deviation of the ad scores of the advertisements in the plurality of advertisements and μSCORE is a mean of the ad scores of the advertisements in the plurality of advertisements.
  • Topical cohesiveness is a feature that measures how topically related the advertisements of the plurality of advertisements are to each other. For example, if a term “cheap hotels” is obtained from the content of a search query and the bid phrases associated with the plurality of advertisements are “cheap cars,” “hotel discounts,” and “swimming pools,” then the plurality of advertisements have a low topical cohesiveness since they relate to very different topics. However, if the term “cheap hotels” is obtained from the content of the search query and the bid phrases associated with the plurality of advertisements are “hotel discounts,” “inexpensive hotels,” and “vacation hotels,” then the results are more topically cohesive and more likely to be satisfying to an Internet user.
  • Typically, if a plurality of advertisements is of a high quality, the advertisements of the plurality of advertisements will also be topically related. Conversely, if the plurality of advertisements is of a low quality, the advertisements of the plurality of advertisements are typically not topically related. However, it should be appreciated that because a plurality of advertisements may be topically related to each other, but not related to the content of a search query or a search query, the topical cohesive feature is typically used in conjunction with other features, such as the word overlap, cosine similarity, pointwise mutual information, and chi-squared features described above, to determine a degree of relevance between advertisements and the content of a search query or a search query.
  • To measure a topical cohesiveness of the plurality of advertisements, the relevance module may build a relevance model over terms and/or semantic classes. With respect to terms, the relevance module may first build a statistical model using the equation:
  • θ w = α A P ( w | A ) P ( A | WP )
  • where P(w|A) is a likelihood that term w is present in an advertisement, as explained below; P(A|WP) is a likelihood of an advertisement given the search query (WP), as explained below; and θw is shorthand for P(w|WP), which is a multinomial distribution over items w.
  • The likelihood that a term is present in an advertisement, P(w|A), may be estimated using the equation:
  • P ( w | A ) = tf w , A A
  • where tfw,A is a total number of times a term w occurs in an advertisement (A) and |A| is a total number of terms in the advertisement.
  • The likelihood of an advertisement given a search query, P(A|WP), may be estimated using the equation:
  • P ( A | WP ) = SCORE ( WP , A ) A A SCORE ( WP , A )
  • where SCORE(WP,A) is an ad score for an advertisement given a search query. When θw is estimated using the equations described above, it is often referred to in information retrieval literature as a relevance model.
  • With respect to semantic classes, for each advertisement, the relevance module may generate a number of semantic classes associated with the advertisement and a score associated with the advertisement and the semantic class. As known in the art, a semantic class is a topical classification that an advertisement may relate to. Examples of semantic classes include topics such as entertainment, automobile, and sports. Further, each semantic class may include subclasses, such as golf or tennis for the semantic class sports. It will be appreciated that this hierarchy may continue such that each subclass includes further subclasses.
  • To calculate a relevance model based on semantic classes, the relevance module may estimate P(c|A) using the equation:
  • P ( c | A ) = SCORE ( c , A ) c C SCORE ( c , A )
  • where C is a set of semantic classes and SCORE(c,A) is a score assigned by a classifier to semantic class c for advertisement A. The resulting relevance model, θc, is a multinomial distribution of the semantic classes.
  • After building a relevance model over terms or classes as described above, the relevance module may measure the cohesiveness of the relevance module. For example, the relevance module may calculate a clarity score measuring a KL-divergence between the relevance model and a collection model. For a further discussion on a clarity score, please see Steve Cronen-Townsent, Yun Zhou, and W. Bruce Croft, Predicting Query Performance, Proceedings of the 25th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, 299-306, 2002.
  • The clarity score measures how “far” the relevance model estimated from the plurality of advertisements (θ) is from the model of an entire set of advertisements ({circumflex over (θ)}) available at the ad provider, also known as an ad inventory. If the plurality of advertisements is found to be cohesive and focused on one or two topics, the relevance model will be very different from the collection model. However, if the set of topics represented by the plurality of advertisements is scattered and non-cohesive, the relevance model will be very similar to the collection model.
  • In one implementation, the clarity score may be calculated using the equation:
  • CLARITY ( θ ) = w V θ w log θ w θ ^ w
  • where {circumflex over (θ)} is the collection model, which is a maximum likelihood estimate computed over the entire collection of advertisements available at an ad provider, θw is the relevance model, and V is either the set of terms (for term relevance models) or the set of semantic classes (for semantic class relevance models).
  • The relevance model may additionally be used to calculate an entropy score. Entropy measures how “spread out” a probability distribution is. If a distribution has high entropy, then the distribution is very spread out. Conversely, if the distribution has low entropy, then the distribution is highly peaked and less spread out. By measuring the entropy of either the term relevance model or the semantic class relevance model, the entropy score measures how spread out the terms or semantic classes are with respect to the advertisements. If the entropy is high, then the term or semantic class distribution is very spread out, meaning that the advertisements are not very cohesive. However, if the entropy is low, then the term or semantic class distribution is very peaked and less spread out, meaning that the advertisements are more cohesive.
  • For example, if a term relevance model is built over five advertisements, where each advertisement includes the term “cars,” then the entropy of the relevance model would be 0, since the relevance model would be peaked around the term “cars” since P(cars|model)=1 and P(other words|model)=0. However, of the five advertisements, if a first advertisement includes the term “cat,” a second advertisement includes the term “dog,” a third advertisement includes the term “rabbit,” a fourth advertisement includes the term “turtle,” and a fifth advertisement includes the term “fish,” then the entropy of the relevance model would be very large, since the distribution is spread across five different terms, instead of just one.
  • In one implementation, the relevance module may calculate an entropy score using the equation:
  • H ( θ ) = - w V θ w log θ w
  • It will be appreciated that the calculation of an entropy score does not require the calculation of a background model as described above with respect to the clarity score.
  • In some implementations, the relevance module computes both clarity and entropy scores based on relevance models estimated from terms in an ad title, an ad description, and ad semantic classes, resulting in a total of six topical cohesiveness scores.
  • After extracting the set of features from the plurality of advertisements and the content of the search query at block 510, the method loops to block 500 and the above-described process is repeated for another plurality of advertisements and another search query. This process is repeated until at block 515 the relevance module generates a prediction model that may be utilized to predict whether a set of candidate advertisements is relevant to the content of a set of search queries based on the indications of relevance received from one or more human operators received at block 505 and the set of features extracted at block 510. In one implementation, the relevance module generates the prediction model using machine-learning algorithms.
  • Additionally, in some implementations, the relevance module may extract information from a different number of advertisements for each feature. For example, for one set of candidate advertisements, the relevance module may extract information from five advertisements of the set of candidate advertisements for the word overlap feature and extract information from ten advertisements of the set of candidate advertisements for the pointwise mutual information feature.
  • FIG. 6 illustrates a general computer system, which may represent a sponsored search web server 105, terminal 120, or any of the other computing devices referenced herein. The computer system 600 may include a set of instructions 645 that may be executed to cause the computer system 600 to perform any one or more of the methods or computer based functions disclosed herein. The computer system 600 may operate as a standalone device or may be connected, e.g., using a network, to other computer systems or peripheral devices.
  • In a networked deployment, the computer system may operate in the capacity of a server or as a client user computer in a server-client user network environment, or as a peer computer system in a peer-to-peer (or distributed) network environment. The computer system 600 may also be implemented as or incorporated into various devices, such as a personal computer (PC), a tablet PC, a set-top box (STB), a personal digital assistant (PDA), a mobile device, a palmtop computer, a laptop computer, a desktop computer, a communications device, a wireless telephone, a land-line telephone, a control system, a camera, a scanner, a facsimile machine, a printer, a pager, a personal trusted device, a web appliance, a network router, switch or bridge, or any other machine capable of executing a set of instructions 645 (sequential or otherwise) that specify actions to be taken by that machine. In one embodiment, the computer system 600 may be implemented using electronic devices that provide voice, video or data communication. Further, while a single computer system 600 may be illustrated, the term “system” shall also be taken to include any collection of systems or sub-systems that individually or jointly execute a set, or multiple sets, of instructions to perform one or more computer functions.
  • As illustrated in FIG. 6, the computer system 600 may include a processor 605, such as, a central processing unit (CPU), a graphics processing unit (GPU), or both. The processor 605 may be a component in a variety of systems. For example, the processor 605 may be part of a standard personal computer or a workstation. The processor 605 may be one or more general processors, digital signal processors, application specific integrated circuits, field programmable gate arrays, servers, networks, digital circuits, analog circuits, combinations thereof, or other now known or later developed devices for analyzing and processing data. The processor 605 may implement a software program, such as code generated manually (i.e., programmed).
  • The computer system 600 may include a memory 610 that can communicate via a bus 620. For example, the advertisement database 115 and the query rewrite database may be stored in the memory. The memory 610 may be a main memory, a static memory, or a dynamic memory. The memory 610 may include, but may not be limited to computer readable storage media such as various types of volatile and non-volatile storage media, including but not limited to random access memory, read-only memory, programmable read-only memory, electrically programmable read-only memory, electrically erasable read-only memory, flash memory, magnetic tape or disk, optical media and the like. In one case, the memory 610 may include a cache or random access memory for the processor 605. Alternatively or in addition, the memory 610 may be separate from the processor 605, such as a cache memory of a processor, the system memory, or other memory. The memory 610 may be an external storage device or database for storing data. Examples may include a hard drive, compact disc (“CD”), digital video disc (“DVD”), memory card, memory stick, floppy disc, universal serial bus (“USB”) memory device, or any other device operative to store data. The memory 610 may be operable to store instructions 645 executable by the processor 605. The functions, acts or tasks illustrated in the figures or described herein may be performed by the programmed processor 605 executing the instructions 645 stored in the memory 610. The functions, acts or tasks may be independent of the particular type of instructions set, storage media, processor or processing strategy and may be performed by software, hardware, integrated circuits, firm-ware, micro-code and the like, operating alone or in combination. Likewise, processing strategies may include multiprocessing, multitasking, parallel processing and the like.
  • The computer system 600 may further include a display 630, such as a liquid crystal display (LCD), an organic light emitting diode (OLED), a flat panel display, a solid state display, a cathode ray tube (CRT), a projector, a printer or other now known or later developed display device for outputting determined information. The display 630 may act as an interface for the user to see the functioning of the processor 605, or specifically as an interface with the software stored in the memory 610 or in the drive unit 615.
  • Additionally, the computer system 600 may include an input device 630 configured to allow a user to interact with any of the components of system 600. The input device 625 may be a number pad, a keyboard, or a cursor control device, such as a mouse, or a joystick, touch screen display, remote control or any other device operative to interact with the system 600.
  • The computer system 600 may also include a disk or optical drive unit 615. The disk drive unit 615 may include a computer-readable medium 640 in which one or more sets of instructions 645, e.g. software, can be embedded. Further, the instructions 645 may perform one or more of the methods or logic as described herein. The instructions 645 may reside completely, or at least partially, within the memory 610 and/or within the processor 605 during execution by the computer system 600. The memory 610 and the processor 605 also may include computer-readable media as discussed above.
  • The present disclosure contemplates a computer-readable medium 640 that includes instructions 645 or receives and executes instructions 645 responsive to a propagated signal; so that a device connected to a network 650 may communicate voice, video, audio, images or any other data over the network 650. The instructions 645 may be implemented with hardware, software and/or firmware, or any combination thereof. Further, the instructions 645 may be transmitted or received over the network 650 via a communication interface 635. The communication interface 635 may be a part of the processor 605 or may be a separate component. The communication interface 635 may be created in software or may be a physical connection in hardware. The communication interface 635 may be configured to connect with a network 650, external media, the display 630, or any other components in system 600, or combinations thereof. The connection with the network 650 may be a physical connection, such as a wired Ethernet connection or may be established wirelessly as discussed below. Likewise, the additional connections with other components of the system 600 may be physical connections or may be established wirelessly.
  • The network 650 may include wired networks, wireless networks, or combinations thereof. Information related to business organizations may be provided via the network 650. The wireless network may be a cellular telephone network, an 802.11, 802.16, 802.20, or WiMax network. Further, the network 650 may be a public network, such as the Internet, a private network, such as an intranet, or combinations thereof, and may utilize a variety of networking protocols now available or later developed including, but not limited to TCP/IP based networking protocols.
  • The computer-readable medium 640 may be a single medium, or the computer-readable medium 640 may be a single medium or multiple media, such as a centralized or distributed database, and/or associated caches and servers that store one or more sets of instructions. The term “computer-readable medium” may also include any medium that may be capable of storing, encoding or carrying a set of instructions for execution by a processor or that may cause a computer system to perform any one or more of the methods or operations disclosed herein.
  • The computer-readable medium 640 may include a solid-state memory such as a memory card or other package that houses one or more non-volatile read-only memories. The computer-readable medium 640 also may be a random access memory or other volatile re-writable memory. Additionally, the computer-readable medium 640 may include a magneto-optical or optical medium, such as a disk or tapes or other storage device to capture carrier wave signals such as a signal communicated over a transmission medium. A digital file attachment to an e-mail or other self-contained information archive or set of archives may be considered a distribution medium that may be a tangible storage medium. Accordingly, the disclosure may be considered to include any one or more of a computer-readable medium or a distribution medium and other equivalents and successor media, in which data or instructions may be stored.
  • Alternatively or in addition, dedicated hardware implementations, such as application specific integrated circuits, programmable logic arrays and other hardware devices, may be constructed to implement one or more of the methods described herein. Applications that may include the apparatus and systems of various embodiments may broadly include a variety of electronic and computer systems. One or more embodiments described herein may implement functions using two or more specific interconnected hardware modules or devices with related control and data signals that may be communicated between and through the modules, or as portions of an application-specific integrated circuit. Accordingly, the present system may encompass software, firmware, and hardware implementations.
  • From the foregoing, it may be seen that the embodiments disclosed herein provide an approach for predicting a degree of relevance between query rewrites and a search query. By using a relevance model to predict a degree of relevance between the query rewrites and search query before serving advertisements, an ad provider is able to more accurately serve relevant advertisements.
  • While the method and system has been described with reference to certain embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted without departing from the scope. In addition, many modifications may be made to adapt a particular situation or material to the teachings without departing from its scope. Therefore, it is intended that the present method and system not be limited to the particular embodiment disclosed, but that the method and system include all embodiments falling within the scope of the appended claims.

Claims (21)

1. A method for predicting a degree of relevance between search queries, the method comprising:
receiving a search query;
identifying a candidate query rewrite associated with the search query;
extracting a first set of features from advertisements associated with the candidate query rewrite and the search query;
determining a first degree of relevance between the advertisements associated with the candidate query rewrite and the search query based on the first set of features and a second set of features extracted from advertisements and query terms of known relevance; and
determining a second degree of relevance between the candidate query rewrite and the search query based on the first degree of relevance between the advertisements associated with the candidate query rewrite and the search query.
2. The method according to claim 1, wherein the first degree of relevance corresponds to an average relevance between the advertisements associated with the candidate query rewrite and the search query.
3. The method according to claim 1, further comprising serving advertisements associated with the query rewrite that have a second degree of relevance higher than a threshold.
4. The method according to claim 1, further comprising determining a third degree of relevance between advertisements associated with the query rewrite that have a second degree of relevance higher than a first threshold and the search query, and serving those advertisements that have a third degree of relevance higher than a second threshold.
5. The method of claim 1, wherein extracting the first set of features comprises:
determining a degree to which terms associated with the advertisements associated with the candidate query rewrite overlaps with terms in the search query.
6. The method of claim 1, wherein extracting the first set of features comprises:
determining a degree to which terms associated with the advertisements associated with the candidate query rewrite overlaps with terms in the search query, weighted based on a number of times a term appears in both the advertisements associated with the candidate query rewrite and the first search query.
7. The method of claim 1, wherein extracting the first set of features comprises:
determining a degree of relevance between the advertisements associated with the candidate query rewrite and the search query based on the co-occurrence of a first term and a second term, which is different from the first term but is related to the first term, in the advertisements associated with the candidate query rewrite and the first search query.
8. The method of claim 1, wherein extracting the first set of comprises:
determining a quality of the advertisements associated with the candidate query rewrite based on a bid price associated with two or more advertisements of the advertisements associated with the candidate query rewrite.
9. The method of claim 1, wherein extracting the first set of features comprises:
determining a quality of the advertisements associated with the candidate query rewrite based on a coefficient of variation of an ad score associated with two or more advertisements of the advertisements associated with the candidate query rewrite.
10. The method of claim 1, wherein extracting the first set of comprises:
determining a quality of the advertisements associated with the candidate query rewrite based on a degree of topical cohesiveness of two or more advertisements of the advertisements associated with the candidate query rewrite.
11. The method of claim 10, wherein determining a quality of the advertisements associated with the candidate query rewrite based on a degree of topical cohesiveness of two or more advertisements of the advertisements associated with the candidate query rewrite comprises:
building a relevance model over at least one of terms or semantic classes associated with two or more advertisements of the advertisements associated with the candidate query rewrite; and
determining a clarity score for the advertisements associated with the candidate query rewrite based on a difference between the relevance model and a model of an ad inventory of an ad provider.
12. The method of claim 10, wherein determining a quality of the advertisements associated with the candidate query rewrite based on a degree of topical cohesiveness of two or more advertisements of the advertisements associated with the candidate query rewrite comprises:
building a relevance model over at least one of terms or semantic classes associated with two or more advertisements of the advertisements associated with the candidate query rewrite; and
determining an entropy score for the advertisements associated with the candidate query rewrite based on a probability distribution of the terms or semantic classes over which the relevance model was built.
13. A machine-readable storage medium having stored thereon, a computer program comprising at least one code section for predicting a degree of relevance between search queries, the at least one code section being executable by a machine for causing the machine to perform acts of:
receiving a search query;
identifying a set of candidate query rewrites associated with the search query;
extracting a set of features from advertisements associated with the set of candidate query rewrites and the search query;
determining a degree of relevance between the advertisements associated with the set of candidate query rewrites and the search query based on a prediction model and the set of features extracted from the advertisements associated with the set of candidate query rewrites and the search query; and
determining a degree of relevance between the set of candidate query rewrites and the search query based on the determined degree of relevance between the advertisements associated with the set of candidate query rewrites and the search query receiving a search query;
identifying a candidate query rewrite associated with the search query;
extracting a first set of features from advertisements associated with the candidate query rewrite and the search query;
determining a first degree of relevance between the advertisements associated with the candidate query rewrite and the search query based on the first set of features and a second set of features extracted from advertisements and query terms of known relevance; and
determining a second degree of relevance between the candidate query rewrite and the search query based on the first degree of relevance between the advertisements associated with the candidate query rewrite and the search query.
14. The machine-readable storage medium according to claim 13, the first degree of relevance corresponds to an average relevance between the advertisements associated with the candidate query rewrite and the search query.
15. The machine-readable storage medium according to claim 13, wherein the at least one code section comprises code that enables serving advertisements associated with the query rewrite that have a second degree of relevance higher than a threshold.
16. The machine-readable storage medium according to claim 13, wherein the at least one code section comprises code that enables determining a third degree of relevance between advertisements associated with the query rewrite that have a second degree of relevance higher than a first threshold and the search query, and serving those advertisements that have a third degree of relevance higher than a second threshold.
17. A system for predicting a degree of relevance between search queries, the system comprising:
a receiver operative to receive a search query;
identification circuitry operative to identify a candidate query rewrite associated with the search query; and
a relevance module operative to extract a first set of features from advertisements associated with the candidate query rewrite and the search query, determine a first degree of relevance between the advertisements associated with the candidate query rewrite and the search query based on the first set of features and a second set of features extracted from advertisements and query terms of known relevance, and determine a second degree of relevance between the candidate query rewrite and the search query based on the first degree of relevance between the advertisements associated with the candidate query rewrite and the search query.
18. The system according to claim 17, wherein the first degree of relevance corresponds to an average relevance between the advertisements associated with the candidate query rewrite and the search query.
19. The system according to claim 17, wherein the relevance module is operative to serve advertisements associated with the query rewrite that have a second degree of relevance higher than a threshold.
20. The system according to claim 17, wherein the relevance module is operative to determine a third degree of relevance between advertisements associated with the query rewrite that have a second degree of relevance higher than a first threshold and the search query, and serving those advertisements that have a third degree of relevance higher than a second threshold.
21. A system for predicting a degree of relevance between search queries, the system comprising:
means for receiving a search query;
means for identifying a candidate query rewrite associated with the search query; and
means for extracting a first set of features from advertisements associated with the candidate query rewrite and the search query;
means for determining a first degree of relevance between the advertisements associated with the candidate query rewrite and the search query based on the first set of features and a second set of features extracted from advertisements and query terms of known relevance; and
means for determining a second degree of relevance between the candidate query rewrite and the search query based on the first degree of relevance between the advertisements associated with the candidate query rewrite and the search query.
US12/169,218 2008-07-08 2008-07-08 Prediction of a degree of relevance between query rewrites and a search query Abandoned US20100010895A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/169,218 US20100010895A1 (en) 2008-07-08 2008-07-08 Prediction of a degree of relevance between query rewrites and a search query
PCT/US2009/048681 WO2010005801A2 (en) 2008-07-08 2009-06-25 Prediction of a degree of relevance between query rewrites and a search query

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/169,218 US20100010895A1 (en) 2008-07-08 2008-07-08 Prediction of a degree of relevance between query rewrites and a search query

Publications (1)

Publication Number Publication Date
US20100010895A1 true US20100010895A1 (en) 2010-01-14

Family

ID=41505999

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/169,218 Abandoned US20100010895A1 (en) 2008-07-08 2008-07-08 Prediction of a degree of relevance between query rewrites and a search query

Country Status (2)

Country Link
US (1) US20100010895A1 (en)
WO (1) WO2010005801A2 (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080140699A1 (en) * 2005-11-09 2008-06-12 Rosie Jones System and method for generating substitutable queries
US20080319962A1 (en) * 2007-06-22 2008-12-25 Google Inc. Machine Translation for Query Expansion
US20120307662A1 (en) * 2009-12-23 2012-12-06 7Signal Oy Method for monitoring and intelligent control of the parameters in radio networks
US20130103493A1 (en) * 2011-10-25 2013-04-25 Microsoft Corporation Search Query and Document-Related Data Translation
US8478699B1 (en) * 2010-04-30 2013-07-02 Google Inc. Multiple correlation measures for measuring query similarity
US20130332225A1 (en) * 2011-03-29 2013-12-12 Nec Corporation Risk-profile generation device
US8732151B2 (en) 2011-04-01 2014-05-20 Microsoft Corporation Enhanced query rewriting through statistical machine translation
US8756241B1 (en) * 2012-08-06 2014-06-17 Google Inc. Determining rewrite similarity scores
US8799312B2 (en) 2010-12-23 2014-08-05 Microsoft Corporation Efficient label acquisition for query rewriting
US8874581B2 (en) 2010-07-29 2014-10-28 Microsoft Corporation Employing topic models for semantic class mining
US8903817B1 (en) * 2011-08-23 2014-12-02 Amazon Technologies, Inc. Determining search relevance from user feedback
US8972411B1 (en) * 2009-01-27 2015-03-03 Google Inc. Selection of sponsored content using multiple sets of query terms
US9177341B2 (en) 2011-08-23 2015-11-03 Amazon Technologies, Inc. Determining search relevance from user feedback
US9507861B2 (en) 2011-04-01 2016-11-29 Microsoft Technolgy Licensing, LLC Enhanced query rewriting through click log analysis
US9659248B1 (en) 2016-01-19 2017-05-23 International Business Machines Corporation Machine learning and training a computer-implemented neural network to retrieve semantically equivalent questions using hybrid in-memory representations
US20180107636A1 (en) * 2016-04-29 2018-04-19 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for establishing sentence editing model, sentence editing method and apparatus
CN112132601A (en) * 2019-06-25 2020-12-25 百度在线网络技术(北京)有限公司 Advertisement title rewriting method, device and storage medium
US11113714B2 (en) * 2015-12-30 2021-09-07 Verizon Media Inc. Filtering machine for sponsored content
US11113291B2 (en) 2018-09-17 2021-09-07 Yandex Europe Ag Method of and system for enriching search queries for ranking search results
US11194878B2 (en) 2018-12-13 2021-12-07 Yandex Europe Ag Method of and system for generating feature for ranking document
US11562292B2 (en) 2018-12-29 2023-01-24 Yandex Europe Ag Method of and system for generating training set for machine learning algorithm (MLA)
US11681713B2 (en) 2018-06-21 2023-06-20 Yandex Europe Ag Method of and system for ranking search results using machine learning algorithm

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020107853A1 (en) * 2000-07-26 2002-08-08 Recommind Inc. System and method for personalized search, information filtering, and for generating recommendations utilizing statistical latent class models
US20050071325A1 (en) * 2003-09-30 2005-03-31 Jeremy Bem Increasing a number of relevant advertisements using a relaxed match
US20050080772A1 (en) * 2003-10-09 2005-04-14 Jeremy Bem Using match confidence to adjust a performance threshold
US20060155751A1 (en) * 2004-06-23 2006-07-13 Frank Geshwind System and method for document analysis, processing and information extraction
US20060161534A1 (en) * 2005-01-18 2006-07-20 Yahoo! Inc. Matching and ranking of sponsored search listings incorporating web search technology and web content
US20060259455A1 (en) * 2002-09-24 2006-11-16 Darrell Anderson Serving advertisements based on content
US20070185847A1 (en) * 2006-01-31 2007-08-09 Intellext, Inc. Methods and apparatus for filtering search results
US20070208714A1 (en) * 2006-03-01 2007-09-06 Oracle International Corporation Method for Suggesting Web Links and Alternate Terms for Matching Search Queries
US20070299826A1 (en) * 2006-06-27 2007-12-27 International Business Machines Corporation Method and apparatus for establishing relationship between documents
US20090198644A1 (en) * 2008-02-05 2009-08-06 Karolina Buchner Learning query rewrite policies
US20090216710A1 (en) * 2008-02-27 2009-08-27 Yahoo! Inc. Optimizing query rewrites for keyword-based advertising
US20090248627A1 (en) * 2008-03-27 2009-10-01 Yahoo! Inc. System and method for query substitution for sponsored search

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020107853A1 (en) * 2000-07-26 2002-08-08 Recommind Inc. System and method for personalized search, information filtering, and for generating recommendations utilizing statistical latent class models
US20060259455A1 (en) * 2002-09-24 2006-11-16 Darrell Anderson Serving advertisements based on content
US20050071325A1 (en) * 2003-09-30 2005-03-31 Jeremy Bem Increasing a number of relevant advertisements using a relaxed match
US20050080772A1 (en) * 2003-10-09 2005-04-14 Jeremy Bem Using match confidence to adjust a performance threshold
US20060155751A1 (en) * 2004-06-23 2006-07-13 Frank Geshwind System and method for document analysis, processing and information extraction
US20060161534A1 (en) * 2005-01-18 2006-07-20 Yahoo! Inc. Matching and ranking of sponsored search listings incorporating web search technology and web content
US20070185847A1 (en) * 2006-01-31 2007-08-09 Intellext, Inc. Methods and apparatus for filtering search results
US20070208714A1 (en) * 2006-03-01 2007-09-06 Oracle International Corporation Method for Suggesting Web Links and Alternate Terms for Matching Search Queries
US20070299826A1 (en) * 2006-06-27 2007-12-27 International Business Machines Corporation Method and apparatus for establishing relationship between documents
US20090198644A1 (en) * 2008-02-05 2009-08-06 Karolina Buchner Learning query rewrite policies
US20090216710A1 (en) * 2008-02-27 2009-08-27 Yahoo! Inc. Optimizing query rewrites for keyword-based advertising
US20090248627A1 (en) * 2008-03-27 2009-10-01 Yahoo! Inc. System and method for query substitution for sponsored search

Cited By (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7962479B2 (en) * 2005-11-09 2011-06-14 Yahoo! Inc. System and method for generating substitutable queries
US20080140699A1 (en) * 2005-11-09 2008-06-12 Rosie Jones System and method for generating substitutable queries
US20080319962A1 (en) * 2007-06-22 2008-12-25 Google Inc. Machine Translation for Query Expansion
US9569527B2 (en) 2007-06-22 2017-02-14 Google Inc. Machine translation for query expansion
US9002869B2 (en) 2007-06-22 2015-04-07 Google Inc. Machine translation for query expansion
US8972411B1 (en) * 2009-01-27 2015-03-03 Google Inc. Selection of sponsored content using multiple sets of query terms
US20120307662A1 (en) * 2009-12-23 2012-12-06 7Signal Oy Method for monitoring and intelligent control of the parameters in radio networks
US8478699B1 (en) * 2010-04-30 2013-07-02 Google Inc. Multiple correlation measures for measuring query similarity
US8825571B1 (en) 2010-04-30 2014-09-02 Google Inc. Multiple correlation measures for measuring query similarity
US8874581B2 (en) 2010-07-29 2014-10-28 Microsoft Corporation Employing topic models for semantic class mining
US8799312B2 (en) 2010-12-23 2014-08-05 Microsoft Corporation Efficient label acquisition for query rewriting
US20130332225A1 (en) * 2011-03-29 2013-12-12 Nec Corporation Risk-profile generation device
US8732151B2 (en) 2011-04-01 2014-05-20 Microsoft Corporation Enhanced query rewriting through statistical machine translation
US9507861B2 (en) 2011-04-01 2016-11-29 Microsoft Technolgy Licensing, LLC Enhanced query rewriting through click log analysis
US9177341B2 (en) 2011-08-23 2015-11-03 Amazon Technologies, Inc. Determining search relevance from user feedback
US8903817B1 (en) * 2011-08-23 2014-12-02 Amazon Technologies, Inc. Determining search relevance from user feedback
US9501759B2 (en) * 2011-10-25 2016-11-22 Microsoft Technology Licensing, Llc Search query and document-related data translation
US20130103493A1 (en) * 2011-10-25 2013-04-25 Microsoft Corporation Search Query and Document-Related Data Translation
US8756241B1 (en) * 2012-08-06 2014-06-17 Google Inc. Determining rewrite similarity scores
US11113714B2 (en) * 2015-12-30 2021-09-07 Verizon Media Inc. Filtering machine for sponsored content
US9659248B1 (en) 2016-01-19 2017-05-23 International Business Machines Corporation Machine learning and training a computer-implemented neural network to retrieve semantically equivalent questions using hybrid in-memory representations
US10191892B2 (en) * 2016-04-29 2019-01-29 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for establishing sentence editing model, sentence editing method and apparatus
US20180107636A1 (en) * 2016-04-29 2018-04-19 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for establishing sentence editing model, sentence editing method and apparatus
US11681713B2 (en) 2018-06-21 2023-06-20 Yandex Europe Ag Method of and system for ranking search results using machine learning algorithm
US11113291B2 (en) 2018-09-17 2021-09-07 Yandex Europe Ag Method of and system for enriching search queries for ranking search results
US11194878B2 (en) 2018-12-13 2021-12-07 Yandex Europe Ag Method of and system for generating feature for ranking document
US11562292B2 (en) 2018-12-29 2023-01-24 Yandex Europe Ag Method of and system for generating training set for machine learning algorithm (MLA)
CN112132601A (en) * 2019-06-25 2020-12-25 百度在线网络技术(北京)有限公司 Advertisement title rewriting method, device and storage medium

Also Published As

Publication number Publication date
WO2010005801A2 (en) 2010-01-14
WO2010005801A3 (en) 2010-04-22

Similar Documents

Publication Publication Date Title
US20100010895A1 (en) Prediction of a degree of relevance between query rewrites and a search query
US8572011B1 (en) Outcome estimation models trained using regression and ranking techniques
US8412648B2 (en) Systems and methods of making content-based demographics predictions for website cross-reference to related applications
US10134053B2 (en) User engagement-based contextually-dependent automated pricing for non-guaranteed delivery
US8073850B1 (en) Selecting key phrases for serving contextually relevant content
US8103650B1 (en) Generating targeted paid search campaigns
US8311957B2 (en) Method and system for developing a classification tool
US8600809B1 (en) Predictive model performance
US8725562B2 (en) Keyword advertisement using ranking of advertisers
US20120059713A1 (en) Matching Advertisers and Users Based on Their Respective Intents
US11762908B1 (en) Node graph pruning and fresh content
US20080256034A1 (en) System and method for understanding relationships between keywords and advertisements
US20110270672A1 (en) Ad Relevance In Sponsored Search
US20110264507A1 (en) Facilitating keyword extraction for advertisement selection
US20160132935A1 (en) Systems, methods, and apparatus for flexible extension of an audience segment
US20110040604A1 (en) Systems and Methods for Providing Targeted Content
US20210192460A1 (en) Using content-based embedding activity features for content item recommendations
US8688514B1 (en) Ad selection using image data
US20110131093A1 (en) System and method for optimizing selection of online advertisements
US20190080348A1 (en) Exploration for search advertising
US9064007B1 (en) Co-click based similarity score of queries and keywords
US8234265B1 (en) Content selection data expansion
Wang et al. Viewability prediction for online display ads
US20150310487A1 (en) Systems and methods for commercial query suggestion
US20130166371A1 (en) Predicting advertiser keyword performance indicator values

Legal Events

Date Code Title Description
AS Assignment

Owner name: YAHOO| INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GABRILOVICH, EVGENIY;METZLER, DONALD;JOSIFOVSKI, VANJA;AND OTHERS;REEL/FRAME:021272/0979;SIGNING DATES FROM 20080626 TO 20080703

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: YAHOO HOLDINGS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAHOO| INC.;REEL/FRAME:042963/0211

Effective date: 20170613

AS Assignment

Owner name: OATH INC., NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAHOO HOLDINGS, INC.;REEL/FRAME:045240/0310

Effective date: 20171231