Home
Building Taxonomy of Web Search Intents for Name Entity Queries
Contents
1. Class of entity Num Wikipedia categories or Web source Entity car models 859 2000s_automobiles U S clothing stores 103 clothing retailers_of the _united_states film actors 19432 film_actors musicians 21091 female singers _male_singers music groups restaurants 694 restaurants universities colleges 7191 universities and colleges _ U S cities 246 www mongabay com igapo US htm U S presidents 57 presidents _of the _united_states U S retail companies 180 retail companies_of the _united_states U S TV networks 276 american_television_networks of entities from one or multiple Wikipedia categories or their des cendent categories represent wildcard match and one is from a web page For the nine classes from Wikipedia we ignore all entities with a Wikipedia disambiguation page We use the query click logs of Bing search formerly as Live search from 2008 01 01 to 2008 12 31 For each query that ap pears at least 6 times in the logs we get the numbers of clicks and skips of each URL for this query For each pair of queries with common clicked URL s we compute the degree of the intent of one query being included in the other using MapReduce on a PC cluster Based on this we compute the relationships between the intents of each pair of intent phrases as in Section 4 2 We compare the performances of three approaches 1 DMST Method based on directed maximum spanning tree 2 HAC
2. method based on hierarchical agglomerative clustering and 3 PAM method based on pachinko allocation model baseline approach For each class of entities we use each approach to build an intent tree The top 400 intent phrases that co appear with most entities in search queries are considered for each class of entities We ignore remaining intent phrases as they are usually obscure or do not carry generic intents 6 2 Measurements by Mechanical Turk Based on an intent tree we say two intent phrases are either equivalent if they belong to the same node one belonging to another if the node of one phrase is a descendent of the node of the other or unrelated if none of the above In order to measure the accuracy of each approach we test how accurate it is in judg ing the relationship between two queries containing the same entity and different intent phrases We use Amazon Mechanical Turk 1 to judge the relationships between queries For each pair of queries q and q2 a Mechanical Turk worker is asked to figure out their intents we provide buttons for showing Google s results of each query and ask them to search on Google if they are not sure and select one of the following four options about the in tents of q and q2 1 Unrelated 2 q s intents belongs to those of q2 3 qy s intents contain those of qo and 4 equivalent We provide four to six examples for each type of cases such as tom hanks movies an
3. H Cui J R Wen J Y Nie and W Y Ma Probabilistic query expansion using query logs WWW 02 W H E Day and H Edelsbrunner Efficient algorihtms for agglomerative hierarchical clustering methods Journal of Classification 1 1 7 24 1984 9 J Dean and S Chemawat MapReduce Simplified data processing on large clusters OSDI 04 10 J Edmonds Optimum branching J Research of the National Bureau of Standards 71 B pp 233 240 1967 11 A Fuxman P Tsaparas K Achan and R Agrawal Using the wisdom of the crowds for keyword generation WWW 08 12 H N Gabow Z Galil T Spencer and R E Tarjan Efficient algorithms for finding minimum spanning trees in undirected and directed graphs Combinatorica 6 pp 109 122 1986 13 F Guo C Liu and Y M Wang Efficient multiple click models in web search WSDM 09 14 J Guo G Xu X Cheng and H Li Named entity recogni tion in query SIGIR 09 15 T Joachims L Granka B Pan H Hembrooke F Radlinski and G Gay Evaluating the accuracy of implicit feedback from clicks and query reformulations in web search In ACM Trans Information Systems 25 2 2007 16 J Hu G Wang F Lochovsky J T Sun and Z Chen Un derstanding User s Query Intent with Wikipedia VWW 09 17 U Lee Z Liu and J Cho Automatic identification of user goals in web search WWW 05 18 W Li and A McCallum Pachinko Allocation DAG structured mixtur
4. 824 700 854 796 824 848 886 867 belongs 0 0 0 500 737 596 625 263 370 equivalent 684 542 605 857 324 470 762 865 810 Table 9 summarizes the accuracy of different approaches Con sidering this is a four class classification problem unrelated belongs in either direction or equivalent our approaches achieve reasonably high accuracy It is shown that aggregating query relationships involving different entities is helpful as intent phrase relationships are more accurate than query relationships HAC achieves significantly higher accuracy because clustering helps to better capture the relationships between intent phrases For example even if the inferred relationship between bio and wikipedia is very weak for a class of entities such as musicians HAC can probably group bio and biography into a node and identify wikipedia as a child node of that node and thus infers the intents of wikipedia belongs to those of bio Table 9 Summary of accuracy of different approaches PAM baseline DMST HAC aeeassoried body parts Accuracy 532 560 675 auto parts 2 door prec recl F prec rec l F prec recl Fi miles per galon Ja unrelated 497 924 646 678 817 741 727 867 791 new car lt sedan belongs 220 050 082 308 405 350 389 198 262 root used f
5. Examples include musicians soccer athletes presidents of U S movies car manufacturers cell phone models superstores TV networks cities etc Almost every entity can be categorized into some classes It is usually easy to get a class of entities For most classes we can think about there is corresponding categories or lists on Wikipedia such as en wikipedia org wiki Category Lists_of musicians for musicians Such lists can also be gathered by approaches such as 19 Because a large portion of web search queries are entity queries and many others contain names of entities improving user expe riences for entity queries is very important for search engines The goal of our study is to find generic search intents associated with a class of entities identify relationships between different generic intents and organize them into a hierarchical structure Similar to studies on query suggestion 5 7 22 we represent a specific search intent by a query such as britney spears biography We represent generic intents by intent phrases For a query con taining an entity name as a substring we define the query exclud ing the entity name as an intent phrase For example for query britney spears biography with britney spears being an entity biography is an intent phrase We consider an intent phrase carrying the same generic intent no matter whether it appears before or after the entities in queries Different intent phrases
6. For each query pair q q2 with judgments by Mechanical Turk we get the degree of intents of one query being included in those of the other i e d q E q2 and d qz q The distribu tions are shown in Figure 5 We can see the distribution of rela tionships between unrelated query pairs and equivalent ones are mostly separated while query pairs with one belonging to the other are less separated from other types Unrelated 0 2 Belongs a Ce rere Belongs other direction o 015 y _ _ iene ao a a o1 a 5 0 05 oO 0 0 1 0 2 03 04 05 06 07 08 0 9 d qi S q2 or d qz S q1 Figure 5 Distribution of relationships between queries for each type of query pairs 1 WWW 2010 Full Paper Then we use our manually labeled examples and Mechanical Turk data to measure the accuracy of query relationships We set a threshold t For each query pair q q2 we predict they are 1 unrelated if d q E q2 lt T and d qz E q1 lt T 2 q belongs to qo ifd q S q2 gt T and d q E q2 gt 2 d q2 E q 3 2 belongs to q likewise and 4 equivalent if none of above We found z 0 09 leads to highest accuracy w r t manually labeled examples The accuracy precision recall and F measure are shown in Table 4 Table 4 Precision Recall of Inferred Query Relationships By manually labeled data By Mechanical Turk data Accura
7. For every other node n suppose the original weight of edge from n or n to ngis dix or diz Let fw be the total frequency of all queries consisted of intent phrase w and an entity in the specific class and f n Xwew n f w where W n is the set of phrases of n The new weight from merged node n to n is an average of dx and dix weighted by f and fn as follows _ Fp dint f nj djx I d jr f np f nj 6 It can be easily proved that if d or dix is the weighted average degree of any intent phrase in n or n belonging to that of n then d is that for the newly merged node n After updating the edges the algorithm extracts the next edge with highest weight and performs the merging Lines 17 19 create the final intent tree and output it This algorithm is different from traditional average link algo rithm in two aspects First in the input of our algorithm there can be two directed edges between each two objects indicating bi directional relationships In contrast average link algorithm only Algorithm 1 Average link for intent trees Input Intent phrases wy wx for a certain class of entities d w E wj for 1 lt i 7 lt K Output An intent tree whose nodes contain w wx Procedure 1 build graph G with a node n containing each phrase w 2 for each d w Sc wj gt 0 3 add edge n n with weight d wi e wj and type 1 4 ifr a w c w d w S w gt max dw w d w S
8. With the evaluation by human judges via Mechanical Turk it is shown that our approaches can build trees of phrases that capture the relationships between important search intents Categories and Subject Descriptors H 3 3 Information Storage and Retrieval Information Search and Retrieval search process General Terms Algorithms Measurement Experimentation Keywords Web search intent query clustering 1 INTRODUCTION Until a few years ago most major search engines return only ten result snippets to user and let the user find useful results by reading the snippets Although this has been very successful it costs much user effort in reading snippets and it is not always possible for the search engine to accurately capture the user s intent when it is not clearly indicated in the query This problem is more obvious for name entity queries since different users may search for different aspects of a name entity using the same query and it is very difficult for the search engine to infer the exact search intent Name entity queries are a most popular type of que ries According to an internal study of Microsoft at least 20 30 of queries submitted to Bing search are simply name entities and Copyright is held by the International World Wide Web Conference Committee IW3C2 Distribution of these papers is limited to classroom use and personal use by others WWW 2010 April 26 30 2010 Raleigh North Carolina USA
9. angles The first approach is based on directed maximum spanning tree 10 The second one is based on hierarchical ag glomerative clustering 8 The third one is a baseline approach based on Pachinko Allocation models 18 5 1 Method Based on Directed MST Given the relationships between different intent phrases for a class of entities a good intent tree should have the following property For all pairs of intent phrases w w2 that are in the same node or w in a child node of wz Yiw w dW S w2 is high This problem is similar to finding a maximum spanning tree WWW 2010 Full Paper MST in the directed graph containing a node for each intent phrase and an edge between each pair of related intent phrases Directed MST has been studied in 6 10 These papers study how to find Minimum Spanning Tree which is equivalent to find ing Maximum Spanning Tree In order to apply such algorithms we first define the goodness of an intent tree DEFINITION 3 Score of intent tree Given the intent phrases W Wx for a class of entities and an intent tree T with nodes N containing w wx we define the score of T as 4 s T Vi n echild root L F Za hienoi S Wi i d w w child n is the child nodes of n W n is the intent phrases of node n p is the base weight of each node For a node n if d w Cc wj lt P for any w ofn and w of some other node then we do not consider n represent a sub concep
10. is shown in Figure 2 which is generated by our approach Our work can help search engines in three ways First it can help to organize the search tasks by selecting tasks for representa tive intents and avoiding redundancy With the taxonomy of in tents we can also easily create a tree of search tasks so that a user can navigate in the tree to find the right task Second our work can help to better organize the direct answers so that their layout is consistent with the taxonomy of user intents The direct answers can also be organized into a tree according to the tax onomy whose nodes can be dynamically shown to users when necessary Third our work can help web developers and users better understand queries containing name entities and provide an overview of different search intents for a class of entities Outline and contributions Baeza Yates et al propose the concept of query relationships including equivalence IS A and overlapping relationships 2 However only equivalence relationship is studied in 2 which is also studied in 3 5 24 In this paper we study both equivalence and IS A relationships between search intents and how to build taxonomy of intents based on such relationships The first challenge is how to identify generic search intents for a class of entities For each class of entities we can find the ge neric search intents from user queries For example if many users search for musician lyrics
11. of parts In the intent tree for U S presidents the node containing biography has child nodes like presidency wikipedia education and history Such intent trees are very useful in providing the search tasks for search engines like Bing and help ing to organize the result pages for search engines like Kosmix 6 5 Run time and Scalability The first two steps of our approach are 1 Computing the rela tionships between queries and 2 computing the intent phrases for each class of entities and relationships between them We use children wife presidency administration wikipedia education president biography bio history photos images pictures pictures president interesting fac fun facts for kids trivia facts facts about school school district high school middle school elementary school Figure 10 Intent tree of U S Presidents by HAC WWW 2010 Full Paper Table 10 Run time seconds of each approach URL Class of entities PAM car_model 0 109 clothing store 0 265 film_actors 0 250 musician 0 374 restaurants 0 125 147531 0 062 0 078 1 342 0 484 0 062 universities _and_colleges us cities us_president us_retail_ companies us_tv_networks Average a PC cluster supporting MapReduce 9 to perform these two tasks The
12. ries involving a specific entity Many existing approaches on rank ing query suggestions can be applied here 4 INTENT PHRASES AND THEIR RELA TIONSHIPS 4 1 Intent Phrases We hope to find all intent phrases that represent generic search intents involving entities of a certain class For each generic intent e g lyrics of a musician there are usually many users express ing the intent explicitly in their queries and thus we can capture most generic intents using all intent phrases On the other hand to make sure an intent phrase represents a generic intent for different entities the intent phrase should appear with at least 0 entities in queries 0 5 in our study We exclude a small set of stop phras es that do not indicate clear intents such as web online 2009 and inappropriate phrases such as nude and sex There are 185 stop phrases including misspelled terms Table 1 shows the top intent phrases for some classes of entities sorted by the number of entities they co appear within queries We can see they represent very important intents for each class of entities Some of them are redundant e g jobs and employment and we will study how to infer relationships between intent phrases 4 2 Relationships between Intent Phrases Before defining the relationships between intent phrases we need to first define the relationships between the intents of differ ent queries We say the int
13. t a which controls when to merge intent phras es into one node no such merging is done when a 1 Figure 7 b shows accuracy w r t 8 which controls whether to put a node as a child of root or that of another node A smaller 6 means more nodes are put as child nodes of root DMST achieves highest ac curacy when a 0 95 and 0 08 which are used in all experi ments including this one unless specified otherwise Figure 8 a shows the average accuracy of HAC w r t y which controls when to merge intent phrases into one node When y 0 5 no two intent phrases are merged into one node and the result is a tree with a phrase on each node When y 1 no node is a child of another node and the result is a set of clusters generated by traditional hierarchical agglomerative clustering None of these two cases leads to best accuracy Figure 8 b shows the accuracy w r t min sim which is the similarity threshold for merging nodes Highest accuracy is achieved when y 0 9 and min_sim 0 03 which are used in all experiments including this one Table 6 shows the accuracy of PAM DMST and HAC on ten classes of entities HAC achieves highest accuracy on most classes and DMST is slightly more accurate than PAM whose Table 6 Accuracy on ten classes of entities Class of entities PAM baseline DMST HAC car models 0 7075 U S clothing stores 0 4976 film actors 0 7055 musicians 0 7652 restaurants 0 4781 universities
14. the problem of building taxonomies of search intents for entity queries We use intent phrases to represent the important generic search intents and propose an approach based on user logs for measuring the belonging relationships be tween intents of queries We propose three approaches for build ing hierarchies of intent phrases based on such relationships It is shown by experiments that these approaches can generate intent trees that capture the hierarchical structure of user intents An important direction of future work is to study how to rank the child nodes of each node in the intent tree There have been 1010 April 26 30 Raleigh NC USA many mature studies on ranking query suggestions which can be adapted for ranking intent phrases 8 REFERENCES 1 Amazon Mechanical Turk https www mturk com mturk 2 R Baeza Yates A Tiberi Extracting semantic relations from query logs KDD 07 3 H Cao D Jiang J Pei Q He Z Liao E Chen and H Li Context aware query suggestion by mining click through and session data KDD 08 S L Chuang L F Chien Automatic query taxonomy gen eration for information retrieval applications Online Infor mation Review 27 4 2003 S Chien and N Immorlica Semantic similarity between search engine queries using temporal correlation WWW 05 Y J Chu and T H Liu On the shortest arborescence of a directed graph Science Sinica vol 14 pp 1396 1400 1965
15. users there exist two limitations First the direct answers only work well for certain classes of entities For example for query university of washington Yahoo provides no direct answer while Kosmix shows a simple list of direct answers without most popular query intents such as sports admissions etc Second these approaches put the major generic search intents k DING britney spears BRITNEY SPEARS News News about britney spears Britney Spears Delivers Her Pretaped Late Ever since her disastrous zombified performa Music Awards the team of handlers that man Nymag com 18 minutes ago Musician Bio Britney Spears Songs Britney Spears Tickets Britney Spears Tour Celeb Pix Bikini clad Britney Spears deliver Britney Spears Inside Bay Area 1 hour ago Albums Britney Spears shows her new figure on Lett Blueridgenow com 1 hour ago Britney Spears Biography Images Fact Database Britney Spears official web site ai Access Britney Spears photos galleries anc Britney on her oficial blog a Search tasks of Bing left pane Videos b Top part of _Kosmix result page Britney Spears Official Site www britneyspears com Albums Lyrics Photos Videos Play Full Length Tracks IfU Seek Amy Circus Piece Of Me amp Gimme More fe oh Watch Music Videos z KEN IEN 3 c Top part of Yahoo result page Figure 1 The result pages containing search tasks or di
16. 7 z 0 6 8 0 6 wen Sgt aN 5 0 4 0 4 0 3 0 3 0 2 02 05 06 07 08 09 1 0 0 04 0 08 0 12 0 16 a merging condition a accuracy w r t a Figure 7 Accuracy of DMST B min similarity b accuracy w r t p 1008 April 26 30 Raleigh NC USA 0 8 0 8 0 7 0 7 e a 0 6 eo 408 z 0 5 50 5 2 2 0 4 0 4 0 3 0 3 0 2 4 1 1 1 1 0 2 1 T 1 05 06 07 08 09 1 0 01 0 03 0 05 0 07 y merging condition min_sim a accuracy w r t y b accuracy w r t min_sim Figure 8 Accuracy of HAC tage of predictions being correct Figure 6 a shows the average accuracy of PAM on all classes of entities w r t 7 which controls when to put an intent phrase on a super topic top level node or a sub topic 2 level node Fig ure 6 b shows accuracy w r t number of super topics We always set the number of sub topics to be four times that of super topics PAM achieves highest accuracy with 7 1 8 and 10 super topics which are used in all experiments including this one unless spe cified otherwise PAM fails to run on three classes of entities film actors musicians and U S cities because the input size is too large PAM is also very slow and for Figure 6 b we only test on three classes on which PAM finishes in 24 hours U S clothing stores U S presidents and U S TV networks Figure 7 a shows the average accuracy on ten classes of enti ties by DMST w r
17. ACM 978 1 60558 799 8 10 04 1001 Sarthak Shah Microsoft Corp One Microsoft Way Redmond WA 98052 sarthaks microsoft com it is reported 71 of queries contain name entities 14 For sim plicity we will use the word entity to refer to name entity Recently there are some new methods used by different search engines to help users find the right information faster and more effectively for entity queries The first method is to provide rele vant search tasks Bing search For example when the user searches for an entity Bing usually shows several popular tasks for this entity As shown in Figure l a for query Britney Spears Bing shows five tasks Songs Tickets Tour Albums and Biography together with three result snippets for each task The tasks are all generic ones for a certain class of entities e g musicians which are found through mining the search logs The second method is to show direct answers for what the user might be searching for which saves the efforts of inspecting result snippets Kosmix and WolframAlpha show only such direct an swers while Yahoo and Google mix such information with regular results Given the query Britney Spears Kosmix shows her biography similar artists albums genres etc Figure 1 b and Yahoo shows her image videos official site songs and links to albums lyrics photos and videos Figure 1 c Although these methods bring more convenience to
18. G of topics with root at the top level super topics at the second level sub topics at the third level and intent phrases at the bottom level Each intent phrase has a certain probability to belong to each sub topic and so does each sub topic to each super topic To build an intent tree we put each intent phrase as a child of the sub topic that it belongs to with highest probability and put each sub topic as a child of a super topic in the same way A tree can be built in this way This tree is not an intent tree yet because the super topics in the output of PAM are not associated with any intent phrases We assign an intent phrase to a super topic if its probability of belong ing to this super topic is at least 7 times its total probability of belonging to any sub topic where 7 gt 1 In this way an intent tree can be built using PAM which has three levels of nodes The root the super topics and the sub topics 6 EXPERIMENTS We now present the experimental evaluation of our approaches All experiments are run on a Windows server with dual 2 66GHz Intel quad core CPU and 32GB main memory or on the PC clus ter of Microsoft All experiments on the Windows server are run with a single core 6 1 Data Collection We test the approaches on ten datasets each containing a cer tain class of entities as shown in Table 2 Nine of them are the set WWW 2010 Full Paper Table 2 Data sources for ten classes of entities
19. L 1 Clicked 2 skipped i e a URL ranked below it is clicked and 3 neither clicked nor skipped According to the studies of Joachims et al 15 the assumption that a clicked URL is more relevant than a skipped URL can provide a large number of preferences between pairs of URLs with reasonable accuracy In order to determine the relationships between intents of dif ferent queries through their clicked URLs we need to be able to estimate the relevance of a URL for a query based on the search logs We adopt a model similar to the one proposed in 13 Sup pose we are given the query click logs of a search engine in a period of time We say a URL is viewed for a query if it is either clicked or skipped Consider a query q in the logs and a URL u that is viewed for q Let click q u be the number of times a user clicks u for q and skip q u be the number of times a user skips u for q We define the relevance of u for q as click q u 1 click q u skip q u E rel q u rel q u is the estimated probability for u being clicked when it is viewed for q e is a smoothing factor which reflects the fact that we are not confident about the estimated relevance when we do not see many clicks skips for a URL e is set to 1 0 in our study As discussed in 13 23 the measure based on probability of being clicked for viewed URLs is quite resistant to positional bias For two queries q and q2 we want to estimate the degree
20. WWW 2010 Full Paper April 26 30 Raleigh NC USA Building Taxonomy of Web Search Intents for Name Entity Queries Xiaoxin Yin Microsoft Research One Microsoft Way Redmond WA 98052 xyin microsoft com ABSTRACT A significant portion of web search queries are name entity que ries The major search engines have been exploring various ways to provide better user experiences for name entity queries such as showing search tasks Bing search and showing direct answers Yahoo Kosmix In order to provide the search tasks or direct answers that can satisfy most popular user intents we need to capture these intents together with relationships between them In this paper we propose an approach for building a hierarchical taxonomy of the generic search intents for a class of name entities e g musicians or cities The proposed approach can find phrases representing generic intents from user queries and organize these phrases into a tree so that phrases indicating equivalent or similar meanings are on the same node and the parent child relationships of tree nodes represent the relationships between search intents and their sub intents Three different methods are proposed for tree building which are based on directed maximum spanning tree hierarchical agglomerative clustering and pachinko alloca tion model Our approaches are purely based on search logs and do not utilize any existing taxonomies such as Wikipedia
21. colleges 0 6613 US cities 0 8101 U S presidents 0 7613 U S retail companies 0 6429 U S TV networks 0 7171 Average 0 5323 0 5604 0 6747 WWW 2010 Full Paper Table 7 Precision recall and F of each type of query pairs w r t Mechanical Turk judgments April 26 30 Raleigh NC USA erformance parts aftermarket parts parts accuracy is computed on seven classes Considering there are four possible types of relationship for each query pair HAC achieves reasonably high prediction accuracy of 0 6747 We also study the precision recall and F for each type of query pairs judged by Mechanical Turk As shown in Table 7 HAC achieves best precision and recall for unrelated and equivalent query pairs while DMST is the best for belongs relationship We also use the 100 manually labeled query pairs to measure the accuracy precision and recall or each approach which is shown in Table 8 The accuracy is higher than when measured with Mechanical Turk data which contains noises The relative performances of different approaches remain the same Table 8 Precision recall and F of each type of query pairs w r t manually labeled data PAM baseline DMST HAC Accuracy 586 610 760 prec rec l F prec recl F prec recl Fy unrelated 609
22. cy 0 540 0 543 prec rec Fi prec rec l Fi unrelated 0 763 0 659 0 707 0 698 0 789 0 741 belongs 0 125 0 211 0 157 0 195 0 180 0 187 equivalent 0 700 0 568 0 627 0 623 0 564 0 592 We also measure the accuracy of relationships between intent phrases as defined in Def 2 The same method is used to predict the relationship between two queries as mentioned above except t 0 26 which leads to optimal results as shown in Table 5 Table 5 Precision Recall of Intent Phrase Relationships By manually labeled data By Mechanical Turk data Accuracy 0 630 0 578 prec rec l F prec rec l F unrelated 0 612 0 932 0 739 0 544 0 939 0 689 belongs 0 333 0 105 0 160 0 810 0 111 0 195 equivalent 0 769 0 541 0 635 0 652 0 598 0 624 6 4 Accuracy of Intent Trees Here we measure the accuracy of intent trees generated by PAM DMST and HAC For each class of entities we use the intent tree generated by each approach to predict the relationship between each pair of queries The accuracy is defined as percen 0 8 0 8 0 7 0 7 2 0 6 a 0 6 fos et te E ji a E 8 0 4 0 4 0 3 0 3 0 2 0 2 4 1 12 14 16 18 n a accuracy w r t y Figure 6 Accuracy of PAM 2 10 super topic 12 14 b accuracy w r t super topic 0 8 0 8 0 7 0
23. cy of q and let e w be the query consisted of entity e and intent phrase w for simplicity we do not distinguish the queries in which w appears before e or after e For a class of entities E and two intent phrases w and w the degree that the search intents of w are included in those of w is defined as Yidle w cetw f e w ecE f e w PO aero E slew ecE f e w p0 3 d w S w3 is the weighted average degree of the intents of a query containing w being included in those of a query containing w2 if these two queries contain the same entity d w w3 E 0 1 for any w and wz If both d w S wz and d w2 w1 are high then the intents of w and w should be very similar and w and w are likely to be equivalent If d w E w3 is high but d w S w is low then w should correspond to a sub concept of w2 We will use such relationships between intent phrases to or ganize the intent phrases into a tree 5 ORGANIZING INTENT PHRASES In this section we describe our approaches for organizing intent phrases into intent trees Each node of the intent tree contains one or more intent phrases with same or similar search intents The child nodes of a node n should contain intent phrases that represent sub concepts of the intent phrase s of n There is a root node that contains no intent phrase An example intent tree has been shown in Figure 2 We will present three approaches that solve this problem from different
24. d how to contact tom hanks unrelated day ton ohio furnitures and furniture in ohio belongs go transit and go transit schedule contains and track income tax re turn and track tax return equivalent For each class of entities we randomly select 250 query pairs so that each pair of queries contain same entity and different intent phrases The chance of a query being selected is proportional to the frequency of this query To balance the number of query pairs of each type of relationship we select one third of query pairs being unrelated one third being equivalent and one third in which 1007 April 26 30 Raleigh NC USA one query belongs to another In this step the relationship between a pair of queries is predicted by the intent tree generated either by DMST or HAC one of them is randomly selected to make each prediction This will not lead to biased results because the pre dicted relationship is not indicated to the worker in any manner We have 2500 Mechanical Turk questions for the 10 classes of entities and assign each question to three workers paying 1 cent to each We find the chance that a worker agrees with another one is 69 8 We only consider the questions on which at least two workers agree and there are 2327 of such questions The distribu tion of answers to these questions is as follows 930 unrelated 384 belongs 288 contains and 725 equivalent In order to s
25. e models of topic correlations JCML 06 19 M Pasca Organizing and searching the world wide web of facts step two harnessing the wisdom of the crowds www 07 20 M Pasca and E Alfonseca Web derived resources for Web information retrieval From conceptual hierarchies to attribute hierarchies SIGIR 09 21 R Sibson SLINK An optimally efficient algorithm for the single link cluster method Computer Journal 16 30 34 1973 22 X Wang D Chakrabarti K Punera Mining broad latent query aspects from search sessions KDD 09 23 K Wang T Walker and Z Zheng PSkip estimating relev ance ranking quality from web search clickthrough data KDD 09 24 J R Wen J Y Nie H J Zhang Clustering user queries of a search engine WWW 01 4 5 6 7 8
26. ect such relationships which are the goals of our study WWW 2010 Full Paper By instruments Bands and artists Instruments Agents and managers Bands and artists Regional Figure 3 ODP taxonomy for music Another related category of studies is how to find related que ries as query suggestions which has been studied from may pers pectives Cui et al 7 uses the clicked URLs of queries to find related queries In 5 Chien and Immorlica find related queries using the temporal query frequency information Cao et al 3 first group queries into clusters in an offline process and then utilize query sequences to find related queries In 22 Wang et al propose an approach for generating the broad intent aspects for query suggestions based on search session data In 19 Pasca studies extracting different classes of entities and top attributes for each class Our study is different from these studies because we study a different problem of building taxonomy of query in tents We infer relationships between generic intents for a class of entities which is more feasible than inferring relationships be tween queries and does not suffer from ambiguity of keywords 3 PROBLEM FORMULATION Our study is about the generic intents on certain classes of enti ties A class of entities is a set of entities that people usually con sider them to be of the same type
27. ee whether the answers by Mechanical Turk are ac curate we randomly choose 100 query pairs with Mechanical Turk judgments ten from each class and manually identify the relationships between each pair of queries Unless the relationship is obvious by common sense we figure out the intents of each query according to the most clicked URL on Bing The average accuracy of Mechanical Turk judgments is 0 83 The precision and recall for query pairs with each type of relationships is shown in Table 3 Query pairs with contains relationship is converted into those of belongs by switching the order In general Mechanical Turk judgments are quite accurate Please note although there are only three categories of results in Table 3 for each question there are four possible answers because belongs relationship can happen in both directions Table 3 Precision Recall of Mechanical Turk Precision Recall Fi Unrelated 1 000 0 727 0 842 Belongs 0 680 0 895 0 773 Equivalent 0 944 0 919 0 931 In general the judgments by Mechanical Turk are reasonably accurate There are some errors such as qvc handbags is consi dered to belong to qvc outlet while qvc is mainly an online store and these two queries have very different intents 6 3 Query and Intent Phrase Relationship We first check whether the query relationship defined in Def 1 in Section 4 2 is consistent with the judgments by Mechanical Turk
28. een two intent phrases based on user clicks which indicate their intents Although it is often difficult to accu rately infer the relationship between two queries there are usually many entities that co appear with two intent phrases in queries and we can aggregate the relationships between these queries to infer the relationship between the two intent phrases Our experi ments show such aggregation improves accuracy significantly The third challenge is how to organize intent phrases into a tree so that each node contains intent phrases corresponding to the same intent and the child nodes of a node correspond to sub concepts of this node We propose three algorithms for this task The first algorithm is based on Directed Maximum Spanning Tree 6 10 The second algorithm is adapted from Hierarchical Ag glomerative Clustering 8 21 The third algorithm is based on Pachinko allocation models 18 for building hierarchical topic models for documents These algorithms are evaluated with hu man labeled data on ten classes of entities The rest of this paper is organized as follows We discuss re lated work in Section 2 Section 3 presents the problem formula tion and Section 4 describes how to infer the relationships be tween intent phrases We present approaches for building intent trees in Section 5 and empirical study in Section 6 This study is concluded in Section 7 2 RELATED WORK There are some studies using existing taxonom
29. ents of query q belong to those of q if the information or web pages that can satisfy the needs of q can also satisfy the needs of q3 For example a user searching for Seattle is usually looking for tourism general information or government of Seattle Therefore the intents of queries like Seat tle tourism and Seattle government belong to those of Seat tle Queries q and q are likely to be equivalent if the intents of each of them belong to those of the other The search intent of a query is inherently subjective and varies Table 1 Top intent phrases for certain classes of entities Actors Cities Musicians Universities actor city lyrics library photos city of music employment biography news youtube jobs pictures real estate wikipedia bookstore imdb hospital songs address bio apartments Wiki athletics wikipedia jobs discography alumni movies map biography tuition WWW 2010 Full Paper for different users As shown by Lee et al 17 the intent of a query can usually be indicated by clicks of the user It is also pro posed by Baeza Yates et al 2 that relationships between queries can be defined with clicked URLs Therefore we propose an ap proach for deciding the relationship between intents of two que ries by their clicked URLs For a query q submitted by a user a search engine returns a ranked list of result URLs There are three possibilities for each result UR
30. er intents These incorrect relationships still capture the user behaviors and the intent trees can still be reasonable and useful with these errors The intent tree for musicians generated by DMST is shown in Figure 2 Here we show two other intent trees by HAC Figure 9 shows the intent tree of car models and Figure 10 shows that of U S Presidents Because of space limitation we only show nodes with most popular intent phrases The query frequency of a node n is defined as the sum of frequencies of queries consisting of an entity and an intent phrase in n or any descendants of n We show all nodes with query frequency higher than 0 5 of the query frequency of root node and ignore all intent phrases that are only slightly different with another intent phrase in the same node with higher frequency e g we ignore biography of if biography is shown The intent phrases within each node and child nodes of each node are ordered by their query frequencies From the intent trees we can see the hierarchical structures of user intents being captured For example in the intent tree for musicians the node music videos has child nodes of song albums cd music videos mp3 etc which are different types of music In the intent tree for car models the node parts accessories has child nodes of performance parts aftermarket parts body parts and auto parts which are popular types
31. f node has a single parent node To convert a semi intent tree into an intent tree we simply merge all nodes connected by equivalent edges into a single node Given the intent phrases w wx for a class of entities we build a semi intent tree T of nodes n nx with n containing phrase w The score of T is measured as s T Yaw cw Yaw cw Vi j n echild n Vi j n equivalent n JM j Jm eeg j max d w ow t Vi jnjechild n wy W n w x Z Vi n echild root 5 where equivalent n is the equivalent nodes of n Building Optimal Semi Intent Tree In order to build an optimal semi intent tree we first convert the intent phrases and their relationships into a directed graph G For each phrase w we create a node n containing w For each two phrases w and wj with d w Cc wj gt 0 we add a belonging edge from n to n with weight d w wj We add two equiva lent edges between n and n in both directions with weight d w S wj d w Cw if d w E w2 gt d w E w gt B d wy S w2 gt a d w S w and d w S w1 gt a 1005 April 26 30 Raleigh NC USA d w S w3 is a parameter between 0 and 1 and a larger a indicates more strict criterion for considering two nodes as equiv alent If 1 then no nodes are considered equivalent Given the graph G we apply Edmond s algorithm 10 to build a directed maximum spanning tree Edmond
32. fferent aspects of information by Bing Kosmix and Yahoo We use x to represent a web search query x WWW 2010 Full Paper for a class of entities into a simple flat list But these intents ac tually form a taxonomy as some intents are related and some are sub concepts of others Take Britney Spears or any musician as an example Albums songs lyrics and music videos are all about her music biography and profile are about informa tion of her life and career and tours and tickets are about her concerts Showing a flat list of search tasks or direct answers that mix up different aspects of information makes it difficult for users to find what he wants or get an overview of the entity The purpose of our study is to organize the generic search in tents for a class of entities into a taxonomy according to the rela tionship between different intents A class of entities is a set of entities that are usually considered to be of the same type such as musicians movies car models cities etc We work on classes of entities instead of individual ones because the sparseness and noises in the data prevent us from accurately inferring the rela tionships between intents of different queries involving a single entity By aggregating data involving many entities we can better capture the relationships between intents An example taxonomy of major generic search intents for musicians
33. for many different musicians we can know lyrics is an important search intent for musicians song albums song lyrics lyrics for lyrics downloads videos de pictures photos pics pictures of images band tickets tour dates oncert schedule fan club fan concert dates biography wikipedia wiki _whois _ bio singer death Figure 2 Result intent tree of musicians most popular phrases only 1002 April 26 30 Raleigh NC USA There are usually many users explicitly indicating their search intents in queries and thus we can capture almost every major generic search intent for a class of entities We represent such generic intents using words and phrases co appearing with entities in user queries If a word or phrase co appears with a number of entities in user queries e g lyrics biography it usually represents a generic intent and we call it an intent phrase The second challenge is how to infer the relationship between two intent phrases Some intent phrases carry the same intent such as pictures pics photos images in queries for mu sicians Some intents are sub concepts of some other intents For example wallpapers are a type of pictures and wikipedia is source of biography We propose a method for inferring the relationship betw
34. involved in computing s 7 and add edge between their corresponding nodes to construct a semi intent tree T add ing equivalent edges for phrases of same node in 7 In this way s T s T and T can be converted into T THEOREM 1 The intent tree T converted from the semi intent tree T built by directed maximum spanning tree algorithm is the intent tree with maximum score PROOF If T is not optimal then there exists intent tree T so that s T gt s T By Lemma 1 there exists semi intent tree T so that s I gt s T which contradicts with T being the maximum semi spanning tree 5 2 Method Based on Hierarchical Agglomer ative Clustering The above algorithm can find the optimal intent tree defined based on the maximum spanning tree Its disadvantage is that the algorithm cannot merge nodes or edges during the procedure for finding maximum spanning tree When we merge multiple equiva lent intent phrases into a single node or put one node as a child of another node we should consider all intent phrases in the merged node and descendant nodes as a whole and measure its overall relationships with other nodes To accomplish this task we modi fy the hierarchical agglomerative clustering average link algo rithm 8 so that it can run on a directed graph and perform two types of merging operations 1 Putting one node as a child of another node and 2 merging two nodes into one O
35. may represent the same generic intent Take intent phrases for musicians as an example Biography bio biography of all represent the same generic intent Some intent phrases may indicate generic intents that are sub concepts of others For example history life and death represent sub concepts of biography 1003 April 26 30 Raleigh NC USA For each class of entities our approach performs three tasks 1 Find all important intent phrases 2 Through the query click behaviors of users determine the relationships between intent phrases i e whether one intent phrase is equivalent to unrelated to or represents a sub concept of another intent phrase 3 Based on the relationships organize the intent phrases into a hierarchical structure like the one in Figure 2 so that each node represents a generic intent and contains all intent phrases for it and the child nodes of each node represent sub concepts of that node The cor rectness of the intent trees will be evaluated by human judges Non Goals In this paper we do not study the following problems 1 How to find a class of entities which can usually be done through pars ing Wikipedia categories lists or using automatic approaches such as that proposed by Pasca 19 2 Ranking search tasks or gener ic intents for a specific entity which can be done by analyzing query frequencies query refinements and clicked URLs of que
36. of the search intents of q being included in those of q2 This is a chal lenging problem as search intent is implicitly implied by user behaviors Because clicked URLs indicate query intents distribu tion of clicks on the clicked URLs of a query is a good indicator of distribution of query intents For example the major clicked URLs of four queries involving Seattle are shown in Figure 4 together with the percentage of Bing search users clicking on each URL for each query We can see the major user intents for query Seattle are about tourism official site of city and ho tels attractions restaurants Most users querying for city of Seat tle and Seattle tourism click on URLs frequently clicked for query Seattle which indicate the intents of city of Seattle and Seattle tourism are included in that of Seattle DEFINITION 1 Relationship between intents of queries Let U q be the set of URLs clicked for q For two queries q and qn the degree of q s intents being included in q s is defined as X click q u rel q u d ueU q q S q X click q u uc q d q q2 is the average relevance to q2 of each clicked URL of q and d q E q2 E 0 1 If we could assume a user only clicks on URLs that satisfy his search intent and only skips URLs that do not then rel q u is the estimated probability of a user who submits query q considers URL u to satisfy his intent In this way d q q2 is the estima
37. or sale used cars equivalent 807 549 653 854 379 525 723 873 791 manual query intent PAM Approach MTurk relation Phrase baseline PMST HAG relation Accuracy by 0830 0 540 0 630 0 586 0 610 0 760 labeled examples Accuracy by N A 0 543 0 578 0 532 0 560 0 675 MTurk data Our approaches are less accurate in predicting the belonging re lationships between intent phrases There are two reasons First user clicks do not always represent user intents There are many other factors such as positional bias URLs containing many dif ferent contents etc which may cause our approaches to make mistakes For example HAC assigns quotations as a child of quotes for U S presidents probably because quotations ap pears in much fewer queries than quotes But they are actually equivalent Second user intents are often different from meanings of words For example most users searching for car model transmission are interested in the transmission problems of the 1009 repair manual service manual owners manual maintenance troubleshooting transmission transmission problems problems recalls Figure 9 Intent tree of car models by HAC car model and thus HAC puts transmission as a child of prob lems which is considered to be incorrect because transmission may have oth
38. query relationships are computed from aggregated data of queries and clicked skipped URLs of the year 2008 The job runs on hundreds of CPUs we cannot disclose the exact number and takes 7 hours 10 minutes The output contains about 5B pairs of queries that share clicked URLs The job for the second step takes 2 hours 26 minutes and outputs from 2K to 657K intent phrases for different classes of entities with relationships between them Table 10 shows the run time on each class of entities by each approach with optimal parameters The top 400 intent phras es are used for each class DMST and HAC are much more effi cient than PAM because they use the pre computed relationships between intent phrases We also test the scalability of DMST and HAC w r t number of intent phrases as shown in Figure 11 Both approaches are super linear in time complexity O mn for DMST and O n for HAC for n intent phrases and m relationships between pairs of phrases Their run time grows slower than the theoretical bound 106 times for DMST and 46 times for HAC when number of phrases grows 16 times Usually hundreds of intent phrases are sufficient for capturing all important search intents and both DMST and HAC can generate the intent tree in no more than a few seconds 10 0 1 Runtime sec 0 01 100 200 400 intent phrases 1600 Figure 11 Scalability of DMST and HAC 7 Conclusions and Future Work In this paper we study
39. s algorithm is for building directed minimum spanning tree which is equivalent to building directed maximum spanning tree The main idea of Ed mond s algorithm is as follows 1 Select the edge with smallest weight that points to each node 2 If no cycle formed stop 3 For each cycle formed contract the nodes on the cycle into a pseudo node and modify the weight of each edge pointing to this pseudo node according to weights of edges in the cycle 4 Select the edge with smallest weight pointing to the pseudo node to re place another edge 5 Go to step 2 with the new graph For a graph with n nodes and m edges Edmond s algorithm takes O mn time A more efficient algorithm is presented in 12 which takes O m n logn time Since we only care about intent phrases that co appear with at least a certain number of entities the number of intent phrases is usually limited and Edmond s algorithm is efficient enough The maximum spanning tree on graph G is a semi intent tree which can be easily converted into an intent tree by merging equivalent nodes Proof of Optimality Here we give the sketch of the proof that the above algorithm can generate the optimal intent tree 7 so that s 7 gt s T for any valid intent tree T LEMMA 1 For any intent tree T there exists a semi intent tree T so that s T s T and T can be converted into T PROOF SKETCH Select all pairs of phrases w and w so that d w S w is
40. t of any other node and n should be a child of the root MST W n is the undirected maxi mum spanning tree constructed for a graph built from W n with a node for each phrase w and an edge with weight d w w d w S wg between phrases w and wy s T contains three terms The first and second terms are for re lationships between all pairs of parent and child nodes and the third term is for the inherent consistency of phrases within each node If we consider the graph with each intent phrase as a node and the relationship between each two intent phrases as an edge then s T is defined based on a spanning tree on this graph We will show that the intent tree with maximum score can be built by finding the directed maximum spanning tree in a graph We hope to build an intent tree from the graph of intent phrases so that each tree node may contain multiple equivalent intent phrases However this is not allowed in the tree generated by a maximum spanning tree MST algorithm To use a MST algo rithm we define a semi intent tree which can be generated by an MST algorithm and can be converted into an intent tree Each non roof node in a semi intent tree contains a single intent phrase Between each pair of nodes n and m there is zero or one directed edge Each edge n n2 is either a belonging edge which indi cates n is a child of m2 or an equivalence edge which indicates n and n contain equivalent intent phrases Each non roo
41. ted of w and an entity in the corresponding class We find the intent phrase with highest query frequency is usually a good representative for a node which is shown in the example intent trees in Section 6 4 5 3 Baseline Method Based on Pachinko Allo cation models The problem of generating hierarchical topic models has been studied in language modeling and 18 presents a most popular approach called Pachinko Allocation models PAM which builds a hierarchy of topics from a set of documents with each topic being a distribution of words We adapt this approach for building intent trees which serves as our baseline approach To apply topic modeling on intent phrases for a class of entities we consider all intent phrases from queries that lead to clicks on a certain URL u to be intent phrases of u represented by W u W u is a multi set and the count of intent phrase w in W u is the sum of click q u for any query q consisted of w and an entity in the specific class In language modeling it is assumed that key words appearing in the same document are generated from the topic distribution this document We make a similar assumption that each intent phrase in W u is generated from the topic distri bution of URL u In this way we can convert each URL into a pseudo document that contains a bag of intent phrases and can apply PAM on these pseudo documents to generate hierarchical topic models The output of PAM is a four level DA
42. ted probability that a clicked URL for q can satisfy the search intent of a user querying for q2 This will be a good estimation of the probability that q s intents are included in q s intents if each URL does not satisfy multiple different intents In reality some URLs can satisfy different but 2 1004 April 26 30 Raleigh NC USA F 75 www seattle gov hyo Seale official site of city en wikipedia org wiki seattle 18 www visitseattle org Seattle Seasin convention and visitor s bureau eattle tourism 585 www seattle gov html visitor visiting seattle 18 www seattle com mcd hotels attractions restaurants eattle hotels 145 www seattle downtown com ese Seattle hotels Figure 4 Percentage of users clicking each URL for four queries involving Seattle usually related intents For example www seattle com contains information about Seattle s hotels restaurants attractions etc which may confuse our algorithm by mixing up two weakly re lated or even unrelated intents Fortunately our goal is to identify the relationships between dif ferent intent phrases and each intent phrase usually co appears with many different entities in many queries We can achieve higher accuracy by aggregating the relationships between many pairs of queries involving different entities DEFINITION 2 Relationship between intent phrases Let f g be the query frequen
43. ther hierar chical agglomerative clustering algorithms Single link and com plete link are not used because they are highly vulnerable to noises and the addition or removal of one edge can often signifi cantly change the result clusters Our algorithm is described in Algorithm 1 Lines 1 6 are for in itializing a graph whose nodes represent the intent phrases and edges represent relationships between intent phrases There are two types of edges A type 1 edge n n indicates putting node n WWW 2010 Full Paper as a child of n A type 2 edge n nj indicates merging n as an equivalent node of n which means merging the phrases and child nodes of n into those of n and removing n Let d n Cc nj represent the average degree of any intent phrase of n belongs to that of n We add a type 1 edge from n to n if d n e nj gt 0 and add a type 2 edge if r a n Cn t d n S nj gt max a n cn d y S n s where r is a parameter between 0 5 and 1 controlling whether to merge two nodes into one or put one as a child of another Ifr 0 5 we only put nodes as children of other nodes if r 1 we only merge two nodes into one Lines 7 16 are for repeatedly merging nodes In each iteration the edge with highest weight is selected If it is a type 1 edge we put one node as a child of the other otherwise we merge the two nodes into one After merging node n into n in either way we need to update the weights of edges involving n
44. wi 5 add edge n n with weight r aw Cw d wj E wi and type 2 6 E lt set of all edges in G 7 while max c e weight gt min_sim 8 n n edge in E with maximum weight 9 ifn or n has been merged with other nodes then continue 10 if n nj type 1 11 merge n as a child of n 12 if n n type 2 13 merge the phrases and children of n as into those of n 14 remove n 15 for each node n with no parent 16 update n n weight and n n weight 17 create intent tree T with T root being an empty node 18 put every node n with no parent as child of T root 19 output 7 1006 April 26 30 Raleigh NC USA handles simple undirected similarity between objects Second our algorithm can put one node as a child node of another besides merging two nodes into one as in average link algorithm Com paring to the DMST based algorithm described in Section 5 1 this algorithm updates the relationships between different nodes after any node is modified When deciding which nodes to be merged it considers the relationship between all descendants of each node and all intent phrases on these descendants This makes it more accurate than the DMST based algorithm according to our expe riments although it is a greedy algorithm It is sometimes necessary to select a representative intent phrase within a node We define the query frequency of an intent phrase w as the sum of frequency of all queries consis
45. y to generate re lated queries or organize queries or information into an existing taxonomy Ariel et al 11 utilizes query click logs and the tax onomy of Open Directory Project ODP www dmoz org to gen erate related keywords for advertisers In 20 Pasca and Alfonse ca utilize WordNet and Wikipedia to organize the information extracted from Web into a hierarchy Hu et al 16 identify search intents of queries by considering Wikipedia categories and articles as possible search intents Our approach is different from above approaches as we do not utilize any existing manually edited taxonomies This is because these taxonomies are very different from user intents For example the ODP taxonomy for Music is shown in Figure 3 Neither this taxonomy nor the taxonomies from Wikipedia or WordNet can be used to represent the search intents for musicians Therefore we purely rely on user behaviors to build hierarchies of generic search intents Baeza Yates et al study relationships between queries 2 in which IS A relationship is mentioned but not studied or com puted Wen et al 24 study query clustering by their similarities In 4 Chuang et al use hierarchical agglomerative clustering to organize queries into hierarchical clusters These papers study how to organize queries using their similarities But they do not infer the IS A relationships between queries or search intents and do not build taxonomies that refl
Download Pdf Manuals
Related Search
Related Contents
2 Bedienung des ND 780 - Raytech Measuring Systems SERVICE MANUAL & PARTS LIST 1600P-QC 3-Finger-Zentrischgreifer Type: PZN-plus 3 Controle su uso y costos de energía de manera local o desde - Security Server - Security Server REGOLAMENTO COMUNALE D`IGIENE バックアップ/リストア 取扱説明書 Emtec 8GB Batman Dell PowerVault MD3220i User's Manual Copyright © All rights reserved.
Failed to retrieve file