proximity language model
play

Proximity Language Model A Language Model beyond Bag of Words - PowerPoint PPT Presentation

Proximity Language Model A Language Model beyond Bag of Words through Proximity Jinglei Zhao 1 1,2 Yeogirl Yun 1 iZENEsoft, Inc. 2 Wisenut, Inc. 2 Wi t I Outline Introduction The proposed model Proximity Language Model


  1. Proximity Language Model A Language Model beyond Bag of Words through Proximity Jinglei Zhao 1 1,2 Yeogirl Yun 1 iZENEsoft, Inc. 2 Wisenut, Inc. 2 Wi t I

  2. Outline Ⅰ Introduction Ⅱ The proposed model Proximity Language Model Modeling Proximate Centrality of Terms Ⅲ Experiment and Result

  3. Introduction Introduction The proposed model Experiment and Result Background Probabilistic models are prevalent in IR Probabilistic models are prevalent in IR. Documents are represented as “bag of words” (BOW). Statistics usually exploited under BOW: y p Term frequency,inverse document frequency Document length, etc. Merits Simplicity in modeling. Effectiveness in parameter estimation Effectiveness in parameter estimation. Model more under the BOW assumption. BOW are criticized for not capturing the relatedness between terms. Could we model term relatedness while retain the simplicity of probabilistic modeling under BOW?

  4. Introduction Introduction The proposed model Experiment and Result Background Proximity information. Represents the closeness or compactness of the query terms appearing in a document. Underlying intuition of using proximity in ranking: y g g p y g The more compact the terms, the more likely that they are topically related. The closer the query terms appear the more possible the The closer the query terms appear, the more possible the document is relevant. It can be seen as a kind of indirect measure of term relatedness or dependence.

  5. Introduction Introduction The proposed model Experiment and Result Objective Integrate proximity information into Unigram language modeling. g Language modeling has become a very promising direction in IR. Solid theoretical background. g Empirical good performance. This paper’s focus: Develop a systematic way to integrate the term proximity Develop a systematic way to integrate the term proximity information into the unigram language modeling.

  6. Introduction Introduction The proposed model Experiment and Result Related Work Dependency Modeling Dependency Modeling General language model, dependency language model,etc. Shortcoming: The parameter estimation become much more difficult to g p compute and sensitive to data sparse and noise. Phrase Indexing Ph I d i Incorporate bigger unit than word such as phrase or loose phrase in text representation text representation. Shortcoming: The improvement of using phrases is not consistent. Previous Proximity Modeling Span-based, pair-based. Shortcoming: Combining with relevant score at document level Shortcoming: Combining with relevant score at document-level, intuitive, without theoretical ground.

  7. Introduction Introduction The proposed model Experiment and Result Our Approach Integrate Proximity with Unigram Language Model Integrate Proximity with Unigram Language Model View query term’s proximate centrality as Dirichlet hyper-parameters. Combines the score at the term level Combines the score at the term level. Boost a term’s score contribution when the term is at a central place in the proximity structure. Merits A uniform ranking formula. Mathematically grounded. Performs better empirically.

  8. Introduction Proximity Language Model The proposed model Experiment and Result Unigram Language Model Represent query and document as vectors of term counts Represent query and document as vectors of term counts Query and document are generated by multinomial distribution Q d d t t d b lti i l di t ib ti q The relevance of d to is measured by the probability of l q generating by the language model estimated from d l

  9. Introduction Proximity Language Model The proposed model Experiment and Result Integration with Proximity Our belief and expectation Our belief and expectation Given d and d , supposing all others being equal while the query terms a b of q appears more proximate in d than in d , we believe that d should a a b b a a be more relevant to the query than d . b ∧ ∧ θ θ In other words, if and represent the language model estimated a b from d and d respective ly, we believe that the probabilit y that q is a b ∧ ∧ θ θ generated from should be higher than . a b Express our expectation θ Term' s emission probabilit y should be proportion al to each term' s l , , i proximity centrailit y score Prox (w ) with respect to other query terms. d i l θ View Prox (w ) as the weight on . d i l , i 1 θ θ E Express th the above b t two points i t by b using i a conjugate j t prior i on . l

  10. Introduction Proximity Language Model The proposed model Experiment and Result Integration with Proximity θ θ Dirichlet prior on Dirichlet prior on l l θ The posterior estimation of l Th The proximity integrated estimation of the word emission i it i t t d ti ti f th d i i probability

  11. Introduction Proximity Language Model The proposed model Experiment and Result Integration with Proximity Interpretation on proximity document model Interpretation on proximity document model Transform proximity information to word count information. Boost a term’s likelihood when it is proximate to other terms Boost a term s likelihood when it is proximate to other terms. From the original bag of words to a pseudo “bag of words”. More generally, a way of model term relatedness under BOW? Relation with smoothing. g The proximity factor mainly functions to adjust the parameters for seen matching terms with respect to a query in a document. Smoothing is motivated to weight the unseen words in the document.

  12. Introduction Proximity Language Model The proposed model Experiment and Result Integration with Proximity Further smoothing with collection language model The ranking formula under KL divergence framework

  13. Introduction The proposed model Modeling Proximate Centrality of Terms Experiment and Result Term Proximity Measure Term’s Proximate centrality Term s Proximate centrality A key notion in PLM is the estimation of term proximity : Prox (w ). d l i For For non non - query query terms, terms, they they are are assumed assumed to to have have a a constant constant score score of of zero. For a query term, it should be computed according to a proximity measure th t that reflects fl t th the t terms closeness l t to other th query term' t ' s. Measuring Proximity via Pair Distance Measuring Proximity via Pair Distance Represent a term’s proximity by measuring its distance to other query terms in the document terms in the document. How to define a term’s distance to other terms in a document? how to map term distance to the term’s proximate centrality score?

  14. Introduction The proposed model Modeling Proximate Centrality of Terms Experiment and Result Term Proximity Measure Pairwise term distance Pairwise term distance Represented as the distance between the closest occurring positions of the two terms in the document. Pairwise proximity

  15. Introduction The proposed model Modeling Proximate Centrality of Terms Experiment and Result Computation of Term’s Proximate Centrality Term Proximity based on Minimum Distance Term Proximity based on Average Distance Term Proximity Summed over Pair Proximity

  16. Introduction The proposed model Modeling Proximate Centrality of Terms Experiment and Result An example − dist dist P Proximity computed by different measures (f = 1.5 i i d b diff (f 1 5 ) )

  17. Introduction Experiment and Result The proposed model Experiment and Result Experimental Setting Data Set Experimental platform Lemur toolkit. A naive tokenizer. A very small stopword list.

  18. Introduction Experiment and Result The proposed model Experiment and Result Experimental Setting Baselines Basic KL divergence language model (LM) Tao’s document-level linear score combination (LLM).

  19. Introduction Experiment and Result The proposed model Experiment and Result Parameter Setting LM LM The prior collection sample size μ is set to 2000 across all the experiments which is also used in LLM and PLM. LLM Parameter is optimized by searching : 0.1, 0.2, ..., 1.0. PLM Proximity argument λ :controls the proportional weight of prior proximity factor relative to the observed word count information. Exponential weight para:controls the proportional ratio of proximity score between different query terms. q y Optimization space: para : 1.1, 1.2, ..., 2.0, λ : 0.1, 1, 2, 3, ..., 10.

  20. Introduction Experiment and Result The proposed model Experiment and Result PLM’s parameter Sensitivity using P_MinDist.

  21. Introduction Experiment and Result The proposed model Experiment and Result Comparison of Best Performance

  22. Introduction Experiment and Result The proposed model Experiment and Result Main Observation The observations e obse a o s PLM performs empirically better than LM and LLM. LLM fails on Ohsumed collection (more verbose in queries). PLM performs very well on verbose queries. For the three proposed term proximity measures used in PLM, P SumProx and P MinDist performs better than P AveDist . P_ SumProx and P_ MinDist performs better than P_ AveDist .

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend