Generalized vector space model: Difference between revisions

Content deleted Content added
Tag: gettingstarted edit
m Reverted edits by Babe12345654321 (talk) to last version by ClueBot NG
Line 14:
Term correlation <math>t_i \cdot t_j</math> can be implemented in several ways. As an example Wong et al. use as input to their algorithm the term occurrence frequency matrix obtained from automatic indexing and the output is term correlation between any pair of index terms.
 
==Semantic information on GVSM==
this does not exist. youre a babe
 
There are at least two basic directions for embedding term to term relatedness, other than exact keyword matching, into a retrieval model:
# compute semantic correlations between terms
# compute frequency co-occurrence statistics from large corpora
 
Recently Tsatsaronis<ref>{{cite | title=A Generalized Vector Space Model for Text Retrieval Based on Semantic Relatedness | url=http://www.aclweb.org/anthology/E/E09/E09-3009.pdf | last= Tsatsaronis | first=George | coauthors=Vicky Panagiotopoulou | date=2009}}</ref> focused on the first approach.
 
They measure semantic relatedness (''SR'') using a thesaurus (''O'') like [[WordNet]]. It considers the path length, captured by compactness (''SCM''), and the path depth, captured by semantic path elaboration (''SPE'').
They estimate the <math>t_i \cdot t_j</math> inner product by:
 
<math>t_i \cdot t_j = SR((t_i, t_j), (s_i, s_j), O)</math>
 
where ''s<sub>i</sub>'' and ''s<sub>j</sub>'' are senses of terms ''t<sub>i</sub>'' and ''t<sub>j</sub>'' respectively, maximizing <math>SCM \cdot SPE</math>.
 
== References ==