Content deleted Content added
Tom.Reding (talk | contribs) m →Criticisms and other Lesk-based methods: Rep typographic ligatures like "fi" with plain text; possible ref cleanup; WP:GenFixes on, replaced: fi → fi (2) using AWB |
|||
Line 5:
==Overview==
The Lesk algorithm is based on the assumption that words in a given "neighborhood" (section of text) will tend to share a common topic. A simplified version of the Lesk algorithm is to compare the dictionary definition of an ambiguous word with the terms contained in its neighborhood. Versions have been adapted to use [[WordNet]].<ref>Satanjeev Banerjee and Ted Pedersen. ''[http://www.cs.cmu.edu/~banerjee/Publications/cicling2002.ps.gz An Adapted Lesk Algorithm for Word Sense Disambiguation Using WordNet]'', Lecture Notes
</ref> An implementation might look like this:
# for every sense of the word being disambiguated one should count the amount of words that are in both neighborhood of that word and in the dictionary definition of that sense
|