Content deleted Content added
m hyperlinks |
m Removed invisible unicode characters + other fixes, replaced: → using AWB (12020) |
||
Line 11:
The application of document clustering can be categorized to two types, online and offline. Online applications are usually constrained by efficiency problems when compared to offline applications.
In general, there are two common algorithms. The first one is the hierarchical based algorithm, which includes single link, complete linkage, group average and Ward's method. By aggregating or dividing, documents can be clustered into hierarchical structure, which is suitable for browsing. However, such an algorithm usually suffers from efficiency problems. The other algorithm is developed using the [[K-means algorithm]] and its variants. Generally hierarchical algorithms produce more in-depth information for detailed analyses, while algorithms based around variants of the [[K-means algorithm]] is more efficient and provides sufficient information for most purposes.<ref>Manning, Chris, and Hinrich Schütze,''''' Foundations of Statistical Natural Language Processing'Italic text''', MIT Press. Cambridge, MA: May 1999. Chapter 14'</ref>
These algorithms can further be classified as hard or soft clustering algorithms. Hard clustering computes a hard assignment – each document is a member of exactly one cluster. The assignment of soft clustering algorithms is soft – a document’s assignment is a distribution over all clusters. In a soft assignment, a document has fractional membership in several clusters.<ref>Manning, Chris, and Hinrich Schütze,''''' Foundations of Statistical Natural Language Processing'Italic text''', MIT Press. Cambridge, MA: May 1999. Pg 499'</ref>
Other algorithms involve graph based clustering, ontology supported clustering and order sensitive clustering.
Line 28:
== Clustering v. Classifying ==
Clustering algorithms in computational text analysis groups documents into what are called subsets or ''clusters'' where the algorithm's goal is to create internally coherent clusters that are distinct from one another.<ref>{{Cite web|url=http://nlp.stanford.edu/IR-book/|title=Introduction to Information Retrieval|website=nlp.stanford.edu|pages=349|access-date=2016-05-03}}</ref> Classification on the other hand, is a form of [[supervised learning]] where the individual coder creates internal, coherent clusters that are based on either [[Inductive reasoning|inductive]], [[Deductive reasoning|deductive]], or [[Abductive reasoning|abductive]] reasoning. Clustering relies on no supervisory teacher imposing previously derived categories upon the data, just types of distances, of which the most commonly found distance is [[Euclidean distance|Euclidean]].<ref>{{Cite web|url=http://nlp.stanford.edu/IR-book/|title=Introduction to Information Retrieval|website=nlp.stanford.edu|pages=349–50|access-date=2016-05-03}}</ref>
Clustering algorithms rely on [[Latent semantic analysis|Latent Semantic Analysis]] where the documents being analyzed are treated as a "[[Bag-of-words model|bag of words]]," or multidimensional spaces rather than [[
== References ==
|