Content deleted Content added
m link face recognition using Find link |
Altered journal. | Use this tool. Report bugs. | #UCB_Gadget | Alter: title, doi, template type. Add: isbn, chapter, chapter-url. Removed or converted URL. Removed parameters. Some additions/deletions were parameter name changes. | Use this tool. Report bugs. | #UCB_Gadget |
||
(30 intermediate revisions by 18 users not shown) | |||
Line 1:
{{Short description|Geometric algorithm}}
[[File:Diffusion_map_of_a_torodial_helix.jpg|thumb|right|Given non-uniformly sampled data points on a toroidal helix (top), the first two Diffusion Map coordinates with Laplace–Beltrami normalization are plotted (bottom). The Diffusion Map unravels the toroidal helix recovering the underlying intrinsic circular geometry of the data.]]
'''Diffusion maps''' is a [[dimensionality reduction]] or [[feature extraction]] algorithm introduced by [[Ronald Coifman|
==Definition of diffusion maps==
Following <ref name="DifussionMap"/> and
===Connectivity===
Diffusion maps exploit the relationship between [[heat diffusion]] and [[random walk]] [[Markov chain]]. The basic observation is that if we take a random walk on the data, walking to a nearby data-point is more likely than walking to another that is far away. Let <math>(X, \mathcal{A}, \mu)</math> be a [[measure space]], where <math>X</math> is the data set and <math>\mu</math> represents the distribution
Based on this, the connectivity <math>k</math> between two data points, <math>x</math> and <math>y</math>, can be defined as the probability of walking from <math>x</math> to <math>y</math> in one step of the random walk. Usually, this probability is specified in terms of a kernel function of the two points: <math>k: X \times X \rightarrow \mathbb{R}</math>. For example, the popular Gaussian kernel:
Line 25 ⟶ 26:
The kernel constitutes the prior definition of the ''local'' geometry of the data-set. Since a given kernel will capture a specific feature of the data set, its choice should be guided by the application that one has in mind. This is a major difference with methods such as [[principal component analysis]], where correlations between all data points are taken into account at once.
Given <math>(X, k)</math>, we can then construct a reversible discrete-time Markov chain on <math>X</math> (a process known as the normalized graph Laplacian construction):
: <math>
d(x) = \int_X k(x,y) d\mu(y)
Line 74 ⟶ 75:
</math>
One of the main ideas of the diffusion framework is that running the chain forward in time (taking larger and larger powers of <math>M</math>) reveals the geometric structure of <math>X</math> at larger and larger scales (the diffusion process). Specifically, the notion of a ''cluster'' in the data set is quantified as a region in which the probability of escaping this region is low (within a certain time t). Therefore, t not only serves as a time parameter, but it also has the dual role of scale parameter.
The eigendecomposition of the matrix <math>M^t</math> yields
Line 82 ⟶ 83:
</math>
where <math>\{\lambda_l \}</math> is the sequence of eigenvalues of <math>M</math> and <math>\{\psi_l \}</math> and <math>\{\phi_l \}</math> are the biorthogonal
Due to the spectrum decay of the eigenvalues, only a few terms are necessary to achieve a given relative accuracy in this sum.
====Parameter
The reason to introduce the normalization step involving <math>\alpha</math> is to tune the influence of the data point density on the infinitesimal transition of the diffusion. In some applications, the sampling of the data is generally not related to the geometry of the manifold we are interested in describing. In this case, we can set <math>\alpha=1</math> and the diffusion operator approximates the [[Laplace–Beltrami operator]]. We then recover the Riemannian geometry of the data set regardless of the distribution of the points. To describe the long-term behavior of the point distribution of a system of stochastic differential equations, we can use <math>\alpha=0.5</math> and the resulting Markov chain approximates the [[Fokker–Planck equation|Fokker–Planck diffusion]]. With <math>\alpha=0</math>, it reduces to the classical graph Laplacian normalization.
===Diffusion distance===
Line 123 ⟶ 124:
: <math>
D_t(x_i,x_j)^2
</math>
so the Euclidean distance in the diffusion coordinates approximates the diffusion distance.
Line 130 ⟶ 131:
The basic algorithm framework of diffusion map is as:
Step 1. Given the similarity matrix ''L''.
Step 2. Normalize the matrix according to parameter <math>\alpha</math>: <math>L^{(\alpha)} = D^{-\alpha} L D^{-\alpha} </math>.
Step 3. Form the normalized matrix <math>M=({D}^{(\alpha)})^{-1}L^{(\alpha)}</math>.
Step 4. Compute the ''k'' largest eigenvalues of <math>M^t</math> and the corresponding eigenvectors.
Step 5. Use diffusion map to get the embedding <math>\Psi_t</math>.
==Application==
In the paper
to the distribution of the points and therefore provides a separation of the statistics and the geometry of the
data. Since diffusion maps
| last1 = Barkan
| first1 = Oren
Line 155 ⟶ 156:
| title = Fast high dimensional vector multiplication face recognition
| journal = Proceedings of the IEEE International Conference on Computer Vision 2013
|
}}</ref> [[spectral clustering]], low dimensional representation of images, image segmentation,<ref name="Farbman" /> 3D model segmentation,<ref name="sidi11" /> speaker verification<ref name="spk">{{cite
| last1 = Barkan
| first1 = Oren
| last2 = Aronowitz
| first2 = Hagai
|
|
▲ | journal = Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
| date = 2013
|
| doi = 10.1109/ICASSP.2013.6639149
}}</ref> and identification,<ref name="speakerid2011" /> sampling on manifolds, anomaly detection<ref name="Mishne" /><ref>{{Cite journal|last=Shabat|first=Gil|last2=Segev|first2=David|last3=Averbuch|first3=Amir|date=2018-01-07|title=Uncovering Unknown Unknowns in Financial Services Big Data by Unsupervised Methodologies: Present and Future trends|url=http://proceedings.mlr.press/v71/shabat18a.html|journal=KDD 2017 Workshop on Anomaly Detection in Finance|language=en|volume=71|pages=8–19|via=}}</ref>, image inpainting,<ref name="Gepshtein" /> and so on.▼
| isbn = 978-1-4799-0356-6
▲ }}</ref> and identification,<ref name="speakerid2011" /> sampling on manifolds, anomaly detection,<ref name="Mishne" /><ref>{{Cite journal|
| last1 = Margulies
| first1 = Daniel S.
| last2 = Ghosh
| first2 = Satrajit S.
| last3 = Goulas
| first3 = Alexandros
| last4 = Falkiewicz
| first4 = Marcel
| last5 = Huntenburg
| first5 = Julia M.
| last6 = Langs
| first6 = Georg
| last7 = Bezgin
| first7 = Gleb
| last8 = Eickhoff
| first8 = Simon B.
| last9 = Castellanos
| first9 = F. Xavier
| last10 = Petrides
| first10 = Michael
| last11 = Jefferies
| first11 = Elizabeth
| last12 = Smallwood
| first12 = Jonathan
| title = Situating the default-mode network along a principal gradient of macroscale cortical organization
| journal = Proceedings of the National Academy of Sciences
| pages = 12574–12579
| volume = 113
| issue = 44
| year = 2016
| doi = 10.1073/pnas.1608282113
| pmid = 27791099
| pmc = 5098630
| bibcode = 2016PNAS..11312574M
| doi-access = free
}}</ref> and so on.
Furthermore, the diffusion maps framework has been productively extended to [[complex networks]],<ref>{{cite journal
| last1 = De Domenico
| first1 = Manlio
| url = https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.118.168301
| title = Diffusion geometry unravels the emergence of functional clusters in collective phenomena
| journal = Physical Review Letters
| volume = 118
| issue = 16
| pages = 168301
| year = 2017
| doi = 10.1103/PhysRevLett.118.168301
| pmid = 28474920
| arxiv = 1704.07068
| bibcode = 2017PhRvL.118p8301D
| s2cid = 2638868
}}</ref> revealing a functional organisation of networks which differs from the purely topological or structural one.
==See also==
Line 225 ⟶ 281:
| pmc=1140422
| bibcode = 2005PNAS..102.7426C
| doi-access = free
}}</ref>
Line 246 ⟶ 303:
| pmc=1140426
| bibcode = 2005PNAS..102.7432C
| doi-access = free
}}</ref>
Line 258 ⟶ 316:
| volume = 21
| pages = 5–30
| doi-access=
}}</ref>▼
| s2cid = 17160669
▲ }}</ref>
<ref name="Diffusion">{{cite thesis
Line 284 ⟶ 344:
<ref name="Introduction">{{cite journal
|
|
| last2 = Herbst
| first2 = B M
Line 299 ⟶ 359:
<ref name="sidi11">{{cite conference
|
|
| first2 = Oliver
| last2 = van Kaick
Line 315 ⟶ 375:
}}</ref>
<ref name="speakerid2011">{{cite
|
|
| first2 = Ronen
| last2 = Talmon
Line 329 ⟶ 389:
<ref name="Mishne">{{cite journal
|
|
| first2 = Israel
| last2 = Cohen
| year = 2013
| title = Multiscale Anomaly Detection Using Diffusion Maps
| journal = IEEE Journal of Selected Topics in Signal Processing
| volume = 7
| issue = 1
Line 341 ⟶ 401:
| doi = 10.1109/jstsp.2012.2232279
| bibcode = 2013ISTSP...7..111M
| s2cid = 1954466
}}</ref>
<ref name="Gepshtein">{{cite journal
|
|
| first2 = Yosi
| last2 = Keller
Line 357 ⟶ 418:
| pmid = 23322762
| bibcode = 2013ITIP...22.2983G
| s2cid = 14375333
}}</ref>
}}
|