Deep backward stochastic differential equation method: Difference between revisions

Content deleted Content added
WikiCleanerBot (talk | contribs)
m v2.05b - Bot T20 CW#61 - Fix errors for CW project (Reference before punctuation)
mv citation from section titles to inline
Line 116:
 
==Algorithms==
[[File:Gradient descent Hamiltonian Monte Carlo comparison.gif|thumb|upright=0.9|Gradient descent vs Monte carloCarlo]]
 
* First, we present the pseudocode for the ADAM algorithm as follows:<ref name="Adam2014">{{cite arXiv |first1=Diederik |last1=Kingma |first2=Jimmy |last2=Ba |eprint=1412.6980 |title=Adam: A Method for Stochastic Optimization |year=2014 |class=cs.LG }}</ref>
===Adam optimizer===
===Adam<ref name="Adam2014">{{cite arXiv |first1=Diederik |last1=Kingma |first2=Jimmy |last2=Ba |eprint=1412.6980 |title=Adam: A Method for Stochastic Optimization |year=2014 |class=cs.LG }}</ref> (short for Adaptive Moment Estimation) algorithm===
This function implements the Adam<ref optimizationname="Adam2014">{{cite arXiv |eprint=1412.6980 |class=cs.LG |first1=Diederik |last1=Kingma |first2=Jimmy |last2=Ba |title=Adam: A Method for Stochastic Optimization |year=2014}}</ref> algorithm for minimizing the target function <math>\mathcal{G}(\theta)</math>.
 
'''Function:''' ADAM(<math>\alpha</math>, <math>\beta_1</math>, <math>\beta_2</math>, <math>\epsilon</math>, <math>\mathcal{G}(\theta)</math>, <math>\theta_0</math>) '''is'''
Line 143:
 
* With the ADAM algorithm described above, we now present the pseudocode corresponding to a multilayer feedforward neural network:
===Backpropagation algorithm===
===Backpropagation algorithm<ref name="DLhistory">{{cite arXiv |eprint=2212.11279 |class=cs.NE |first=Juergen |last=Schmidhuber |author-link=Juergen Schmidhuber |title=Annotated History of Modern AI and Deep Learning |date=2022}}</ref> for multilayer feedforward neural networks===
This function implements the backpropagation algorithm for training a multi-layer feedforward neural network.