Content deleted Content added
m Open access bot: url-access updated in citation with #oabot. |
Jario.Zhang (talk | contribs) mNo edit summary |
||
Line 4:
is the term for the collection of all methods in artificial intelligence research that are based on high-level [[physical symbol systems hypothesis|symbolic]] (human-readable) representations of problems, [[Formal logic|logic]] and [[search algorithm|search]].<ref>{{Cite journal|last1=Garnelo|first1=Marta|last2=Shanahan|first2=Murray|date=2019-10-01|title=Reconciling deep learning with symbolic artificial intelligence: representing objects and relations|journal=Current Opinion in Behavioral Sciences|language=en|volume=29|pages=17–23|doi=10.1016/j.cobeha.2018.12.010|s2cid=72336067 |doi-access=free|hdl=10044/1/67796|hdl-access=free}}</ref> Symbolic AI used tools such as [[logic programming]], [[production (computer science)|production rules]], [[semantic nets]] and [[frame (artificial intelligence)|frames]], and it developed applications such as [[knowledge-based systems]] (in particular, [[expert systems]]), [[symbolic mathematics]], [[automated theorem provers]], [[ontologies]], the [[semantic web]], and [[automated planning and scheduling]] systems. The Symbolic AI paradigm led to seminal ideas in [[Artificial intelligence#Search and optimization|search]], [[symbolic programming]] languages, [[Intelligent agent|agents]], [[multi-agent systems]], the [[semantic web]], and the strengths and limitations of formal knowledge and [[automated reasoning|reasoning systems]].
Symbolic AI was the dominant [[paradigm]] of AI research from the mid-1950s until the mid-1990s.{{sfn|Kolata|1982}} Researchers in the 1960s and the 1970s were convinced that symbolic approaches would eventually succeed in creating a machine with [[artificial general intelligence]] and considered this the ultimate goal of their field.<ref>{{
[[Artificial neural network|Neural networks]], a subsymbolic approach, had been pursued from early days and reemerged strongly in 2012. Early examples are [[Frank Rosenblatt|Rosenblatt]]'s [[perceptron]] learning work, the [[backpropagation]] work of Rumelhart, Hinton and Williams,<ref>{{cite journal| doi = 10.1038/323533a0| issn = 1476-4687| volume = 323| issue = 6088| pages = 533–536| last1 = Rumelhart| first1 = David E.| last2 = Hinton| first2 = Geoffrey E.| last3 = Williams| first3 = Ronald J.| title = Learning representations by back-propagating errors| journal = Nature| date = 1986 | bibcode = 1986Natur.323..533R| s2cid = 205001834}}</ref> and work in [[convolutional neural network]]s by LeCun et al. in 1989.<ref>{{Cite journal| volume = 1| issue = 4| pages = 541–551| last1 = LeCun| first1 = Y.| last2 = Boser| first2 = B.| last3 = Denker| first3 = I.| last4 = Henderson| first4 = D.| last5 = Howard| first5 = R.| last6 = Hubbard| first6 = W.| last7 = Tackel| first7 = L.| title = Backpropagation Applied to Handwritten Zip Code Recognition| journal = Neural Computation| date = 1989| doi = 10.1162/neco.1989.1.4.541| s2cid = 41312633}}</ref> However, neural networks were not viewed as successful until about 2012: "Until Big Data became commonplace, the general consensus in the Al community was that the so-called neural-network approach was hopeless. Systems just didn't work that well, compared to other methods. ... A revolution came in 2012, when a number of people, including a team of researchers working with Hinton, worked out a way to use the power of [[GPUs]] to enormously increase the power of neural networks."{{sfn|Marcus |Davis|2019}} Over the next several years, [[deep learning]] had spectacular success in handling vision, [[speech recognition]], speech synthesis, image generation, and machine translation. However, since 2020, as inherent difficulties with bias, explanation, comprehensibility, and robustness became more apparent with deep learning approaches; an increasing number of AI researchers have called for [[Neuro-symbolic AI|combining]] the best of both the symbolic and neural network approaches<ref name="Rossi">
|