Content deleted Content added
←Created page with ''''Bidirectional Recurrent Neural Networks'''('''BRNN''') were invented in 1997 by Schuster & Paliwal <ref name="Schuster"> Schuster, Mike, and Kuldip K. Paliwal...' Tag: Possible vandalism |
No edit summary |
||
(42 intermediate revisions by 31 users not shown) | |||
Line 1:
{{short description|Type of artificial neural network}}
'''Bidirectional Recurrent Neural Networks'''('''BRNN''') were invented in 1997 by Schuster & Paliwal <ref name="Schuster"> Schuster, Mike, and Kuldip K. Paliwal. "Bidirectional recurrent neural networks." Signal Processing, IEEE Transactions on 45.11 (1997): 2673-2681.2. Awni Hannun, Carl Case, Jared Casper, Bryan Catanzaro, Greg Diamos, Erich Elsen, Ryan </ref>. BRNN is introduced to increase the amount of input information available to be referred to. For example, [[multilayer perceptron]](MLP) and [[time delay neural network]](TDNN) have limitation on the input data flexibility, as they require their input data to be fixed. Standard [[recurrent neural network]](RNN) also has restrictions as its future input information cannot be reached from their current state. On the contrary, BRNN do not require their input data to be fixed. Moreover, their future input information is reachable from the current state. The basic idea of BRNN is to connect two hidden layers of opposite directions to the same output. By this structure, the output layer can get information from past and future states. ▼
▲'''Bidirectional
BRNN are especially useful when the context of the input is needed. For example, in handwriting recognition, the performance can be enhanced by knowledge of the letters located before and after the current letter.▼
▲BRNN are especially useful when the context of the input is needed. For example, in [[handwriting recognition]], the performance can be enhanced by knowledge of the letters located before and after the current letter.
==Architecture==
[[File:
The principle of BRNN is to split the neurons of a regular RNN into two directions, one for positive time direction (forward states), and another for negative time direction (backward states). Those two
==Training==
==Applications==
Applications of BRNN include :
*Speech Recognition (Combined with [[Long short-term memory]])<ref>
*Translation<ref>
*Handwritten Recognition<ref>Liwicki, Marcus, et al. "[https://mediatum.ub.tum.de/doc/1289961/file.pdf A novel approach to on-line handwriting recognition based on bidirectional long short-term memory networks]." Proc. 9th Int. Conf. on Document Analysis and Recognition. Vol. 1. 2007.</ref>▼
*Industrial [[Soft sensor]]<ref>Lui, Chun Fai, et al. "[https://ieeexplore.ieee.org/ielx7/19/9717300/09718226.pdf A Supervised Bidirectional Long Short-Term Memory Network for Data-Driven Dynamic Soft Sensor Modeling]." IEEE Transactions on Instrumentation and Measurement 71 (2022): 1-13.</ref>
▲*Handwritten Recognition<ref>Liwicki, Marcus, et al. "A novel approach to on-line handwriting recognition based on bidirectional long short-term memory networks." Proc. 9th Int. Conf. on Document Analysis and Recognition. Vol. 1. 2007.</ref>
*Protein Structure Prediction<ref>
*Part-of-speech tagging
▲*Protein Structure Prediction<ref> Baldi, Pierre, et al. "Exploiting the past and the future in protein secondary structure prediction." Bioinformatics 15.11 (1999): 937-946.</ref><ref>Pollastri, Gianluca, and Aoife Mclysaght. "Porter: a new, accurate server for protein secondary structure prediction." Bioinformatics 21.8 (2005): 1719-1720.</ref>
*Dependency Parsing<ref>{{Cite journal|last1=Kiperwasser|first1=Eliyahu|last2=Goldberg|first2=Yoav|date=2016|title=Simple and Accurate Dependency Parsing Using Bidirectional LSTM Feature Representations|url=https://www.aclweb.org/anthology/Q16-1023/|journal=Transactions of the Association for Computational Linguistics|language=en-us|volume=4|pages=313–327|doi=10.1162/tacl_a_00101|arxiv=1603.04351|bibcode=2016arXiv160304351K|s2cid=1642392}}</ref>
*Entity Extraction<ref>{{Cite arXiv|last1=Dernoncourt|first1=Franck|last2=Lee|first2=Ji Young|last3=Szolovits|first3=Peter|date=2017-05-15|title=NeuroNER: an easy-to-use program for named-entity recognition based on neural networks|eprint=1705.05487|class=cs.CL}}</ref>
==References==
{{reflist}}
==External links==
*[https://github.com/hycis/bidirectional_RNN] Implementation of BRNN/LSTM in Python with Theano
[[Category:Neural network architectures]]
|