Академический Документы
Профессиональный Документы
Культура Документы
11.04.2018, 21)15
Summary: Convolutional Neural Nets are getting all the press but it’s
Recurrent Neural Nets that are the real workhorse of this generation of
AI.
On one scale RNNs have much more in common with the larger family of
https://www.datasciencecentral.com/profiles/blogs/recurrent-neur…73&utm_medium=social&utm_source=twitter.com&utm_campaign=buffer Strona 1 z 6
Recurrent Neural Nets – The Third and Least Appreciated Leg of the AI Stool - Data Science Central 11.04.2018, 21)15
NNs than do CNNs which have very unique architecture. But RNNs are
unique in their own ways, the most important of which are:
Time
Blocks of text.
Audio speech signals.
Strings of stock prices.
Streams of sensor data.
Unlike many other types of NNs which are feedforward networks, RNNs
have bi-directional data flow. Although each unit of data will enter the
RNN and be analyzed one unit at a time, in this model some of the data
from later processing stages is fed back to early stages to influence the
processing of subsequent data items. In other words, it introduces a sense
of TIME.
While strings of stock prices or sensor data may easily be seen as time
series data, it’s also possible to see strings of text or speech as time series
when you understand that these are being processed one character at time
and in the sequence in which they naturally occur (over time). This is
called ‘character level’ language modeling. So RNNs learn on any data
which can be presented as sequence.
https://www.datasciencecentral.com/profiles/blogs/recurrent-neur…73&utm_medium=social&utm_source=twitter.com&utm_campaign=buffer Strona 2 z 6
Recurrent Neural Nets – The Third and Least Appreciated Leg of the AI Stool - Data Science Central 11.04.2018, 21)15
Memory
The feed backwards loop in RNNs is not just one data unit backwards. It
can be as many units backwards as you specify. In other words, the impact
of the current calculation can be felt in the processing of data items dozens
or thousands of units later. This means that the RNN model is said to
‘remember’ and apply learning over a time scale that the data scientist
controls.
The challenge of memory is that the longer the time we ask the RNN to
remember (process together) the weaker is the influence of the current
action on all subsequent data. There are clearly diminishing returns
involved.
Several approaches have evolved to deal with this all of which can be said
to create a kind of sliding window or gate over which the backward flow of
information can extend. In tuning, this gives the author direct control over
how much influence current learning should have over data received later
and how rapidly the RNN should ‘forget’.
https://www.datasciencecentral.com/profiles/blogs/recurrent-neur…73&utm_medium=social&utm_source=twitter.com&utm_campaign=buffer Strona 3 z 6
Recurrent Neural Nets – The Third and Least Appreciated Leg of the AI Stool - Data Science Central 11.04.2018, 21)15
differentiate from the original bard, like this one (part of a much longer
entry).
PANDARUS:
There is a technical challenge for many types of NNs that use common
gradient descent types of fitness functions that is exponentially worse for
RNNs. This is the so-called ‘vanishing gradient’ problem. The reason for
this is that each time step in an RNN is equivalent to an entire layer in a
feedforward NN. So training a RNN for 100 time steps is equivalent to
training a feedforward NN with 100 layers. This leads to exponentially
smaller gradients between layers and a decay of information through time.
Basically this results in early ‘layers’ in the RNN becoming insensitive and
also giving RNNs their reputation for being difficult to train. A number of
solutions based on gating, telling the RNN when to remember and when to
forget can largely resolve the problem. The most common technique is
called LSTM (long short term memory). RNN/LSTM models can also
handle situations where there are long delays (as might occur with IoT
sensors) or where there is a mix of high and low frequency data
components.
https://www.datasciencecentral.com/profiles/blogs/recurrent-neur…73&utm_medium=social&utm_source=twitter.com&utm_campaign=buffer Strona 4 z 6
Recurrent Neural Nets – The Third and Least Appreciated Leg of the AI Stool - Data Science Central 11.04.2018, 21)15
The importance of RNNs and CNNs together is that they are two of three
main legs of the AI stool. CNNs and RNNs are the eyes, ears, and mouth.
SNNs may someday be the brain.
Here’s a brief chart that reflects general agreement of problem type versus
https://www.datasciencecentral.com/profiles/blogs/recurrent-neur…73&utm_medium=social&utm_source=twitter.com&utm_campaign=buffer Strona 5 z 6
Recurrent Neural Nets – The Third and Least Appreciated Leg of the AI Stool - Data Science Central 11.04.2018, 21)15
As you can see, there are more applications for RNNs and they lie mostly in
speech, text, translation, and time series problems.
While ‘deep learning’ takes its name from the large number of layers in
CNNs, it’s RNNs that are the real work horse of this generation of neural
nets.
About the author: Bill Vorhies is Editorial Director for Data Science
Central and has practiced as a data scientist and commercial predictive
modeler since 2001. He can be reached at:
Bill@DataScienceCentral.com
https://www.datasciencecentral.com/profiles/blogs/recurrent-neur…73&utm_medium=social&utm_source=twitter.com&utm_campaign=buffer Strona 6 z 6