Sarcia - Judd Michael - AS4
Sarcia - Judd Michael - AS4
Layers of network nodes that process input and forward output make up
artificial neural networks (ANNs), which are built to mimic the structure and
functions of the human brain. Similar to feed-forward neural networks, they can
process information in two directions. However, they loop information back into
the network via feedback loops like backpropagation through time. Sequence-to-
sequence models can benefit from truncated backpropagation using time neural
networks, which restricts the number of time steps in the input sequence.
1
3. Advantages and disadvantages
Sequence modelling tasks including speech recognition, time series analysis,
and natural language processing are well suited for RNNs. They can recognize
connections and relationships within sequences because they can recall
contextual information from earlier inputs. Furthermore, RNNs can handle
variable-length sequences and share parameters, which reduces their memory
and processing needs.
Among the many drawbacks of RNNs are the disappearing and exploding
gradient issues that arise during training. In addition to memory capacity
limitations, bias towards recent data, and challenges with parallelization, these
issues make it more difficult for the network to grasp global context and long-
range dependencies. As a result of sequential data processing, RNNs are longer
to train and have difficulty retaining knowledge over many time steps.
4. Examples of use
Some examples of recurrent neural networks use in the industry include:
2
processes natural language data, mostly using machine learning language
models.
Deep Q-Networks
1. Definition
The first stage in developing a Q-matrix to optimize long-term rewards for an
agent is called Q-Learning. Larger systems with more states and actions,
however, make this approach unfeasible. To solve this problem, Deep Q-
Learning approximates values while preserving their relative importance using a
neural network. By using this method, the neural network receives the initial state
as input and outputs the Q-values for every action that could be taken.
3
trained to become more adept at making decisions. To stabilize the training
process, the target network is updated on a regular basis to match the
parameters of the main neural network. Iteratively repeating the process enables
the agent to become more adept at making decisions. By training the agent to
predict optimal Q-values for every state-action pair, reinforcement learning
improves performance in the given environment and yields bigger rewards.
4. Examples of use
Some examples of recurrent neural networks use in the industry include
4
2. Autonomous vehicles – With deep Q-learning, autonomous car performance,
safety, and efficiency can all be greatly enhanced by teaching agents to make
decisions in real time based on real-time information.
3. Other uses – Beyond gaming and driverless cars, deep Q-learning has
potential uses in robotics, finance, healthcare, and recommendation systems
for intelligent decision-making across a range of industries.
Reference:
advantages-and-disadvantages-of-a-recurrent-neural-network-rnn/
https://www.geeksforgeeks.org/deep-q-learning/
https://www.techtarget.com/searchenterpriseai/definition/recurrent-neural-
networks#:~:text=Recurrent%20neural%20networks%20recognize
%20data's,activity%20in%20the%20human%20brain.
Bento, C. (2022, May 31). Recurrent Neural Networks Explained with a Real Life
recurrent-neural-networks-explained-with-a-real-life-example-and-python-
code-e8403a45f5de
5
Kvartalnyi, N. (2023, November 24). Deep Q-Learning Explained:Step-by-Step
comprehensive-guide/