Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 2 additions & 0 deletions units/en/unit3/deep-q-network.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -39,3 +39,5 @@ Finally, we have a couple of fully connected layers that output a Q-value for ea
<img src="https://huggingface.co/datasets/huggingface-deep-rl-course/course-images/resolve/main/en/unit4/deep-q-network.jpg" alt="Deep Q Network"/>

So, we see that Deep Q-Learning uses a neural network to approximate, given a state, the different Q-values for each possible action at that state. Now let's study the Deep Q-Learning algorithm.

NOTE: We are using same network to calculate predicting (Q_current) and targeting (Q_target). This will cause some issues of Unstable training since the target is moving every time network’s parameters is updated.