mirror of
https://github.com/labmlai/annotated_deep_learning_paper_implementations.git
synced 2025-11-10 23:30:33 +08:00
📚 annotations
This commit is contained in:
@@ -34,7 +34,7 @@ and
|
||||
#### ✨ [Reinforcement Learning](http://lab-ml.com/labml_nn/rl/)
|
||||
* [Proximal Policy Optimization](http://lab-ml.com/labml_nn/rl/ppo/) with
|
||||
[Generalized Advantage Estimation](http://lab-ml.com/labml_nn/rl/ppo/gae.html)
|
||||
* [Deep Q Netowrks](http://lab-ml.com/labml_nn/rl/dqn/) with
|
||||
* [Deep Q Networks](http://lab-ml.com/labml_nn/rl/dqn/) with
|
||||
with [Dueling Network](http://lab-ml.com/labml_nn/rl/dqn/model.html),
|
||||
[Prioritized Replay](http://lab-ml.com/labml_nn/rl/dqn/replay_buffer.html)
|
||||
and Double Q Network.
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
"""
|
||||
# Neural Network Model for Deep Q Network (DQN)
|
||||
# Deep Q Network (DQN) Model
|
||||
"""
|
||||
|
||||
import torch
|
||||
|
||||
@@ -33,7 +33,7 @@ and
|
||||
#### ✨ [Reinforcement Learning](http://lab-ml.com/labml_nn/rl/)
|
||||
* [Proximal Policy Optimization](http://lab-ml.com/labml_nn/rl/ppo/) with
|
||||
[Generalized Advantage Estimation](http://lab-ml.com/labml_nn/rl/ppo/gae.html)
|
||||
* [Deep Q Netowrks](http://lab-ml.com/labml_nn/rl/dqn/) with
|
||||
* [Deep Q Networks](http://lab-ml.com/labml_nn/rl/dqn/) with
|
||||
with [Dueling Network](http://lab-ml.com/labml_nn/rl/dqn/model.html),
|
||||
[Prioritized Replay](http://lab-ml.com/labml_nn/rl/dqn/replay_buffer.html)
|
||||
and Double Q Network.
|
||||
|
||||
Reference in New Issue
Block a user