diff --git a/README.md b/README.md index a1dbebf..7e7b31f 100644 --- a/README.md +++ b/README.md @@ -8,14 +8,14 @@ - bilibili:[李宏毅《深度强化学习》](https://www.bilibili.com/video/BV1MW411w79n) ## 目录 -- [P1 Policy Gradient](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter1/chapter1) -- [P2 Proximal Policy Optimization (PPO)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter2/chapter2) -- [P3 Q-learning (Basic Idea)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter3/chapter3) -- [P4 Q-learning (Advanced Tips)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter4/chapter4) -- [P5 Q-learning (Continuous Action)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter5/chapter5) -- [P6 Actor-Critic](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter6/chapter6) -- [P7 Sparse Reward](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter7/chapter7) -- [P8 Imitation Learning](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter8/chapter8) +- [P1 Policy Gradient](https://datawhalechina.github.io/leedeeprl-notes/#/chapter1/chapter1) +- [P2 Proximal Policy Optimization (PPO)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter2/chapter2) +- [P3 Q-learning (Basic Idea)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter3/chapter3) +- [P4 Q-learning (Advanced Tips)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter4/chapter4) +- [P5 Q-learning (Continuous Action)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter5/chapter5) +- [P6 Actor-Critic](https://datawhalechina.github.io/leedeeprl-notes/#/chapter6/chapter6) +- [P7 Sparse Reward](https://datawhalechina.github.io/leedeeprl-notes/#/chapter7/chapter7) +- [P8 Imitation Learning](https://datawhalechina.github.io/leedeeprl-notes/#/chapter8/chapter8) ## 主要贡献者 - [@qiwang067](https://github.com/qiwang067) diff --git a/docs/README.md b/docs/README.md index 1dde697..a6bbb82 100644 --- a/docs/README.md +++ b/docs/README.md @@ -1,19 +1,19 @@ # 李宏毅深度强化学习笔记(LeeDeepRL-Notes) ## 笔记在线阅读地址 -在线阅读地址:https://datawhalechina.github.io/LeeDeepRL-Notes +在线阅读地址:https://datawhalechina.github.io/leedeeprl-notes/ ## 课程在线观看地址 - bilibili:[李宏毅《深度强化学习》](https://www.bilibili.com/video/BV1MW411w79n) ## 目录 -- [P1 Policy Gradient](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter1/chapter1) -- [P2 Proximal Policy Optimization (PPO)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter2/chapter2) -- [P3 Q-learning (Basic Idea)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter3/chapter3) -- [P4 Q-learning (Advanced Tips)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter4/chapter4) -- [P5 Q-learning (Continuous Action)](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter5/chapter5) -- [P6 Actor-Critic](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter6/chapter6) -- [P7 Sparse Reward](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter7/chapter7) -- [P8 Imitation Learning](https://datawhalechina.github.io/LeeDeepRL-Notes/#/chapter8/chapter8) +- [P1 Policy Gradient](https://datawhalechina.github.io/leedeeprl-notes/#/chapter1/chapter1) +- [P2 Proximal Policy Optimization (PPO)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter2/chapter2) +- [P3 Q-learning (Basic Idea)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter3/chapter3) +- [P4 Q-learning (Advanced Tips)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter4/chapter4) +- [P5 Q-learning (Continuous Action)](https://datawhalechina.github.io/leedeeprl-notes/#/chapter5/chapter5) +- [P6 Actor-Critic](https://datawhalechina.github.io/leedeeprl-notes/#/chapter6/chapter6) +- [P7 Sparse Reward](https://datawhalechina.github.io/leedeeprl-notes/#/chapter7/chapter7) +- [P8 Imitation Learning](https://datawhalechina.github.io/leedeeprl-notes/#/chapter8/chapter8) ## 主要贡献者 - [@qiwang067](https://github.com/qiwang067)