reinforcement_learning_course_materials
Lecture notes, tutorial tasks including solutions as well as online videos for the reinforcement learning course hosted by Paderborn University
https://github.com/upb-lea/reinforcement_learning_course_materials
Science Score: 36.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
✓Committers with academic emails
3 of 11 committers (27.3%) from academic institutions -
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (8.3%) to scientific vocabulary
Keywords
Keywords from Contributors
Repository
Lecture notes, tutorial tasks including solutions as well as online videos for the reinforcement learning course hosted by Paderborn University
Basic Info
Statistics
- Stars: 997
- Watchers: 27
- Forks: 224
- Open Issues: 0
- Releases: 3
Topics
Metadata Files
README.md
Reinforcement learning course
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Lecture notes, tutorial tasks including solutions as well as online videos for a reinforcement learning course originally hosted at Paderborn University and transferred to University of Siegen. Source code for the entire course material is open and everyone is cordially invited to use it for self-learning (students) or to set up your own course (lecturers).
Lecture slides (click on preview picture)
- Introduction to reinforcement learning
- Markov decision processes
- Dynamic programming
- Monte Carlo methods
- Temporal-difference learning
- Multi-step bootstrapping
- Planning and learning with tabular methods
- Function approximation with supervised learning
- On-policy prediction with function approximation
- Value-based control with function approximation
- Stochastic policy gradient methods
- Deterministic policy gradient methods
- Further contemporary RL algorithms (TRPO, PPO)
Outlook and Research Insights
Summary of part one: reinforcement learning in finite state and action spaces
Summary of part two: reinforcement learning in continuous state and action spaces
Exercise content
All exercises are based on Python 3.12 and site-packages according to the requirements.txt: ```
pip install -r requirements.txt ```
- Basics of Python for scientific computing
- Tutorial video (only 2022 edition available due to technical outage)
- Tutorial template
- Tutorial solution
- Manually solving basic Markov chain, reward and decision problems
- The beer-bachelor and dynamic programming (the shortest beer problem)
- Drive through the race track with Monte Carlo learning
- Drive even faster using temporal-difference learning
- Stabilizing the inverted pendulum by tabular multi-step methods
- Boosting the inverted pendulum by integrating learning & planning (Dyna framework)
- Predicting the operating behavior of a real electric drive systems with supervised learning
- Evaluate the performance of given agents in the mountain car problem using function approximation
- Escape from the mountain car valley using semi-gradient SARSA & least square policy iteration
- Landing on the moon with REINFORCE and actor-critic methods
- Shoot for the moon with DDPG & PPO
Contributions
We highly appreciate any feedback and input to the course material e.g. * typos or content-related discussions (please raise an issue) * adding new contents (please provide a pull request)
If you like to contribute to the repo to a larger extent, please do not hesitate to contact us directly.
Credits
The lecture notes are inspired by * Richard S. Sutton, Andrew G. Barto, 'Reinforcement Learning: An Introduction' Second Edition MIT Press, Cambridge, MA, 2018 * David Silver, UCL Course on Reinforcement Learning, 2015
The tutorials are partly using pre-packed environments from * Gymnasium (maintained branch of OpenAI's Gym)
Owner
- Name: Paderborn University - LEA
- Login: upb-lea
- Kind: organization
- Location: Paderborn, Germany
- Website: https://ei.uni-paderborn.de/en/lea/
- Repositories: 29
- Profile: https://github.com/upb-lea
Department of power electronics and electrical drives
GitHub Events
Total
- Release event: 1
- Watch event: 63
- Push event: 3
- Pull request event: 3
- Fork event: 13
- Create event: 1
Last Year
- Release event: 1
- Watch event: 63
- Push event: 3
- Pull request event: 3
- Fork event: 13
- Create event: 1
Committers
Last synced: 9 months ago
Top Committers
| Name | Commits | |
|---|---|---|
| Wilhelm Kirchgässner | 1****n | 120 |
| Oliver Wallscheid | w****d@l****e | 96 |
| XyDrKRulof | d****t@w****e | 86 |
| Maximilian Schenke | 6****e | 24 |
| Daniel Weber | w****r@l****e | 21 |
| Marvin Meyer | 2****r | 13 |
| webbah | d****o@g****e | 10 |
| bhk11 | 9****1 | 9 |
| Hendrik Vater | 8****r | 4 |
| Darius Jakobeit | j****t@a****e | 3 |
| Maximilian Schenke | m****e | 1 |
Committer Domains (Top 20 + Academic)
Issues and Pull Requests
Last synced: 9 months ago
All Time
- Total issues: 15
- Total pull requests: 6
- Average time to close issues: 3 months
- Average time to close pull requests: about 4 hours
- Total issue authors: 8
- Total pull request authors: 5
- Average comments per issue: 1.8
- Average comments per pull request: 0.17
- Merged pull requests: 5
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 0
- Pull requests: 1
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Issue authors: 0
- Pull request authors: 1
- Average comments per issue: 0
- Average comments per pull request: 0.0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
- eatam23 (6)
- VikasChidananda (2)
- wallscheid (2)
- stwerner97 (1)
- MarvinMeyer (1)
- wkirgsn (1)
- max-schenke (1)
- XyDrKRulof (1)
Pull Request Authors
- AliAbdelwanis (2)
- hvater (2)
- wallscheid (1)
- wkirgsn (1)
- XyDrKRulof (1)
- Webbah (1)
Top Labels
Issue Labels
Pull Request Labels
Dependencies
- gym *
- gymnasium ==0.28.1
- jupyter ==1.0.0
- matplotlib ==3.7.1
- numpy ==1.23.5
- openpyxl *
- pandas ==1.4.2
- pygame ==2.4.0
- pyglet ==1.5.27
- scikit-learn ==1.1.2
- scipy ==1.8.0
- seaborn ==0.11.2
- stable-baselines3 *
- torch *
- tqdm ==4.65.0
- actions/checkout v4 composite
- actions/deploy-pages v4 composite
- actions/upload-pages-artifact v3 composite
- xu-cheng/texlive-action v2 composite

