irl-maxent
Maximum Entropy and Maximum Causal Entropy Inverse Reinforcement Learning Implementation in Python
Science Score: 44.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
✓CITATION.cff file
Found CITATION.cff file -
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
○Committers with academic emails
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (9.7%) to scientific vocabulary
Keywords
Repository
Maximum Entropy and Maximum Causal Entropy Inverse Reinforcement Learning Implementation in Python
Basic Info
Statistics
- Stars: 293
- Watchers: 4
- Forks: 63
- Open Issues: 3
- Releases: 0
Topics
Metadata Files
README.md
Maximum Entropy Inverse Reinforcement Learning
This is a python implementation of the Maximum Entropy Inverse Reinforcement Learning (MaxEnt IRL) algorithm based on the similarly named paper by Ziebart et al. and the Maximum Causal Entropy Inverse Reinforcement Learning (MaxCausalEnt IRL) algorithm based on his PhD thesis. Project for the Advanced Seminar in Imitation Learning, summer term 2019, University of Stuttgart.
This implementation is available as python package at https://pypi.org/project/irl-maxent/ and can be installed via pip install irl-maxent.
You may also want to have a look at the accompanying presentation.
For an example demonstrating how the Maximum (non-causal) Entropy IRL algorithm works, see the corresponding Jupyter notebook (notebooks/maxent.ipynb).
Note that the provided python files (src/) contain a slightly more optimized implementation of the algorithms.
To run a demonstration without the notebook, you can directly run ./src/example.py.
Also have a look at this file on how to use the provided framework.
The framework contains:
- Two GridWorld implementations for demonstration (irl_maxent.gridworld)
- The algorithm implementations (irl_maxent.maxent)
- A gradient based optimizer framework (irl_maxent.optimizer)
- Plotting helper functions (irl_maxent.plot)
- A MDP solver framework, i.e. value iteration and corresponding utilities (irl_maxent.solver)
- A trajectory/trajectory generation framework (irl_maxent.trajectory)
This project solely relies on the following dependencies: numpy, matplotlib, itertools, and pytest.
Owner
- Name: Maximilian Luz
- Login: qzed
- Kind: user
- Location: Freiburg i. Br., Germany
- Website: https://mxnluz.io/
- Twitter: mxnluz
- Repositories: 36
- Profile: https://github.com/qzed
PhD Student at the Robot Learning Lab, University of Freiburg.
Citation (CITATION.cff)
cff-version: 1.2.0
title: Maximum Entropy Inverse Reinforcement Learning - An Implementation
message: 'If you use this software, please cite it as below.'
type: software
authors:
- family-names: Luz
given-names: Maximilian
orcid: 'https://orcid.org/0000-0002-1123-6604'
email: m@mxnluz.io
url: 'https://github.com/qzed/irl-maxent'
license: MIT
version: 0.1.0
date-released: '2019-07-01'
GitHub Events
Total
- Watch event: 71
- Fork event: 7
Last Year
- Watch event: 71
- Fork event: 7
Committers
Last synced: 9 months ago
Top Committers
| Name | Commits | |
|---|---|---|
| qzed | q****d | 105 |
| Daniel Boateng | 9****J | 1 |
Issues and Pull Requests
Last synced: 7 months ago
All Time
- Total issues: 5
- Total pull requests: 2
- Average time to close issues: 7 months
- Average time to close pull requests: 2 days
- Total issue authors: 5
- Total pull request authors: 2
- Average comments per issue: 2.2
- Average comments per pull request: 0.0
- Merged pull requests: 1
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 0
- Pull requests: 0
- Average time to close issues: N/A
- Average time to close pull requests: N/A
- Issue authors: 0
- Pull request authors: 0
- Average comments per issue: 0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
- ArezooAalipanah (1)
- catubc (1)
- rosewang2008 (1)
- siddhya (1)
- kierad (1)
Pull Request Authors
- herambnemlekar (1)
- NiftyJ (1)
Top Labels
Issue Labels
Pull Request Labels
Packages
- Total packages: 1
-
Total downloads:
- pypi 2,054 last-month
- Total dependent packages: 0
- Total dependent repositories: 1
- Total versions: 1
- Total maintainers: 1
pypi.org: irl-maxent
A small package for Maximum Entropy Inverse Reinforcement Learning on simple MDPs
- Homepage: https://github.com/qzed/irl-maxent
- Documentation: https://irl-maxent.readthedocs.io/
- License: MIT License
-
Latest release: 0.1.0
published almost 4 years ago
Rankings
Maintainers (1)
Dependencies
- matplotlib >=3.1.0
- numpy >=1.18.0
- pytest >=5.3.0
- matplotlib *
- numpy *