bsk_rl

RL environments and tools for spacecraft autonomy research, built on Basilisk. Developed by the AVS Lab.

https://github.com/avslab/bsk_rl

Science Score: 54.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
  • Academic publication links
  • Committers with academic emails
    1 of 5 committers (20.0%) from academic institutions
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (7.1%) to scientific vocabulary

Keywords

autonomy earth-observation reinforcement-learning satellite-tasking
Last synced: 6 months ago · JSON representation ·

Repository

RL environments and tools for spacecraft autonomy research, built on Basilisk. Developed by the AVS Lab.

Basic Info
Statistics
  • Stars: 73
  • Watchers: 6
  • Forks: 3
  • Open Issues: 31
  • Releases: 4
Topics
autonomy earth-observation reinforcement-learning satellite-tasking
Created over 2 years ago · Last pushed 6 months ago
Metadata Files
Readme Contributing License Citation

README.md

BSK-RL: Environments for Spacecraft Planning and Scheduling

BSK-RL (Basilisk + Reinforcement Learning) is a Python package for constructing Gymnasium environments for spacecraft tasking problems. It is built on top of Basilisk, a modular and fast spacecraft simulation framework, making the simulation environments high-fidelity and computationally efficient.

Usage

Installation instructions, examples, and documentation can be found on the BSK-RL website.

Acknowledgment

BSK-RL is developed by the Autonomous Vehicle Systems (AVS) Lab at the University of Colorado Boulder.

Owner

  • Name: AVSLab
  • Login: AVSLab
  • Kind: organization
  • Location: United States of America

Citation (CITATION.cff)

cff-version: 1.2.0
title: >-
  BSK-RL: Modular, High-Fidelity Reinforcement Learning
  Environments for Spacecraft Tasking
message: >-
  If you use this software, please cite it using the
  metadata from this file.
type: software
authors:
  - given-names: Mark
    family-names: Stephenson
    email: Mark.A.Stephenson@colorado.edu
    affiliation: 'University of Colorado, Boulder'
    orcid: 'https://orcid.org/0009-0004-3438-8127'
  - given-names: Hanspeter
    family-names: Schaub
    orcid: 'https://orcid.org/0000-0003-0002-6035'
    affiliation: 'University of Colorado, Boulder'
    email: Hanspeter.Schaub@colorado.edu
identifiers:
  - type: url
    value: 'https://hanspeterschaub.info/Papers/Stephenson2024c.pdf'
repository-code: 'https://github.com/AVSLab/bsk_rl/'
url: 'https://avslab.github.io/bsk_rl/'
abstract: >-
  Reinforcement learning (RL) is a highly adaptable
  framework for generating autonomous agents across a wide
  domain of problems. While RL has been successfully applied
  to highly complex, real-world systems, a significant
  amount of the literature studies abstractions and
  idealized versions of problems. This is especially the
  case for the field of spacecraft tasking, in which even
  traditional preplanning approaches tend to use highly
  simplified models of spacecraft dynamics and operations.
  When simplified methods are tested in a full-fidelity
  simulation, they often lead to conservative solutions that
  are suboptimal or aggressive solutions that are
  infeasible. As a result, there is a need for a
  high-fidelity spacecraft simulation environment to
  evaluate RL-based and other tasking algorithms. This paper
  introduces BSK-RL, an open-source Python package for
  creating and customizing reinforcement learning
  environments for spacecraft tasking problems. It combines
  Basilisk --- a high-speed and high-fidelity spacecraft
  simulation framework --- with abstractions of satellite
  tasks and operational objectives within the standard
  Gymnasium API wrapper for RL environments. The package is
  designed to meet the needs of RL and spacecraft operations
  researchers: Environment parameters are easily
  reproducible, customizable, and randomizable. Environments
  are highly modular: satellite state and action spaces can
  be specified, mission objectives and rewards can be
  defined, and the satellite dynamics and flight software
  can be configured, implicitly introducing operational
  limitations and safety constraints. Heterogeneous
  multi-agent environments can be created for more complex
  mission scenarios that consider communication and
  collaboration. Training and deployment using the package
  are demonstrated for an Earth-observing satellite with
  resource constraints.
license: MIT
version: 1.0.1
date-released: '2024-08-27'

GitHub Events

Total
  • Create event: 45
  • Release event: 2
  • Issues event: 60
  • Watch event: 31
  • Delete event: 39
  • Issue comment event: 34
  • Push event: 331
  • Pull request review comment event: 86
  • Pull request review event: 118
  • Pull request event: 77
Last Year
  • Create event: 45
  • Release event: 2
  • Issues event: 60
  • Watch event: 31
  • Delete event: 39
  • Issue comment event: 34
  • Push event: 331
  • Pull request review comment event: 86
  • Pull request review event: 118
  • Pull request event: 77

Committers

Last synced: about 2 years ago

All Time
  • Total Commits: 81
  • Total Committers: 5
  • Avg Commits per committer: 16.2
  • Development Distribution Score (DDS): 0.123
Past Year
  • Commits: 81
  • Committers: 5
  • Avg Commits per committer: 16.2
  • Development Distribution Score (DDS): 0.123
Top Committers
Name Email Commits
Mark Stephenson m****n@g****m 71
LorenzzoQM l****i@c****u 7
Adam Herrmann a****e@g****m 1
Hanspeter Schaub 8****h 1
johnowagon 9****n 1
Committer Domains (Top 20 + Academic)

Issues and Pull Requests

Last synced: 6 months ago

All Time
  • Total issues: 97
  • Total pull requests: 94
  • Average time to close issues: about 1 month
  • Average time to close pull requests: 8 days
  • Total issue authors: 10
  • Total pull request authors: 4
  • Average comments per issue: 0.51
  • Average comments per pull request: 0.4
  • Merged pull requests: 79
  • Bot issues: 0
  • Bot pull requests: 0
Past Year
  • Issues: 35
  • Pull requests: 44
  • Average time to close issues: about 1 month
  • Average time to close pull requests: 15 days
  • Issue authors: 8
  • Pull request authors: 2
  • Average comments per issue: 0.37
  • Average comments per pull request: 0.2
  • Merged pull requests: 34
  • Bot issues: 0
  • Bot pull requests: 0
Top Authors
Issue Authors
  • Mark2000 (88)
  • LorenzzoQM (19)
  • ai-in-aerospace (5)
  • RainGameon (2)
  • xiaoJ136 (1)
  • MingJunF (1)
  • gautam1858 (1)
  • mvinovivek (1)
  • arya-pathak (1)
  • Yume27 (1)
  • Rishoban (1)
  • nmharmon8 (1)
Pull Request Authors
  • Mark2000 (114)
  • LorenzzoQM (15)
  • johnowagon (4)
  • nmharmon8 (2)
  • hhlei (1)
  • ai-in-aerospace (1)
Top Labels
Issue Labels
enhancement (59) bug (45) triage (28) infrastructure (14) modeling (11) refactor (9) documentation (5) question (2) good first issue (2) wontfix (1)
Pull Request Labels
enhancement (4) bug (3) triage (1)

Dependencies

.github/workflows/commit_checks.yml actions
  • actions/checkout v3 composite
  • chartboost/ruff-action v1 composite
  • gsactions/commit-message-checker v2 composite
  • isort/isort-action v1 composite
  • psf/black stable composite
pyproject.toml pypi
  • Deprecated *
  • deap ==1.3.3
  • gymnasium *
  • matplotlib *
  • numpy *
  • pandas *
  • pytest *
  • pytest-cov *
  • pytest-repeat *
  • requests *
  • scikit-learn *
  • scipy *
  • stable-baselines3 *
  • tensorflow *
  • torch *
.github/workflows/documentation.yml actions
  • actions/checkout v3 composite
  • actions/setup-python v3 composite
  • peaceiris/actions-gh-pages v3 composite