https://github.com/aehrc/cxrmate
CXRMate: Longitudinal Data and a Semantic Similarity Reward for Chest X-Ray Report Generation
Science Score: 49.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
○CITATION.cff file
-
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
✓DOI references
Found 4 DOI reference(s) in README -
✓Academic publication links
Links to: arxiv.org, sciencedirect.com -
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (8.1%) to scientific vocabulary
Keywords
Repository
CXRMate: Longitudinal Data and a Semantic Similarity Reward for Chest X-Ray Report Generation
Basic Info
- Host: GitHub
- Owner: aehrc
- License: apache-2.0
- Language: Python
- Default Branch: main
- Homepage: https://huggingface.co/aehrc/cxrmate
- Size: 4.03 MB
Statistics
- Stars: 15
- Watchers: 9
- Forks: 3
- Open Issues: 14
- Releases: 0
Topics
Metadata Files
README.md
CXRMate: Leveraging Longitudinal Data and a Semantic Similarity Reward for Chest X-Ray Report Generation
Paper: https://doi.org/10.1016/j.imu.2024.101585, arXiv: https://arxiv.org/abs/2307.09758
@article{NICOLSON2024101585,
title = {Longitudinal data and a semantic similarity reward for chest X-ray report generation},
journal = {Informatics in Medicine Unlocked},
volume = {50},
pages = {101585},
year = {2024},
issn = {2352-9148},
doi = {https://doi.org/10.1016/j.imu.2024.101585},
url = {https://www.sciencedirect.com/science/article/pii/S2352914824001424},
author = {Aaron Nicolson and Jason Dowling and Douglas Anderson and Bevan Koopman},
}
CXRMate is a longitudinal, multi-image CXR report generation encoder-to-decoder model that conditions the report generation process on the report from the previous patient's study if available. The CXRMate checkpoint trained on MIMIC-CXR is available on the Hugging Face Hub: https://huggingface.co/aehrc/cxrmate.
|
|
|----|
|
Generated reports:
Generated reports for the single-image, multi-image, and longitudinal, multi-image CXR generators (both prompted with the radiologist and the generated reports) are located in the generated_reports directory.
Hugging Face models:
- Longitudinal, multi-image CXR report generation with SCST & CXR-BERT reward and generated previous reports: https://huggingface.co/aehrc/cxrmate
Longitudinal, multi-image CXR report generation with SCST & CXR-BERT reward and radiologist previous reports: https://huggingface.co/aehrc/cxrmate-tf
Longitudinal, multi-image CXR report generation with TF: https://huggingface.co/aehrc/cxrmate-tf
Multi-image CXR report generation with TF: https://huggingface.co/aehrc/cxrmate-multi-tf
Single-image CXR report generation with TF: https://huggingface.co/aehrc/cxrmate-single-tf
SCST: Self-Critical Sequence Training, TF: Teacher Forcing
Notebook examples:
Notebook examples for the models can be found in the examples directory.
Dataset:
- The MIMIC-CXR-JPG dataset is available at:
https://physionet.org/content/mimic-cxr-jpg/2.0.0/
Installation:
After cloning the repository, install the required packages in a virtual environment.
The required packages are located in requirements.txt:
shell script
python -m venv --system-site-packages venv
source venv/bin/activate
python -m pip install --upgrade pip
python -m pip install --upgrade -r requirements.txt --no-cache-dir
Test the Hugging Face checkpoints:
The model configurations for each task can be found in its config directory, e.g. config/test_huggingface_longitudinal_gen_prompt_cxr-bert.yaml. To run testing:
shell
dlhpcstarter -t cxrmate_hf -c config/test_huggingface/longitudinal_gen_prompt_cxr-bert.yaml --stages_module tools.stages --test
See dlhpcstarter==0.1.4 for more options.
Note:
- Data will be saved in the experiment directory (exp_dir in the configuration file).
- See https://github.com/MIT-LCP/mimic-cxr/tree/master/txt to extract the sections from the reports.
Training:
To train with teacher forcing:
dlhpcstarter -t cxrmate -c config/train/longitudinal_gt_prompt_tf.yaml --stages_module tools.stages --train
The model can then be tested with the --test flag:
dlhpcstarter -t cxrmate -c config/train/longitudinal_gt_prompt_tf.yaml --stages_module tools.stages --test
To then train with Self-Critical Sequence Training (SCST) with the CXR-BERT reward:
- Copy the path to the checkpoint from the
exp_dirfor the configuration above, then paste it in the configuration for SCST aswarm_start_ckpt_path, then: -
dlhpcstarter -t mimic_cxr -c config/train/longitudinal_gen_prompt_cxr-bert.yaml --stages_module tools.stages --train
Note:
- See dlhpcstarter==0.1.4 for more options.
- See https://github.com/MIT-LCP/mimic-cxr/tree/master/txt to extract the sections from the reports.
Help/Issues:
If you need help, or if there are any issues, please leave an issue and we will get back to you as soon as possible.
Owner
- Name: The Australian e-Health Research Centre
- Login: aehrc
- Kind: organization
- Website: https://aehrc.com
- Twitter: ehealthresearch
- Repositories: 101
- Profile: https://github.com/aehrc
The Australian e-Health Research Centre (AEHRC) is CSIRO’s digital health research program.
GitHub Events
Total
- Issues event: 3
- Watch event: 2
- Issue comment event: 3
- Push event: 2
Last Year
- Issues event: 3
- Watch event: 2
- Issue comment event: 3
- Push event: 2
Issues and Pull Requests
Last synced: 6 months ago
All Time
- Total issues: 19
- Total pull requests: 0
- Average time to close issues: about 1 month
- Average time to close pull requests: N/A
- Total issue authors: 5
- Total pull request authors: 0
- Average comments per issue: 3.89
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 5
- Pull requests: 0
- Average time to close issues: about 1 month
- Average time to close pull requests: N/A
- Issue authors: 2
- Pull request authors: 0
- Average comments per issue: 3.0
- Average comments per pull request: 0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
- yihp (9)
- AceMcAwesome77 (6)
- oscarloch (3)
- passby111 (1)
- Liqq1 (1)
Pull Request Authors
Top Labels
Issue Labels
Pull Request Labels
Dependencies
- bert-score ==0.3.12
- dlhpcstarter ==0.1.4
- numpy ==1.23.5
- peft ==0.3.0
- pycocoevalcap ==1.2
- pycocotools ==2.0.4
- sacremoses ==0.0.53
- stanza ==1.4.2