Science Score: 44.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
✓CITATION.cff file
Found CITATION.cff file -
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (15.3%) to scientific vocabulary
Repository
Basic Info
- Host: GitHub
- Owner: Mahmoud-Da
- License: other
- Language: Python
- Default Branch: master
- Size: 450 KB
Statistics
- Stars: 0
- Watchers: 0
- Forks: 0
- Open Issues: 0
- Releases: 0
Metadata Files
README.md
VTON-X: A Modernized Virtual Try-On Project
Note: This project is a customized and modernized fork of the official CP-VTON+ (CVPRW 2020) implementation. All credit for the core research and model architecture belongs to the original authors.
This repository, VTON-X, aims to make the powerful CP-VTON+ model more accessible and easier to set up by leveraging modern Python development tools and providing a streamlined process for testing and inference.
Original Project Page | Original Paper
What's New in VTON-X?
This fork focuses on improving the developer experience and modernizing the project's foundation. Key changes include:
- Modern Dependency Management: Uses
pipenvfor robust, reproducible environments, replacing therequirements.txtfile. - Updated Libraries: Configured to work with newer library versions, including
PyTorch >= 1.10andTorchvision >= 0.11. - Simplified Setup: A clear, step-by-step guide to get the demo running in minutes.
- Focus on Inference: The primary goal of this fork is to make it easy to run the pre-trained models on new images.
Quickstart: Run the Demo
Get the virtual try-on running with just a few commands.
Prerequisites:
- Python 3.6+
- Git
pipenv(pip install pipenv)- An NVIDIA GPU with a compatible CUDA version installed is highly recommended.
1. Clone the Repository
bash
git clone [repo-url]
cd VTON-X
2. Setup the Environment with Pipenv
TODO
4. Run the Try-On Inference
The pipeline runs in two stages: GMM (warping the cloth) and TOM (generating the final image).
```bash
Stage 1: Run the Geometric Matching Module (GMM)
python test.py --name gmmtest --stage GMM --workers 4 --dataroot ./data --datalist test_pairs.txt --checkpoint checkpoints/gmm.pth
Stage 2: Run the Try-On Module (TOM)
python test.py --name tomtest --stage TOM --workers 4 --dataroot ./data --datalist test_pairs.txt --checkpoint checkpoints/gen.pth ```
5. View Your Results!
The final images are saved in the results/tom_test/test/try-on/ directory.
Advanced Usage (Training)
While this fork focuses on inference, the original training scripts are preserved. To train the models from scratch, please follow the logic from the original repository:
- Prepare Data: Download the full
VITON_PLUSdataset and prepare it in thedatadirectory as described in the originalREADME. - Train GMM: Run
train.pyfor the GMM stage.bash pipenv shell python train.py --name GMM_train --stage GMM --workers 4 --save_count 5000 --shuffle - Generate Warped Clothes: Use the trained GMM to generate warped clothes for the training set by running
test.py. - Train TOM: Run
train.pyfor the TOM stage, using the newly generated warped clothes as input.bash pipenv shell python train.py --name TOM_train --stage TOM --workers 4 --save_count 5000 --shuffle
Using Custom Images
To run the model on your own images, follow these steps:
- Prepare Inputs: You need to generate several input files for each try-on pair. See the original authors' notes on this in the section below.
image(person image, 256x192)cloth(clothing image, 256x192)image-parse(person segmentation map)cloth-mask(binary mask of the cloth)pose(pose keypoints JSON file)
- Organize Files: Place your files in the corresponding subdirectories within the
data/folder. - Update Pair List: Add a new line to
data/test_pairs.txtwith the filenames of your person image and cloth image (e.g.,my_photo.jpg my_shirt.png). - Run Inference: Execute the commands from the Quickstart Step 4.
Notes from the Original Authors on Custom Images:
- You can generate
image-parsewith pre-trained networks like CIHP_PGN or Graphonomy.cloth-maskcan be generated with simple image processing functions in Pillow or OpenCV.posekeypoints can be generated using the official OpenPose repository (COCO-18 model).
Citation & Acknowledgements
This project would not be possible without the foundational work of the original CP-VTON+ authors. If this code helps your research, please cite their paper:
@InProceedings{Minar_CPP_2020_CVPR_Workshops,
title={CP-VTON+: Clothing Shape and Texture Preserving Image-Based Virtual Try-On},
author={Minar, Matiur Rahman and Thai Thanh Tuan and Ahn, Heejune and Rosin, Paul and Lai, Yu-Kun},
booktitle = {The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2020}
}
This implementation is also heavily based on the original CP-VTON. We are extremely grateful for their public implementation, which laid the groundwork for this entire line of research.
Owner
- Name: Mahmoud Dabbagh
- Login: Mahmoud-Da
- Kind: user
- Location: 横浜市
- Repositories: 2
- Profile: https://github.com/Mahmoud-Da
Citation (CITATION.cff)
# This CITATION.cff file was generated with cffinit.
# Visit https://bit.ly/cffinit to generate yours today!
cff-version: 1.2.0
title: >-
Official implementation for "CP-VTON+: Clothing
Shape and Texture Preserving Image-Based Virtual
Try-On" from CVPRW 2020.
message: >-
If you use this code and dataset, please cite it
using the metadata from this file. Also, please
cite the paper separately as well.
type: software
authors:
- given-names: Matiur Rahman
family-names: Minar
email: minar09.bd@gmail.com
orcid: 'https://orcid.org/0000-0002-3128-2915'
- given-names: Thanh Tuan
family-names: Thai
email: thaithanhtuan1987@gmail.com
orcid: 'https://orcid.org/0000-0003-2748-0529'
identifiers:
- type: url
value: 'https://github.com/minar09/cp-vton-plus'
description: public URL for the code and dataset
repository-code: 'https://github.com/minar09/cp-vton-plus'
url: 'https://minar09.github.io/cpvtonplus/'
repository: 'https://github.com/minar09/cp-vton-plus'
abstract: >-
CP-VTON+ (CVPRW 2020)
Official implementation for "CP-VTON+: Clothing
Shape and Texture Preserving Image-Based Virtual
Try-On" from CVPRW 2020.
Project page:
https://minar09.github.io/cpvtonplus/.
Saved/Pre-trained models: Checkpoints
Dataset: VITON_PLUS
The code and pre-trained models are tested with
pytorch 0.4.1, torchvision 0.2.1, opencv-python 4.1
and pillow 5.4 (Python 3 env).
keywords:
- Virtual try-on
- Fashion
license: MIT
GitHub Events
Total
- Delete event: 1
- Push event: 3
- Pull request event: 6
- Create event: 6
Last Year
- Delete event: 1
- Push event: 3
- Pull request event: 6
- Create event: 6
Issues and Pull Requests
Last synced: 6 months ago
All Time
- Total issues: 0
- Total pull requests: 6
- Average time to close issues: N/A
- Average time to close pull requests: less than a minute
- Total issue authors: 0
- Total pull request authors: 1
- Average comments per issue: 0
- Average comments per pull request: 0.0
- Merged pull requests: 6
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 0
- Pull requests: 6
- Average time to close issues: N/A
- Average time to close pull requests: less than a minute
- Issue authors: 0
- Pull request authors: 1
- Average comments per issue: 0
- Average comments per pull request: 0.0
- Merged pull requests: 6
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
Pull Request Authors
- Mahmoud-Da (6)
Top Labels
Issue Labels
Pull Request Labels
Dependencies
- gdown ==5.2.0
- numpy ==2.3.1
- opencv-contrib-python ==4.11.0.86
- pillow ==11.3.0
- tensorboardx ==2.6.4
- torch ==2.7.1
- torchvision ==0.22.1
- tqdm ==4.67.1
- beautifulsoup4 ==4.13.4
- certifi ==2025.6.15
- charset-normalizer ==3.4.2
- filelock ==3.18.0
- fsspec ==2025.5.1
- gdown ==5.2.0
- idna ==3.10
- jinja2 ==3.1.6
- markupsafe ==3.0.2
- mpmath ==1.3.0
- networkx ==3.5
- numpy ==2.3.1
- opencv-contrib-python ==4.11.0.86
- packaging ==25.0
- pillow ==11.3.0
- protobuf ==6.31.1
- pysocks ==1.7.1
- requests ==2.32.4
- setuptools ==80.9.0
- soupsieve ==2.7
- sympy ==1.14.0
- tensorboardx ==2.6.4
- torch ==2.7.1
- torchvision ==0.22.1
- tqdm ==4.67.1
- typing-extensions ==4.14.1
- urllib3 ==2.5.0
- numpy *
- opencv-contrib-python *
- pillow *
- tensorboardX *
- torch >=1.10
- torchvision >=0.11