Science Score: 44.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
  • Academic publication links
  • Academic email domains
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (15.3%) to scientific vocabulary
Last synced: 6 months ago · JSON representation ·

Repository

Basic Info
  • Host: GitHub
  • Owner: Mahmoud-Da
  • License: other
  • Language: Python
  • Default Branch: master
  • Size: 450 KB
Statistics
  • Stars: 0
  • Watchers: 0
  • Forks: 0
  • Open Issues: 0
  • Releases: 0
Created 8 months ago · Last pushed 8 months ago
Metadata Files
Readme License Citation

README.md

VTON-X: A Modernized Virtual Try-On Project

Note: This project is a customized and modernized fork of the official CP-VTON+ (CVPRW 2020) implementation. All credit for the core research and model architecture belongs to the original authors.

This repository, VTON-X, aims to make the powerful CP-VTON+ model more accessible and easier to set up by leveraging modern Python development tools and providing a streamlined process for testing and inference.

Original Project Page | Original Paper


What's New in VTON-X?

This fork focuses on improving the developer experience and modernizing the project's foundation. Key changes include:

  • Modern Dependency Management: Uses pipenv for robust, reproducible environments, replacing the requirements.txt file.
  • Updated Libraries: Configured to work with newer library versions, including PyTorch >= 1.10 and Torchvision >= 0.11.
  • Simplified Setup: A clear, step-by-step guide to get the demo running in minutes.
  • Focus on Inference: The primary goal of this fork is to make it easy to run the pre-trained models on new images.

Quickstart: Run the Demo

Get the virtual try-on running with just a few commands.

Prerequisites:

  • Python 3.6+
  • Git
  • pipenv (pip install pipenv)
  • An NVIDIA GPU with a compatible CUDA version installed is highly recommended.

1. Clone the Repository

bash git clone [repo-url] cd VTON-X

2. Setup the Environment with Pipenv

TODO

4. Run the Try-On Inference

The pipeline runs in two stages: GMM (warping the cloth) and TOM (generating the final image).

```bash

Stage 1: Run the Geometric Matching Module (GMM)

python test.py --name gmmtest --stage GMM --workers 4 --dataroot ./data --datalist test_pairs.txt --checkpoint checkpoints/gmm.pth

Stage 2: Run the Try-On Module (TOM)

python test.py --name tomtest --stage TOM --workers 4 --dataroot ./data --datalist test_pairs.txt --checkpoint checkpoints/gen.pth ```

5. View Your Results!

The final images are saved in the results/tom_test/test/try-on/ directory.


Advanced Usage (Training)

While this fork focuses on inference, the original training scripts are preserved. To train the models from scratch, please follow the logic from the original repository:

  1. Prepare Data: Download the full VITON_PLUS dataset and prepare it in the data directory as described in the original README.
  2. Train GMM: Run train.py for the GMM stage. bash pipenv shell python train.py --name GMM_train --stage GMM --workers 4 --save_count 5000 --shuffle
  3. Generate Warped Clothes: Use the trained GMM to generate warped clothes for the training set by running test.py.
  4. Train TOM: Run train.py for the TOM stage, using the newly generated warped clothes as input. bash pipenv shell python train.py --name TOM_train --stage TOM --workers 4 --save_count 5000 --shuffle

Using Custom Images

To run the model on your own images, follow these steps:

  1. Prepare Inputs: You need to generate several input files for each try-on pair. See the original authors' notes on this in the section below.
    • image (person image, 256x192)
    • cloth (clothing image, 256x192)
    • image-parse (person segmentation map)
    • cloth-mask (binary mask of the cloth)
    • pose (pose keypoints JSON file)
  2. Organize Files: Place your files in the corresponding subdirectories within the data/ folder.
  3. Update Pair List: Add a new line to data/test_pairs.txt with the filenames of your person image and cloth image (e.g., my_photo.jpg my_shirt.png).
  4. Run Inference: Execute the commands from the Quickstart Step 4.

Notes from the Original Authors on Custom Images:

  • You can generate image-parse with pre-trained networks like CIHP_PGN or Graphonomy.
  • cloth-mask can be generated with simple image processing functions in Pillow or OpenCV.
  • pose keypoints can be generated using the official OpenPose repository (COCO-18 model).

Citation & Acknowledgements

This project would not be possible without the foundational work of the original CP-VTON+ authors. If this code helps your research, please cite their paper:

@InProceedings{Minar_CPP_2020_CVPR_Workshops, title={CP-VTON+: Clothing Shape and Texture Preserving Image-Based Virtual Try-On}, author={Minar, Matiur Rahman and Thai Thanh Tuan and Ahn, Heejune and Rosin, Paul and Lai, Yu-Kun}, booktitle = {The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops}, month = {June}, year = {2020} }

This implementation is also heavily based on the original CP-VTON. We are extremely grateful for their public implementation, which laid the groundwork for this entire line of research.

Owner

  • Name: Mahmoud Dabbagh
  • Login: Mahmoud-Da
  • Kind: user
  • Location: 横浜市

Citation (CITATION.cff)

# This CITATION.cff file was generated with cffinit.
# Visit https://bit.ly/cffinit to generate yours today!

cff-version: 1.2.0
title: >-
  Official implementation for "CP-VTON+: Clothing
  Shape and Texture Preserving Image-Based Virtual
  Try-On" from CVPRW 2020.
message: >-
  If you use this code and dataset, please cite it
  using the metadata from this file. Also, please
  cite the paper separately as well.
type: software
authors:
  - given-names: Matiur Rahman
    family-names: Minar
    email: minar09.bd@gmail.com
    orcid: 'https://orcid.org/0000-0002-3128-2915'
  - given-names: Thanh Tuan
    family-names: Thai
    email: thaithanhtuan1987@gmail.com
    orcid: 'https://orcid.org/0000-0003-2748-0529'
identifiers:
  - type: url
    value: 'https://github.com/minar09/cp-vton-plus'
    description: public URL for the code and dataset
repository-code: 'https://github.com/minar09/cp-vton-plus'
url: 'https://minar09.github.io/cpvtonplus/'
repository: 'https://github.com/minar09/cp-vton-plus'
abstract: >-
  CP-VTON+ (CVPRW 2020)

  Official implementation for "CP-VTON+: Clothing
  Shape and Texture Preserving Image-Based Virtual
  Try-On" from CVPRW 2020.

  Project page:
  https://minar09.github.io/cpvtonplus/.

  Saved/Pre-trained models: Checkpoints

  Dataset: VITON_PLUS

  The code and pre-trained models are tested with
  pytorch 0.4.1, torchvision 0.2.1, opencv-python 4.1
  and pillow 5.4 (Python 3 env).
keywords:
  - Virtual try-on
  - Fashion
license: MIT

GitHub Events

Total
  • Delete event: 1
  • Push event: 3
  • Pull request event: 6
  • Create event: 6
Last Year
  • Delete event: 1
  • Push event: 3
  • Pull request event: 6
  • Create event: 6

Issues and Pull Requests

Last synced: 6 months ago

All Time
  • Total issues: 0
  • Total pull requests: 6
  • Average time to close issues: N/A
  • Average time to close pull requests: less than a minute
  • Total issue authors: 0
  • Total pull request authors: 1
  • Average comments per issue: 0
  • Average comments per pull request: 0.0
  • Merged pull requests: 6
  • Bot issues: 0
  • Bot pull requests: 0
Past Year
  • Issues: 0
  • Pull requests: 6
  • Average time to close issues: N/A
  • Average time to close pull requests: less than a minute
  • Issue authors: 0
  • Pull request authors: 1
  • Average comments per issue: 0
  • Average comments per pull request: 0.0
  • Merged pull requests: 6
  • Bot issues: 0
  • Bot pull requests: 0
Top Authors
Issue Authors
Pull Request Authors
  • Mahmoud-Da (6)
Top Labels
Issue Labels
Pull Request Labels

Dependencies

Pipfile pypi
  • gdown ==5.2.0
  • numpy ==2.3.1
  • opencv-contrib-python ==4.11.0.86
  • pillow ==11.3.0
  • tensorboardx ==2.6.4
  • torch ==2.7.1
  • torchvision ==0.22.1
  • tqdm ==4.67.1
Pipfile.lock pypi
  • beautifulsoup4 ==4.13.4
  • certifi ==2025.6.15
  • charset-normalizer ==3.4.2
  • filelock ==3.18.0
  • fsspec ==2025.5.1
  • gdown ==5.2.0
  • idna ==3.10
  • jinja2 ==3.1.6
  • markupsafe ==3.0.2
  • mpmath ==1.3.0
  • networkx ==3.5
  • numpy ==2.3.1
  • opencv-contrib-python ==4.11.0.86
  • packaging ==25.0
  • pillow ==11.3.0
  • protobuf ==6.31.1
  • pysocks ==1.7.1
  • requests ==2.32.4
  • setuptools ==80.9.0
  • soupsieve ==2.7
  • sympy ==1.14.0
  • tensorboardx ==2.6.4
  • torch ==2.7.1
  • torchvision ==0.22.1
  • tqdm ==4.67.1
  • typing-extensions ==4.14.1
  • urllib3 ==2.5.0
requirements.example.txt pypi
  • numpy *
  • opencv-contrib-python *
  • pillow *
  • tensorboardX *
  • torch >=1.10
  • torchvision >=0.11