vizint
Visual intelligence : machines and minds cs-503 project
Science Score: 44.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
✓CITATION.cff file
Found CITATION.cff file -
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (8.3%) to scientific vocabulary
Repository
Visual intelligence : machines and minds cs-503 project
Basic Info
- Host: GitHub
- Owner: DavidLacour
- Language: Python
- Default Branch: main
- Size: 20.6 MB
Statistics
- Stars: 0
- Watchers: 1
- Forks: 0
- Open Issues: 0
- Releases: 0
Metadata Files
README.md
Vision Models Robustness Research Framework
This repository contains a comprehensive framework for training and evaluating robust vision models on CIFAR-10 and TinyImageNet datasets. The framework includes various model architectures, robustness techniques, and visualization tools.
The defauls commands shoud create folders in the directories above the VizInt git repo folders train everything and evaluate everything. It could take more than one day with a gaming laptop GPU.
python main.py --dataset cifar10
python main.py --dataset tinyimagenet
Models
Base Models
- VanillaViT: Vision Transformer with configurable depth and embedding dimensions
- ResNetBaseline: ResNet18 trained from scratch
- ResNetPretrained: ResNet18 with ImageNet pretrained weights
- ResNet18NotPretrainedRobust: ResNet18 trained from scratch with continuous transforms
Robustness-Enhanced Models
- Healer: Predicts and corrects transformations using Wiener deconvolution for noise and inverse transforms
- TTT/TTT3fc: Test-Time Training models that adapt during inference
- BlendedTraining: Predicts applied corruptions in the hope of getting a better feature map and more robust model.
Corrector + Classifier Combinations
- UNet + ResNet18/ViT: UNet corrector preprocessing for classifiers
- Transformer + ResNet18/ViT: Transformer-based corrector
- Hybrid + ResNet18/ViT: Combined CNN and Transformer corrector
Features
Continuous Transforms
The framework applies various transformations with adjustable severity: - Gaussian Noise: Additive noise with σ up to 0.5 - Rotation: Random rotations up to 360° - Affine: Translation and shear transformations
Robust Training
Based Models ending with _robust are trained with continuous transforms applied during training, making them more resilient to input perturbations, wrappers with robust means they contains a based model trained for robustness(with continuous transforms) experimental models are always trained for robustness as they need transformations augmentation to work.
Healer Model Capabilities
- Wiener Deconvolution: Advanced frequency-domain denoising
- Rotation Correction: Inverse rotation based on predicted angles
- Affine Correction: Reverses translation and shear transformations
- Transform Type Classification: Identifies which transformation was applied
Usage
Command Line Arguments
Options: --dataset {cifar10,tinyimagenet} Dataset to use (default: cifar10) --mode {train,evaluate,both} Operation mode (default: both) --models MODEL [MODEL ...] Models to train/evaluate (default: all) --skipmodels MODEL [MODEL ...] Models to skip --robust Train robust versions of TTT models --forceretrain Force retraining existing models --severities SEV [SEV ...] Severity levels for evaluation --debug Enable debug mode with small dataset --device {cuda,cpu,auto} Device to use --seed SEED Random seed ```
Training Individual Models
```bash
Train vanilla ViT
python main.py --dataset cifar10 --models vanilla_vit --mode train
Train ResNet18 with robust training
python main.py --dataset cifar10 --models resnet18notpretrained_robust --mode train
Train Healer model
python main.py --dataset cifar10 --models healer --mode train ```
Evaluation
```bash
Evaluate all models
python main.py --dataset cifar10 --mode evaluate
Evaluate with custom severity levels
python main.py --dataset cifar10 --mode evaluate --severities 0.0 0.25 0.5 0.75 1.0
Evaluate specific model
python main.py --dataset cifar10 --models resnet18notpretrained_robust --mode evaluate ```
Visualization Demos
Healer Visualization Demo
Demonstrates all Healer correction capabilities:
```bash
CIFAR-10
python demohealervisualizations.py
TinyImageNet
python demohealervisualizations.py --dataset tinyimagenet ```
This creates visualizations showing: - Corrections for Gaussian noise, rotation, and affine transforms - Comparison of different denoising methods - Performance across severity levels - Results with trained Healer models
Wiener Denoising Demo
Focuses on the Wiener deconvolution method:
```bash
CIFAR-10
python demohealerwiener.py
TinyImageNet
python demohealerwiener.py --dataset tinyimagenet ```
Shows: - Detailed denoising process with noise/residual maps - PSNR metrics and improvements - Performance across different noise levels
Transform Visualization Demo
Visualizes the continuous transforms:
bash
python demo_transforms.py
Configuration
Configuration files are in YAML format:
- config/cifar10_config.yaml: CIFAR-10 specific settings
- config/tinyimagenet_config.yaml: TinyImageNet specific settings
Key configuration sections: - dataset: Dataset paths and parameters - models: Model-specific hyperparameters - training: Training settings (epochs, learning rate, etc.) - evaluation: Evaluation parameters and severity levels - model_combinations: Defines which models to evaluate
Adding New Models
- Implement model in
models/directory - Register in
model_factory.py - Add to
all_modelslist inmain.py - Add to
all_model_typesinmodel_evaluator.py - Add configuration in YAML files
Results
The evaluation produces: - Clean accuracy: Performance on unmodified images - Robustness scores: Performance at different transformation severities - Transform prediction accuracy: For models that predict transformations - OOD performance: Results on out-of-distribution transforms
Results are saved to:
- Checkpoints: checkpoints/{dataset}/bestmodel_{model_name}/
- Visualizations: visualizations/{dataset}/
- Logs: experiment.log
Example Results Table
Model Clean S0.3 S0.5 S0.7 S1.0
ResNet18_Pretrained 0.9064 0.5826 0.5533 0.5421 0.5398
ResNet18_Baseline 0.8636 0.5787 0.5392 0.5169 0.5071
ResNet18_NotPretrainedRobust 0.8521 0.7234 0.6891 0.6523 0.6102
VanillaViT_Robust 0.7395 0.5255 0.4774 0.4523 0.4540
Advanced Features
Test-Time Training (TTT)
TTT models adapt their parameters during inference by solving a self-supervised task (rotation prediction).
Blended Training
Predict transforms augmentation corruputions in the hope of optaining a better feature map and a more robust model.
Corrector Models
Preprocessors that attempt to clean corrupted inputs before classification: - UNet: CNN-based denoising - Transformer: Attention-based correction - Hybrid: Combines CNN and Transformer approaches
🏆 COMPREHENSIVE RESULTS - CIFAR-10
Model Performance Results
| Model Combination | Description | Clean | S0.3 | S0.5 | S0.7 | S1.0 | |---|---|---|---|---|---|---| | ResNet18Pretrained | ResNet18 (ImageNet pretrained) | 0.9064 | 0.5853 | 0.5555 | 0.5456 | 0.5415 | | ResNet18NotPretrainedRobust | ResNet18 (from scratch, robust training) | 0.8654 | 0.5616 | 0.5213 | 0.5132 | 0.5083 | | ResNet18Baseline | ResNet18 (from scratch) | 0.8636 | 0.5819 | 0.5434 | 0.5195 | 0.5050 | | BlendedResNet18 | Blended wrapper with ResNet18 backbone | 0.8375 | 0.7654 | 0.7253 | 0.7026 | 0.6695 | | Transformer+ResNet18 | Transformer corrector + ResNet18 classifier | 0.8263 | 0.5573 | 0.5079 | 0.4876 | 0.4845 | | UNet+ResNet18 | UNet corrector + ResNet18 classifier | 0.8057 | 0.5602 | 0.5055 | 0.4879 | 0.4782 | | Hybrid+ResNet18 | Hybrid corrector + ResNet18 classifier | 0.7935 | 0.5210 | 0.4809 | 0.4764 | 0.4542 | | VanillaViTRobust | Vanilla ViT (robust training) | 0.7395 | 0.5242 | 0.4690 | 0.4567 | 0.4581 | | BlendedTraining | Blended Training (inherently robust) | 0.7328 | 0.4890 | 0.4753 | 0.4652 | 0.4567 | | VanillaViT | Vanilla ViT (not robust) | 0.7319 | 0.5045 | 0.4823 | 0.4563 | 0.4456 | | BlendedTraining3fc | Blended Training 3fc (inherently robust) | 0.7010 | 0.4773 | 0.4478 | 0.4354 | 0.4317 | | UNet+ViT | UNet corrector + Vision Transformer | 0.6885 | 0.4516 | 0.4257 | 0.4161 | 0.4104 | | Transformer+ViT | Transformer corrector + Vision Transformer | 0.6637 | 0.4881 | 0.4505 | 0.4276 | 0.4197 | | HealerResNet18 | Healer wrapper with ResNet18 backbone | 0.6409 | 0.5151 | 0.4508 | 0.4132 | 0.3876 | | Hybrid+ViT | Hybrid corrector + Vision Transformer | 0.6272 | 0.4058 | 0.3826 | 0.3814 | 0.3635 | | Healer+VanillaViT_Robust | Healer + Vanilla ViT (robust) | 0.2886 | 0.2399 | 0.2440 | 0.2343 | 0.2286 | | Healer+VanillaViT | Healer + Vanilla ViT (not robust) | 0.2304 | 0.1516 | 0.1461 | 0.1494 | 0.1477 | | TTTResNet18 | TTT wrapper with ResNet18 backbone | 0.1023 | 0.1021 | 0.1013 | 0.0977 | 0.0979 | | TTT | TTT (Test-Time Training) | 0.0946 | 0.1018 | 0.1021 | 0.0973 | 0.1003 | | TTT3fc | TTT3fc (Test-Time Training with 3FC) | 0.0935 | 0.0994 | 0.1003 | 0.1045 | 0.0995 |
📊 ANALYSIS
- 🥇 Best Clean Data Performance: ResNet18_Pretrained (0.9064)
- 🛡️ Most Transform Robust: TTTResNet18 (0.4% drop)
📊 TRANSFORMATION ROBUSTNESS SUMMARY
| Model | Sev 0.0 | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | Avg Drop | |---|---|---|---|---|---|---| | ResNet18Pretrained | 0.9064 | 0.5853 | 0.5555 | 0.5456 | 0.5415 | 0.3855 | | ResNet18NotPretrainedRobust | 0.8654 | 0.5616 | 0.5213 | 0.5132 | 0.5083 | 0.3921 | | ResNet18Baseline | 0.8636 | 0.5819 | 0.5434 | 0.5195 | 0.5050 | 0.3777 | | BlendedResNet18 | 0.8375 | 0.7654 | 0.7253 | 0.7026 | 0.6695 | 0.1454 | | Transformer+ResNet18 | 0.8263 | 0.5573 | 0.5079 | 0.4876 | 0.4845 | 0.3836 | | UNet+ResNet18 | 0.8057 | 0.5602 | 0.5055 | 0.4879 | 0.4782 | 0.3696 | | Hybrid+ResNet18 | 0.7935 | 0.5210 | 0.4809 | 0.4764 | 0.4542 | 0.3911 | | VanillaViTRobust | 0.7395 | 0.5242 | 0.4690 | 0.4567 | 0.4581 | 0.3550 | | BlendedTraining | 0.7328 | 0.4890 | 0.4753 | 0.4652 | 0.4567 | 0.3565 | | VanillaViT | 0.7319 | 0.5045 | 0.4823 | 0.4563 | 0.4456 | 0.3549 | | BlendedTraining3fc | 0.7010 | 0.4773 | 0.4478 | 0.4354 | 0.4317 | 0.3608 | | UNet+ViT | 0.6885 | 0.4516 | 0.4257 | 0.4161 | 0.4104 | 0.3813 | | Transformer+ViT | 0.6637 | 0.4881 | 0.4505 | 0.4276 | 0.4197 | 0.3273 | | HealerResNet18 | 0.6409 | 0.5151 | 0.4508 | 0.4132 | 0.3876 | 0.3109 | | Hybrid+ViT | 0.6272 | 0.4058 | 0.3826 | 0.3814 | 0.3635 | 0.3888 | | Healer+VanillaViT_Robust | 0.2886 | 0.2399 | 0.2440 | 0.2343 | 0.2286 | 0.1798 | | Healer+VanillaViT | 0.2304 | 0.1516 | 0.1461 | 0.1494 | 0.1477 | 0.3546 | | TTTResNet18 | 0.1023 | 0.1021 | 0.1013 | 0.0977 | 0.0979 | 0.0249 | | TTT | 0.0946 | 0.1018 | 0.1021 | 0.0973 | 0.1003 | -0.0610 | | TTT3fc | 0.0935 | 0.0994 | 0.1003 | 0.1045 | 0.0995 | -0.0794 |
🔍 HEALER GUIDANCE EVALUATION
🔍 Evaluating Healer+VanillaViT_Robust...
- Severity 0.3: Original: 0.5242, Healed: 0.2399, Improvement: -0.2843
- Severity 0.5: Original: 0.4690, Healed: 0.2440, Improvement: -0.2250
- Severity 0.7: Original: 0.4567, Healed: 0.2343, Improvement: -0.2224
- Severity 1.0: Original: 0.4581, Healed: 0.2286, Improvement: -0.2295
🔍 Evaluating Healer+VanillaViT...
- Severity 0.3: Original: 0.5242, Healed: 0.1516, Improvement: -0.3726
- Severity 0.5: Original: 0.4690, Healed: 0.1461, Improvement: -0.3229
- Severity 0.7: Original: 0.4567, Healed: 0.1494, Improvement: -0.3073
- Severity 1.0: Original: 0.4581, Healed: 0.1477, Improvement: -0.3104
🎯 TRANSFORMATION PREDICTION ACCURACY
| Model | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | Average | |---|---|---|---|---|---| | BlendedResNet18 | 0.9117 | 0.9393 | 0.9427 | 0.9398 | 0.9334 | | BlendedTraining | 0.0983 | 0.0597 | 0.0540 | 0.0467 | 0.0647 | | BlendedTraining3fc | 0.2330 | 0.2488 | 0.2503 | 0.2295 | 0.2404 | | HealerResNet18 | 0.2135 | 0.1878 | 0.1745 | 0.1623 | 0.1845 | | Healer+VanillaViT_Robust | 0.5775 | 0.5856 | 0.5669 | 0.5217 | 0.5629 | | Healer+VanillaViT | 0.5848 | 0.5949 | 0.5655 | 0.5257 | 0.5677 | | TTTResNet18 | 0.3633 | 0.3704 | 0.5288 | 0.5790 | 0.4604 |
📊 DETAILED TRANSFORM TYPE PREDICTION ACCURACY
BlendedResNet18
| Transform Type | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | |---|---|---|---|---| | gaussian_noise | 0.9918 | 0.9988 | 0.9996 | 0.9988 | | none | 0.8602 | 0.8707 | 0.8681 | 0.8658 | | rotate | 0.9233 | 0.9414 | 0.9450 | 0.9419 | | translate | 0.8746 | 0.9462 | 0.9589 | 0.9522 |
BlendedTraining
| Transform Type | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | |---|---|---|---|---| | gaussian_noise | 0.2355 | 0.0956 | 0.0609 | 0.0559 | | none | 0.1252 | 0.1193 | 0.1257 | 0.1085 | | rotate | 0.0000 | 0.0000 | 0.0000 | 0.0000 | | translate | 0.0318 | 0.0227 | 0.0292 | 0.0211 |
BlendedTraining3fc
| Transform Type | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | |---|---|---|---|---| | gaussian_noise | 0.2999 | 0.3652 | 0.3489 | 0.2459 | | none | 0.4388 | 0.4360 | 0.4470 | 0.4452 | | rotate | 0.0043 | 0.0008 | 0.0037 | 0.0037 | | translate | 0.1952 | 0.1803 | 0.1996 | 0.2170 |
HealerResNet18
| Transform Type | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | |---|---|---|---|---| | gaussian_noise | 0.7508 | 0.6442 | 0.5937 | 0.5144 | | none | 0.0027 | 0.0024 | 0.0016 | 0.0041 | | rotate | 0.0666 | 0.0613 | 0.0690 | 0.0674 | | translate | 0.0319 | 0.0331 | 0.0429 | 0.0501 |
Healer+VanillaViT_Robust
| Transform Type | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | |---|---|---|---|---| | gaussian_noise | 1.0000 | 1.0000 | 1.0000 | 1.0000 | | none | 0.0671 | 0.0612 | 0.0578 | 0.0621 | | rotate | 0.7408 | 0.7753 | 0.8061 | 0.7660 | | translate | 0.5042 | 0.5084 | 0.4176 | 0.2555 |
Healer+VanillaViT
| Transform Type | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | |---|---|---|---|---| | gaussian_noise | 0.9996 | 1.0000 | 1.0000 | 1.0000 | | none | 0.0626 | 0.0645 | 0.0523 | 0.0567 | | rotate | 0.7512 | 0.7923 | 0.8024 | 0.7831 | | translate | 0.5235 | 0.5275 | 0.4218 | 0.2627 |
TTTResNet18
| Transform Type | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | |---|---|---|---|---| | gaussian_noise | 0.0000 | 0.0052 | 0.6788 | 0.9948 | | none | 0.0529 | 0.0515 | 0.0555 | 0.0500 | | rotate | 0.5808 | 0.6580 | 0.6675 | 0.6616 | | translate | 0.7908 | 0.7741 | 0.7241 | 0.6100 |
📏 PARAMETER PREDICTION ACCURACY (Mean Absolute Error)
Healer+VanillaViT_Robust
| Parameter | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | Average | |---|---|---|---|---|---| | noise | 0.1371 | 0.0873 | 0.0162 | 0.1304 | 0.0928 |
Healer+VanillaViT
| Parameter | Sev 0.3 | Sev 0.5 | Sev 0.7 | Sev 1.0 | Average | |---|---|---|---|---|---| | noise | 0.1367 | 0.0870 | 0.0164 | 0.1303 | 0.0926 |
🚀 OUT-OF-DISTRIBUTION (FUNKY TRANSFORMS) EVALUATION
This section evaluates model performance on extreme, funky transformations including color inversion, pixelation, extreme blur, masking, etc.
| Model Combination | Description | Funky OOD | |---|---|---| | BlendedResNet18 | Blended wrapper with ResNet18 backbone | 0.4761 | | ResNet18Pretrained | ResNet18 (ImageNet pretrained) | 0.4449 | | ResNet18NotPretrainedRobust | ResNet18 (from scratch, robust training) | 0.4303 | | ResNet18Baseline | ResNet18 (from scratch) | 0.4299 | | Transformer+ResNet18 | Transformer corrector + ResNet18 classifier | 0.4236 | | UNet+ResNet18 | UNet corrector + ResNet18 classifier | 0.4086 | | Hybrid+ResNet18 | Hybrid corrector + ResNet18 classifier | 0.3941 | | VanillaViTRobust | Vanilla ViT (robust training) | 0.3489 | | HealerResNet18 | Healer wrapper with ResNet18 backbone | 0.3438 | | BlendedTraining | Blended Training (inherently robust) | 0.3407 | | UNet+ViT | UNet corrector + Vision Transformer | 0.3288 | | VanillaViT | Vanilla ViT (not robust) | 0.3267 | | Transformer+ViT | Transformer corrector + Vision Transformer | 0.3219 | | BlendedTraining3fc | Blended Training 3fc (inherently robust) | 0.3102 | | Hybrid+ViT | Hybrid corrector + Vision Transformer | 0.2943 | | Healer+VanillaViT_Robust | Healer + Vanilla ViT (robust) | 0.1801 | | Healer+VanillaViT | Healer + Vanilla ViT (not robust) | 0.1451 | | TTT | TTT (Test-Time Training) | 0.1017 | | TTTResNet18 | TTT wrapper with ResNet18 backbone | 0.0993 | | TTT3fc | TTT3fc (Test-Time Training with 3FC) | 0.0932 |
📊 OOD ANALYSIS
🥇 Best Funky OOD Performance: BlendedResNet18 (0.4761)
🔍 OOD vs Clean Performance Gap
- BlendedResNet18: Clean 0.8375 → OOD 0.4761 (Gap: 0.3614, 43.2%)
- ResNet18_Pretrained: Clean 0.9064 → OOD 0.4449 (Gap: 0.4615, 50.9%)
- ResNet18_NotPretrainedRobust: Clean 0.8654 → OOD 0.4303 (Gap: 0.4351, 50.3%)
- ResNet18_Baseline: Clean 0.8636 → OOD 0.4299 (Gap: 0.4337, 50.2%)
- Transformer+ResNet18: Clean 0.8263 → OOD 0.4236 (Gap: 0.4027, 48.7%)
🏆 OOD Robustness Ranking
- BlendedResNet18: 0.4761
- ResNet18_Pretrained: 0.4449
- ResNet18_NotPretrainedRobust: 0.4303
- ResNet18_Baseline: 0.4299
- Transformer+ResNet18: 0.4236
Owner
- Name: David Lacour
- Login: DavidLacour
- Kind: user
- Repositories: 1
- Profile: https://github.com/DavidLacour
Citation (citations.txt)
@article{hendrycks2019robustness,
title={Benchmarking Neural Network Robustness to Common Corruptions and Perturbations},
author={Hendrycks, Dan and Dietterich, Thomas},
journal={Proceedings of the International Conference on Learning Representations},
year={2019}
}
@article{ILSVRC15,
Author = {Olga Russakovsky and Jia Deng and Hao Su and Jonathan Krause and Sanjeev Satheesh and Sean Ma and Zhiheng Huang and Andrej Karpathy and Aditya Khosla and Michael Bernstein and Alexander C. Berg and Li Fei-Fei},
Title = {{ImageNet Large Scale Visual Recognition Challenge}},
Year = {2015},
journal = {International Journal of Computer Vision (IJCV)},
doi = {10.1007/s11263-015-0816-y},
volume={115},
number={3},
pages={211-252}
}
@inproceedings{
li2025laionc,
title={{LAION}-C: An Out-of-Distribution Benchmark for Web-Scale Vision Models},
author={Fanfei Li and Thomas Klein and Wieland Brendel and Robert Geirhos and Roland S. Zimmermann},
booktitle={Workshop on Spurious Correlation and Shortcut Learning: Foundations and Solutions},
year={2025},
url={https://openreview.net/forum?id=t1IBHkU2bt}
}
GitHub Events
Total
- Member event: 2
- Push event: 155
- Pull request event: 34
- Create event: 3
Last Year
- Member event: 2
- Push event: 155
- Pull request event: 34
- Create event: 3
Issues and Pull Requests
Last synced: 6 months ago
All Time
- Total issues: 0
- Total pull requests: 39
- Average time to close issues: N/A
- Average time to close pull requests: less than a minute
- Total issue authors: 0
- Total pull request authors: 2
- Average comments per issue: 0
- Average comments per pull request: 0.0
- Merged pull requests: 35
- Bot issues: 0
- Bot pull requests: 0
Past Year
- Issues: 0
- Pull requests: 39
- Average time to close issues: N/A
- Average time to close pull requests: less than a minute
- Issue authors: 0
- Pull request authors: 2
- Average comments per issue: 0
- Average comments per pull request: 0.0
- Merged pull requests: 35
- Bot issues: 0
- Bot pull requests: 0
Top Authors
Issue Authors
Pull Request Authors
- DavidLacour (37)
- thainamhoang (2)
Top Labels
Issue Labels
Pull Request Labels
Dependencies
- Pillow >=8.0.0
- einops >=0.7.0
- numpy >=1.21.0
- scikit-learn >=1.0.0
- scipy >=1.7.0
- timm >=0.9.0
- torch >=2.0.0
- torchaudio >=2.0.0
- torchvision >=0.15.0
- tqdm >=4.60.0
- wandb >=0.12.0
- einops ==0.8.1
- matplotlib ==3.9.4
- scikit-learn ==1.6.1
- scipy ==1.13.1
- timm ==1.0.15
- torch ==2.7.0
- torchaudio ==2.7.0
- torchvision ==0.22.0
- tqdm ==4.67.1
- wandb ==0.19.11
- GitPython ==3.1.44
- Jinja2 ==3.1.4
- MarkupSafe ==2.1.5
- PyYAML ==6.0.2
- annotated-types ==0.7.0
- certifi ==2025.4.26
- charset-normalizer ==3.4.2
- click ==8.1.8
- contourpy ==1.3.0
- cycler ==0.12.1
- docker-pycreds ==0.4.0
- einops ==0.8.1
- eval_type_backport ==0.2.2
- filelock ==3.13.1
- fonttools ==4.58.0
- fsspec ==2024.6.1
- gitdb ==4.0.12
- huggingface-hub ==0.31.4
- idna ==3.10
- importlib_resources ==6.5.2
- joblib ==1.5.0
- kiwisolver ==1.4.7
- matplotlib ==3.9.4
- mpmath ==1.3.0
- networkx ==3.2.1
- numpy ==1.26.3
- nvidia-cublas-cu11 ==11.11.3.6
- nvidia-cuda-cupti-cu11 ==11.8.87
- nvidia-cuda-nvrtc-cu11 ==11.8.89
- nvidia-cuda-runtime-cu11 ==11.8.89
- nvidia-cudnn-cu11 ==9.1.0.70
- nvidia-cufft-cu11 ==10.9.0.58
- nvidia-curand-cu11 ==10.3.0.86
- nvidia-cusolver-cu11 ==11.4.1.48
- nvidia-cusparse-cu11 ==11.7.5.86
- nvidia-nccl-cu11 ==2.21.5
- nvidia-nvtx-cu11 ==11.8.86
- packaging ==25.0
- pillow ==11.0.0
- platformdirs ==4.3.8
- protobuf ==6.31.0
- psutil ==7.0.0
- pydantic ==2.11.4
- pydantic_core ==2.33.2
- pyparsing ==3.2.3
- python-dateutil ==2.9.0.post0
- requests ==2.32.3
- safetensors ==0.5.3
- scikit-learn ==1.6.1
- scipy ==1.13.1
- sentry-sdk ==2.29.1
- setproctitle ==1.3.6
- six ==1.17.0
- smmap ==5.0.2
- sympy ==1.13.3
- threadpoolctl ==3.6.0
- timm ==1.0.15
- torch ==2.7.0
- torchaudio ==2.7.0
- torchvision ==0.22.0
- tqdm ==4.67.1
- triton ==3.3.0
- typing-inspection ==0.4.1
- typing_extensions ==4.12.2
- urllib3 ==2.4.0
- wandb ==0.19.11
- zipp ==3.21.0