scene_labeling

Compute annotations for a given image that account for spatial relationships in the final summaries

https://github.com/jedavis82/scene_labeling

Science Score: 57.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
    Found 1 DOI reference(s) in README
  • Academic publication links
  • Academic email domains
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (13.6%) to scientific vocabulary
Last synced: 6 months ago · JSON representation ·

Repository

Compute annotations for a given image that account for spatial relationships in the final summaries

Basic Info
  • Host: GitHub
  • Owner: jedavis82
  • License: mit
  • Language: Python
  • Default Branch: main
  • Size: 1.37 MB
Statistics
  • Stars: 3
  • Watchers: 1
  • Forks: 0
  • Open Issues: 0
  • Releases: 0
Created over 3 years ago · Last pushed about 3 years ago
Metadata Files
Readme License Citation

README.md

Scene Labeling

This code base contains the S2T system constructed in CITE PAPER.

This project was created and tested using Python version 3.10.

This code is free to use but the authors ask that if you make use of any of the code during research you cite the work using [7].

Installation

Virtual environment

The codebase was created using Python virtual environments. These environments are recommended as they will not install packages directly to your system. A Python virtual environment can be created using the following command from a terminal:

python -m venv venv

YOLOv3 Object Detection Model

The code base uses the YOLOv3 object detection model. Due to size constraints on the repository, the model weights could not be uploaded. The model weights can be downloaded from the YOLOv3 site.

The required file can be directly downloaded from: - yolov3.weights

The code base originally stored these files in the input/models/yolo/ directory alongside the yolov3.cfg file.

Installing required packages

The requirements.txt file contains the required Python packages in order to perform object detection and scene annotations. To install the required packages, run the following command from the repo root directory in the terminal:

pip install -r requirements.txt

HoFPy

The Histogram of Forces [2] implementation used by this project was supplied by the University of Missouri MINDFUL [6] laboratory.

To install the HOF python package run the following commands from the repo root directory in the terminal:

cd HoFPy

pip install .

Usage

The demo.py script contains an example usage of the scene annotation system.

This script runs the scene annotation system on the six images located in the input/demo_images/ directory.

Example System Output

Below are some example outputs generated by the S2T system. Each table contains the localization results on the left with the General domain annotations and Person domain annotations on the right.

| Object Localization | Level Two Summaries | | ------------------- | ------------------- | | | General: Person1 interacting with bench1
Person: Person1 sitting on bench1 | | | General: Person1 interacting with umbrella1
Person: Person1 carrying umbrella1 | | | General: Person1 interacting with cellphone1
Person: person
1 talking on cellphone1 |

Attribution

This work was published in the 2022 IEEE World Congress on Computational Intelligence [7].

References

[1] Redmon, J. and Farhadi, A., "YOLOv3: An Incremental Improvement," arXiv, 2018.

[2] Matsakis, P. and Wendling, L., “New Way to Represent the Relative Position between Areal Objects,” IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 21, No. 7, 1999, pp. 634-643.

[3] Matsakis, P., Keller, J., Wendling, L., Marjamaa, J. and Sjahputera, O., "Linguistic Description of Relative Positions of Objects in Images", IEEE Transactions on Systems, Man, and Cybernetics, Vol. 31, No. 4, 2001, pp. 573-588.

[4] Matsakis, P., Keller, J., Sjahputera, O., and Marjamaa, J. “The Use of Force Histograms for Affine-Invariant Relative Position Description”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 26, No. 1, 2004, pp.1-18.

[5] Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., and Savarese, S, "Generalized Intersection Over Union: A Metric and A Loss for Bounding Box Regression", IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.

[6] Anderson, D. and Buck, A. "Mizzou Information and Data Fusion Laboratory." Accessed: July 2022. Available: https://github.com/MizzouINDFUL

[7] J. Davis et al., "Spatial Relationship-Driven Computer Vision Image Data Set Annotation," 2022 International Joint Conference on Neural Networks (IJCNN), 2022, pp. 1-8, doi: 10.1109/IJCNN55064.2022.9892975.

Owner

  • Name: Jeremy Davis
  • Login: jedavis82
  • Kind: user

PhD in computer science with a focus on computer vision and scene understanding.

Citation (CITATION.cff)

cff-version: 1.2.0
message: "If you use this software, please cite it as below. This software is free to use with the disclaimer that the authors are not responsible for any misuse."
authors:
- family-names: "Davis"
  given-names: "Jeremy"
title: "Scene Labeling: Annotations for Image Datasets"
version: 1.0.0
doi: 10.5281/zenodo.1234
date-released: 2021-12-20
url: "https://github.com/jedavis82/scene_labeling"

GitHub Events

Total
  • Watch event: 1
Last Year
  • Watch event: 1

Dependencies

requirements.txt pypi
  • Pillow >=8.4.0
  • fiftyone *
  • matplotlib >=3.5.1
  • numpy >=1.22.0
  • opencv-python >=4.6
  • pandas >=1.3.4
  • pybind11 >=2.9.0
  • scikit-fuzzy >=0.4.2
  • scikit-image >=0.18.3
  • scikit-learn >=1.0.1
  • seaborn >=0.11.2
  • torch >=1.11.0
  • torchvision >=0.12.0
  • tqdm >=4.62.3