azsl_dataloader

Data loader for the Azerbaijani Sign Language Dataset

https://github.com/ada-site-jml/azsl_dataloader

Science Score: 67.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
    Found 4 DOI reference(s) in README
  • Academic publication links
    Links to: arxiv.org
  • Academic email domains
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (11.0%) to scientific vocabulary
Last synced: 6 months ago · JSON representation ·

Repository

Data loader for the Azerbaijani Sign Language Dataset

Basic Info
  • Host: GitHub
  • Owner: ADA-SITE-JML
  • License: mit
  • Language: Jupyter Notebook
  • Default Branch: main
  • Size: 8.39 MB
Statistics
  • Stars: 1
  • Watchers: 0
  • Forks: 1
  • Open Issues: 0
  • Releases: 0
Created over 2 years ago · Last pushed about 1 year ago
Metadata Files
Readme License Citation

README.md

Data Loaders for the Azerbaijani Dactyl alphabet and Sign Language datasets

The repository contains the examples of data loaders written in Python, for the PyTorch framework, introduced as Jupyter Notebooks. This document describes the provided code in detail.

Data loader for the dactyl alphabet

The dactyl signs consist of: - short videos, that represent the letters signed as hand or wrist movements: 'D', '', 'Y', '', 'Z', 'C' and '' - images, that show the letter that does not require the motion of the hand, fingers or the wrist (all the remaining letter)

Data loader for the AzSL

The following diagram explains the flow of the word-based data loading process. It considers the following: - based on the given video source (Cam1 or Cam2), feature type (i3d, squeezenet's last convolutional layer or just pixel data of frames) and video processing tool (OpenCV, VidGear or TorchVision) extracts the data from the mp4 files - depending on the configuration, it may keep the frames with only hands, removing the idle parts of the video.

The output of the process is the encoded sentence and the corresponding video data. For debugging purposes, the name of the file is also added.

AzSL_DataLoader1

The data loader code starts with the initialization and config work, that defines the following: - Possibility of running notebook on the host machine, not on Colab's environment. It is possble if the user has good computational resources and do not want to be interrupted after 8 hours or execution - The source where the dataset is located: Global GoogleDrive, the user's Google Drive or any other accessible disk - The target computation unit: CPU, GPU or TPU. The system will always give preference to GPU. For the TPU, it should be set explicitly setting device = 'tpu' - The video processing library that the user prefers. As an option, one of the OpenCV, VidGear or TorchVision could be selected. TorchVision is the default selection - The camera source: a frontal or side camera could be used

The followingfigure graphically describes the flow:

fig_data_loader

During the dataloading process the video frames are resized and converted to float tensors. A user may decide to have all the frames or only frames with the hands. If number of the frames with the hands is less than the maxframes_, then the beginning and end are filled from the original video.

References

The povided works are originally described in the following papers and expected to be cited as a reference:

  1. J. Hasanov, N. Alishzade, A. Nazimzade, S. Dadashzade, T. Tahirov. Development of a hybrid word recognition system and dataset for the Azerbaijani Sign Language dactyl alphabet. Speech Communication, Volume 153, 2023, 102960, ISSN 0167-6393, https://doi.org/10.1016/j.specom.2023.102960.

  2. Nigar Alishzade, Jamaladdin Hasanov, AzSLD: Azerbaijani sign language dataset for fingerspelling, word, and sentence translation with baseline software. Data in Brief, Volume 58, 2025, 111230, ISSN 2352-3409, https://doi.org/10.1016/j.dib.2024.111230.

  3. The preprint of the second paper is available at: https://arxiv.org/abs/2411.12865

The citing could be done through the GitHub's new "Cite this repository" function.

Owner

  • Name: ADA-SITE-JML
  • Login: ADA-SITE-JML
  • Kind: organization

Citation (CITATION.cff)

@software{Data loader code for AzSL,
  author = {Jamal Hasanov},
  doi = {00.000/000.0000},
  month = {12},
  title = {{My Research Software}},
  url = {https://github.com/ADA-SITE-JML/azsl_dataloader},
  version = {3.4},
  year = {2023}
}

GitHub Events

Total
  • Watch event: 1
  • Push event: 2
Last Year
  • Watch event: 1
  • Push event: 2