AASP-processing

Custom scripts for processing arff and TextGrid files to be used in the Automatic Analysis of Speech Prosody application

https://github.com/CentreForDigitalHumanities/AASP-processing

Science Score: 44.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
  • Academic publication links
  • Academic email domains
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (16.1%) to scientific vocabulary

Keywords

praat prosody weka
Last synced: 8 months ago · JSON representation ·

Repository

Custom scripts for processing arff and TextGrid files to be used in the Automatic Analysis of Speech Prosody application

Basic Info
  • Host: GitHub
  • Owner: CentreForDigitalHumanities
  • License: bsd-3-clause
  • Language: Python
  • Default Branch: master
  • Homepage:
  • Size: 15.6 KB
Statistics
  • Stars: 0
  • Watchers: 7
  • Forks: 0
  • Open Issues: 0
  • Releases: 1
Topics
praat prosody weka
Created over 6 years ago · Last pushed over 1 year ago
Metadata Files
Readme License Citation

README.md

AASP-processing

This reposistory contains scripts to train speech prosody classifiers, to be used in the Automatic Analysis of Speech Prosody application. The scripts collected here can be used to use audio features extracted with the speech prosody analysis tool AuToBI to predict human annotations in .TextGrid format.

For more documentation about the file types: - arff format, as used by the machine learning library Weka. These files are a specialized version of a csv, in which each feature name is listed as @ATTRIBUTE on top of the file, and the feature value is a cell in a comma-separated grid - TextGrid format, as used by audio annotation and analysis software Praat.

Prerequesites

  1. Make sure you have Python 3.6 installed (e.g. via Anaconda)
  2. If you do not want to install the required libraries globally, make a conda environment (see instructions here)
  3. Navigate to this directory and run pip install -r requirements.txt

Running

The repository contains two scripts: 1. process_arff can be used to combine audio features computed by AuToBI (in arff format) with human annotations (in TextGrid format). python process_arff.py -a PATH/TO/YOUR/ARFF/FEATURE/FILES -t PATH/TO/YOUR/TEXTGRID/FILES will: - adjust the headers (@ATTRIBUTEs) of the arff feature files and write new files into a directory. By default, this will be subdirectory called 'processed' in the source file directory, you can set another directory with the -c flag. - step through the directory with TextGrid files (including subdirectories) and collect the annotated intonations, and combine them with the features from the arff files in a large output file. By default, the complete table is written to combined.arff, and a table with only numerical parameters will be written to combined_nostring.arff in this directory. You can set another filename with the -o flag. - For help with the script, call python process_arff.py --help 2. As a next step, train_classifiers documents how to train a new classifier using Python's scikit-learn, in which the features extracted from the .arff file will be use to predict the annotations extracted from the .TextGrid file. It shows how to achieve train / test split, and suggests that a random forest classifier may achieve around 0.9 accuracy for prediction of tone and accent, which were the annotated features in data used for making a proof of concept.

Owner

  • Name: Centre for Digital Humanities
  • Login: CentreForDigitalHumanities
  • Kind: organization
  • Email: cdh@uu.nl
  • Location: Netherlands

Interdisciplinary centre for research and education in computational and data-driven methods in the humanities.

Citation (CITATION.cff)

# This CITATION.cff file was generated with cffinit.
# Visit https://bit.ly/cffinit to generate yours today!

cff-version: 1.2.0
title: AASP-processing
message: >-
  If you use this software, please cite it using the
  metadata from this file.
type: software
authors:
  - name: 'Research Software Lab, Centre for Digital Humanities, Utrecht University'
    website: 'https://cdh.uu.nl/centre-for-digital-humanities/research-software-lab/'
    city: Utrecht
    country: NL
identifiers:
  - type: doi
    value: 10.5281/zenodo.10849705
repository-code: 'https://github.com/UUDigitalHumanitieslab/aasp-processing'
abstract: >-
  This reposistory contains scripts to train speech prosody classifiers, to be used in the Automatic Analysis of Speech Prosody application. The scripts collected here can be used to use audio features extracted with the speech prosody analysis tool AuToBI to predict human annotations in .TextGrid format.
keywords:
  - speech
  - prosody
  - classification
license: BSD-3-Clause
version: 0.0.1
date-released: '2024-03-21'

GitHub Events

Total
  • Member event: 1
Last Year
  • Member event: 1

Dependencies

requirements.txt pypi
  • Click ==7.0
  • liac-arff ==2.4.0
  • pandas ==1.1.0
  • pympi-ling ==1.69
  • sklearn ==0.0