tweetynet

Hybrid convolutional-recurrent neural networks for segmentation of birdsong and classification of elements

https://github.com/yardencsgithub/tweetynet

Science Score: 59.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
    Found 6 DOI reference(s) in README
  • Academic publication links
    Links to: zenodo.org
  • Committers with academic emails
    1 of 6 committers (16.7%) from academic institutions
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (14.2%) to scientific vocabulary

Keywords from Contributors

bioacoustics vocalizations animal-communication animal-vocalizations bioacoustic-analysis birdsong spectrograms speech-processing torchvision acoustic-communication
Last synced: 6 months ago · JSON representation

Repository

Hybrid convolutional-recurrent neural networks for segmentation of birdsong and classification of elements

Basic Info
  • Host: GitHub
  • Owner: yardencsGitHub
  • License: bsd-3-clause
  • Language: Jupyter Notebook
  • Default Branch: master
  • Homepage:
  • Size: 356 MB
Statistics
  • Stars: 55
  • Watchers: 6
  • Forks: 11
  • Open Issues: 0
  • Releases: 18
Created almost 9 years ago · Last pushed about 3 years ago
Metadata Files
Readme License Citation

README.md

DOI <!-- ALL-CONTRIBUTORS-BADGE:START - Do not remove or modify this section --> All Contributors <!-- ALL-CONTRIBUTORS-BADGE:END --> PyPI version

TweetyNet

Now published in eLife: https://elifesciences.org/articles/63853
Code to reproduce results from the article is in the directory ./article

What is tweetynet?

A neural network architecture (shown below) that automates annotation of birdsong and other vocalizations by segmenting spectrograms, and then labeling those segments.

neural network architecture

This is an example of the kind of annotations that tweetynet learns to predict:

How is it used?

Installation

Short version (for details see below):

with pip

console $ pip install tweetynet

with conda
on Mac and Linux

console $ conda install tweetynet -c conda-forge

on Windows

On Windows, you need to add an additional channel, pytorch.
You can do this by repeating the -c option more than once. console $ conda install tweetynet -c conda-forge -c pytorch $ # ^ notice additional channel!

Long version:
To facilitate training tweetynet models and using trained models to predict annotation on new datasets, we developed the vak library, that is installed automatically with tweetynet.

If you need more information about installation, please see the vak documentation:
https://vak.readthedocs.io/en/latest/get_started/installation.html

Usage

To train models and use them to predict annotation

For a tutorial on using tweetynet with vak, please see the vak documentation:
https://vak.readthedocs.io/en/latest/get_started/autoannotate.html

To reproduce results from article

In the directory ./article we provide code to reproduce the results in the article
"TweetyNet: A neural network that enables high-throughput, automated annotation of birdsong"
https://elifesciences.org/articles/63853

Please see the README in that directory for instructions on how to install and work with that code.

FAQs

Training data

To train models, you must supply training data in the form of audio files or spectrogram files, and annotations. The package can generate spectrograms from .wav or .cbin audio files. It can also accept spectrograms in the form of Matlab .mat files or .npz files created by numpy. vak uses a separate library to parse annotations, crowsetta, which handles some common formats and can also be used to write custom parsers for other formats. Please see the crowsetta documentation for more detail:
https://crowsetta.readthedocs.io/en/latest/#

Preparing training files

It is possible to train on any manually annotated data but there are some useful guidelines: * Use as many examples as possible - The results will just be better. Specifically, this code will not label correctly syllables it did not encounter while training and will most probably generalize to the nearest sample or ignore the syllable. * Use noise examples - This will make the code very good in ignoring noise. * Examples of syllables on noise are important - It is a good practice to start with clean recordings. The code will not perform miracles and is most likely to fail if the audio is too corrupt or masked by noise. Still, training with examples of syllables on the background of cage noises will be beneficial.

For more details, please see the vak documentation.

Issues

If you run into problems, please use the issue tracker or contact the authors via email in the paper above.

Citation

If you use or adapt this code, please cite its DOI:
DOI

License

Released under BSD license.

Contributors

Thanks goes to these wonderful people (emoji key):


yardencsGitHub


David Nicholson


Zhehao Cheng

This project follows the all-contributors specification. Contributions of any kind welcome!

Owner

  • Name: Yarden Cohen
  • Login: yardencsGitHub
  • Kind: user
  • Location: Rehovot, Israel
  • Company: Weizmann Institute of Science

GitHub Events

Total
  • Watch event: 7
  • Fork event: 1
Last Year
  • Watch event: 7
  • Fork event: 1

Committers

Last synced: over 1 year ago

All Time
  • Total Commits: 813
  • Total Committers: 6
  • Avg Commits per committer: 135.5
  • Development Distribution Score (DDS): 0.616
Past Year
  • Commits: 0
  • Committers: 0
  • Avg Commits per committer: 0.0
  • Development Distribution Score (DDS): 0.0
Top Committers
Name Email Commits
NickleDave n****e 312
David Nicholson N****e 195
David Nicholson n****v@g****m 194
yardencsGitHub y****c@b****u 81
yardencsGitHub y****l@g****m 25
allcontributors[bot] 4****] 6
Committer Domains (Top 20 + Academic)
bu.edu: 1

Issues and Pull Requests

Last synced: 7 months ago

All Time
  • Total issues: 53
  • Total pull requests: 51
  • Average time to close issues: 4 months
  • Average time to close pull requests: 3 days
  • Total issue authors: 8
  • Total pull request authors: 3
  • Average comments per issue: 1.55
  • Average comments per pull request: 0.37
  • Merged pull requests: 51
  • Bot issues: 0
  • Bot pull requests: 3
Past Year
  • Issues: 0
  • Pull requests: 0
  • Average time to close issues: N/A
  • Average time to close pull requests: N/A
  • Issue authors: 0
  • Pull request authors: 0
  • Average comments per issue: 0
  • Average comments per pull request: 0
  • Merged pull requests: 0
  • Bot issues: 0
  • Bot pull requests: 0
Top Authors
Issue Authors
  • NickleDave (46)
  • sfcooke96 (1)
  • abscodeice (1)
  • WalidERRAIS (1)
  • danstowell (1)
  • avanikop (1)
  • vivinastase (1)
  • ashishbijlani (1)
Pull Request Authors
  • NickleDave (44)
  • yardencsGitHub (4)
  • allcontributors[bot] (3)
Top Labels
Issue Labels
writing/paper (15) essential-revision (11) DOC (7) experiment (7) wontfix (6) BUG (3) minor-point-revision (2) figure (2) ENH (1) DEV (1)
Pull Request Labels

Packages

  • Total packages: 2
  • Total downloads:
    • pypi 94 last-month
  • Total dependent packages: 1
    (may contain duplicates)
  • Total dependent repositories: 1
    (may contain duplicates)
  • Total versions: 20
  • Total maintainers: 1
pypi.org: tweetynet

neural network that segments and labels birdsong and other animal vocalizations

  • Versions: 18
  • Dependent Packages: 1
  • Dependent Repositories: 1
  • Downloads: 94 Last month
Rankings
Dependent packages count: 4.7%
Stargazers count: 9.9%
Forks count: 11.4%
Average: 16.2%
Dependent repos count: 21.7%
Downloads: 33.4%
Maintainers (1)
Last synced: 6 months ago
conda-forge.org: tweetynet
  • Versions: 2
  • Dependent Packages: 0
  • Dependent Repositories: 0
Rankings
Dependent repos count: 34.0%
Stargazers count: 39.6%
Average: 42.7%
Forks count: 46.0%
Dependent packages count: 51.2%
Last synced: 6 months ago

Dependencies

.github/workflows/ci.yml actions
  • actions/checkout v2 composite
  • actions/setup-python v2 composite
  • excitedleigh/setup-nox v2.1.0 composite
article/environment.yml pypi
  • absl-py ==1.0.0
  • attrs ==21.4.0
  • cachetools ==4.2.4
  • certifi ==2021.10.8
  • cffi ==1.15.0
  • charset-normalizer ==2.0.9
  • cloudpickle ==2.0.0
  • crowsetta ==3.2.0
  • cycler ==0.11.0
  • dask ==2021.12.0
  • evfuncs ==0.3.3
  • fonttools ==4.28.5
  • fsspec ==2021.11.1
  • google-auth ==2.3.3
  • google-auth-oauthlib ==0.4.6
  • grpcio ==1.43.0
  • idna ==3.3
  • importlib-metadata ==4.10.0
  • joblib ==1.1.0
  • kiwisolver ==1.3.2
  • koumura ==0.2.1.post1
  • locket ==0.2.1
  • markdown ==3.3.6
  • matplotlib ==3.5.1
  • oauthlib ==3.1.1
  • packaging ==21.3
  • pandas ==1.3.5
  • partd ==1.2.0
  • protobuf ==3.19.1
  • pyasn1 ==0.4.8
  • pyasn1-modules ==0.2.8
  • pycparser ==2.21
  • pyparsing ==3.0.6
  • python-dateutil ==2.8.2
  • pytz ==2021.3
  • pyyaml ==6.0
  • requests ==2.26.0
  • requests-oauthlib ==1.3.0
  • rsa ==4.8
  • scipy ==1.7.3
  • soundfile ==0.10.3.post1
  • tensorboard ==2.7.0
  • tensorboard-data-server ==0.6.1
  • tensorboard-plugin-wit ==1.8.0
  • toml ==0.10.2
  • toolz ==0.11.2
  • tqdm ==4.62.3
  • tweetynet ==0.6.0
  • urllib3 ==1.26.7
  • vak ==0.4.0
  • werkzeug ==2.0.2
  • zipp ==3.7.0
article/pyproject.toml pypi
  • hybrid-vocal-classifier >=0.3.1
  • jupyterlab >=3.2.5
  • pyprojroot >=0.2.0
  • scikit-learn >=1.0.1
  • seaborn >=0.11.2
  • tweetynet >=0.6.0
pyproject.toml pypi
  • torch >=1.7.1
  • vak >=0.8.0