Turftopic
Turftopic: Topic Modelling with Contextual Representations from Sentence Transformers - Published in JOSS (2025)
Speakerbox
Speakerbox: Few-Shot Learning for Speaker Identification with Transformers - Published in JOSS (2023)
Jury
Jury: A Comprehensive Evaluation Toolkit - Published in JOSS (2024)
pactus
pactus: A Python framework for trajectory classification - Published in JOSS (2023)
lazyllm-llamafactory
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
decimer
DECIMER Image Transformer is a deep-learning-based tool designed for automated recognition of chemical structure images. Leveraging transformer architectures, the model converts chemical images into SMILES strings, enabling the digitization of chemical data from scanned documents, literature, and patents.
txtai
💡 All-in-one open-source AI framework for semantic search, LLM orchestration and language model workflows
farm-haystack
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
adapters
A Unified Library for Parameter-Efficient and Modular Transfer Learning
rwkv-lm
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
learn_prompting
Prompt Engineering, Generative AI, and LLM Guide by Learn Prompting | Join our discord for the largest Prompt Engineering learning community
chronos-forecasting
Chronos: Pretrained Models for Probabilistic Time Series Forecasting
transformers-interpret
Model explainability that works seamlessly with 🤗 transformers. Explain your transformers model in just 2 lines of code.
mlm-bias
Measuring Biases in Masked Language Models for PyTorch Transformers. Support for multiple social biases and evaluation measures.
alignment-handbook
Robust recipes to align language models with human and AI preferences
tokenizers
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
wiki-entity-summarization-preprocessor
Convert Wikidata and Wikipedia raw files to filterable formats with a focus of marking Wikidata as summaries based on their Wikipedia abstracts.
knn-transformers
PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including an implementation of kNN-LM and kNN-MT
maestro
streamline the fine-tuning process for multimodal models: PaliGemma 2, Florence-2, and Qwen2.5-VL
transformer-srl
Reimplementation of a BERT based model (Shi et al, 2019), currently the state-of-the-art for English SRL. This model implements also predicate disambiguation.
optimum
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
linear-relational
Linear Relational Embeddings (LREs) and Linear Relational Concepts (LRCs) for LLMs in PyTorch
turkish-question-generation
Automated question generation and question answering from Turkish texts using text-to-text transformers
https://github.com/alan-turing-institute/robots-in-disguise
Information and materials for the Turing's "robots-in-disguise" reading group on fundamental AI research.
transformers-tutorials
This repository contains demos I made with the Transformers library by HuggingFace.
huggingsound
HuggingSound: A toolkit for speech-related tasks based on Hugging Face's tools
nerpy
🌈 NERpy: Implementation of Named Entity Recognition using Python. 命名实体识别工具,支持BertSoftmax、BertSpan等模型,开箱即用。
https://github.com/betswish/cross-lingual-consistency
Easy-to-use framework for evaluating cross-lingual consistency of factual knowledge (Supported LLaMA, BLOOM, mT5, RoBERTa, etc.) Paper here: https://aclanthology.org/2023.emnlp-main.658/
hugsvision
HugsVision is a easy to use huggingface wrapper for state-of-the-art computer vision
stormtrooper
Zero/few shot learning components for scikit-learn pipelines with LLMs and transformers.
marqo-fashionclip
State-of-the-art CLIP/SigLIP embedding models finetuned for the fashion domain. +57% increase in evaluation metrics vs FashionCLIP 2.0.
https://github.com/google-research/scenic
Scenic: A Jax Library for Computer Vision Research and Beyond
logfire-callback
A callback for logging training events from Hugging Face's Transformers to Logfire 🤗
https://github.com/superduper-io/superduper
Superduper: End-to-end framework for building custom AI applications and agents.
efficient-task-transfer
Research code for "What to Pre-Train on? Efficient Intermediate Task Selection", EMNLP 2021
napolab
The Natural Portuguese Language Benchmark (Napolab). Stay up to date with the latest advancements in Portuguese language models and their performance across carefully curated Portuguese language tasks.
https://github.com/biomedsciai/biomed-multi-omic
Build foundation model on RNA or DNA data
astronet
Efficient Deep Learning for Real-time Classification of Astronomical Transients and Multivariate Time-series
spacy-wrap
spaCy-wrap is a wrapper library for spaCy for including fine-tuned transformers from Huggingface in your spaCy pipeline allowing you to include existing fine-tuned models within your SpaCy workflow.
https://github.com/chris-santiago/met
Reproducing the MET framework with PyTorch
https://github.com/bioinfomachinelearning/deepinteract
A geometric deep learning framework (Geometric Transformers) for predicting protein interface contacts. (ICLR 2022)
https://github.com/aisuko/notebooks
Implementation for the different ML tasks on Kaggle platform with GPUs.
https://github.com/beomi/transformers-language-modeling
Train 🤗transformers with DeepSpeed: ZeRO-2, ZeRO-3
https://github.com/dadananjesha/ai-text-humanizer-app
Transform AI-generated text into formal, human-like, and academic writing with ease, avoids AI detector!
https://github.com/compvis/geometry-free-view-synthesis
Is a geometric model required to synthesize novel views from a single image?
https://github.com/beomi/bitnet-transformers
0️⃣1️⃣🤗 BitNet-Transformers: Huggingface Transformers Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch with Llama(2) Architecture
https://github.com/rindow/rindow-neuralnetworks
Neural networks library for machine learning on PHP
https://github.com/bramvanroy/lt3-2019-transformer-trainer
Transformer trainer for variety of classification problems that has been used in-house at LT3 for different research topics.
https://github.com/buaadreamer/dlpy
Programming Language for Deep Learning in Python
backprop
Backprop makes it simple to use, finetune, and deploy state-of-the-art ML models.
gpl
Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: Generative Pseudo Labeling for Unsupervised Domain Adaptation of Dense Retrieval" https://arxiv.org/abs/2112.07577
fluence
A deep learning library based on Pytorch focussed on low resource language research and robustness
https://github.com/daniel-furman/sft-demos
Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.
https://github.com/beomi/gemma-easylm
Train GEMMA on TPU/GPU! (Codebase for training Gemma-Ko Series)
https://github.com/cedrickchee/pytorch-pretrained-bert
PyTorch version of Google AI's BERT model with script to load Google's pre-trained models
https://github.com/chirayu-tripathi/paper-implementations
My implementation of Machine Learning and Deep Learning papers from scratch.
https://github.com/cosbidev/naim
Official implementation for the paper ``Not Another Imputation Method: A Transformer-based Model for Missing Values in Tabular Datasets´´
balena
BALanced Execution through Natural Activation : a human-computer interaction methodology for code running.
https://github.com/cyberagentailab/japanese-nli-model
This repository provides the code for Japanese NLI model, a fine-tuned masked language model.
transformers-tf-finetune
Scripts to finetune huggingface transformers models with Tensorflow 2
llm-confidentiality
Whispers in the Machine: Confidentiality in Agentic Systems
https://github.com/dc-research/tempo
The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for forecasting task v1.0 version.
https://github.com/ai4co/routefinder
[ICML'24 FM-Wild Oral] RouteFinder: Towards Foundation Models for Vehicle Routing Problems
bitcoin-trader-ml
Automated 24/7 bitcoin trader for Coinbase using Transformer Neural Networks
master-thesis
One Bit at a Time: Impact of Quantisation on Neural Machine Translation
tsdae
Tranformer-based Denoising AutoEncoder for Sentence Transformers Unsupervised pre-training.
indic-syntax-evaluation
Vyākarana: A Colorless Green Benchmark for Syntactic Evaluation in Indic Languages
https://github.com/chris-santiago/tsfeast
A collection of Scikit-Learn compatible time series transformers and tools.
speechbrain
A PyTorch-based Speech Toolkit
https://github.com/cluebbers/nlp_deeplearning_spring2023
Implementing and fine-tuning BERT for sentiment analysis, paraphrase detection, and semantic textual similarity tasks. Includes code, data, and detailed results.
dpo-rlhf-paraphrase-types
Enhancing paraphrase-type generation using Direct Preference Optimization (DPO) and Reinforcement Learning from Human Feedback (RLHF), with large-scale HPC support. This project aligns model outputs to human-ranked data for robust, safety-focused NLP.
partial-embedding-matrix-adaptation
Vocabulary-level memory efficiency for language model fine-tuning.
pangoling
An R package for estimating the log-probabilities of words in a given context using transformer models.
linktransformer
A convenient way to link, deduplicate, aggregate and cluster data(frames) in Python using deep learning
https://github.com/alan-turing-institute/prompto
An open source library for asynchronous querying of LLM endpoints