Turftopic
Turftopic: Topic Modelling with Contextual Representations from Sentence Transformers - Published in JOSS (2025)
Speakerbox
Speakerbox: Few-Shot Learning for Speaker Identification with Transformers - Published in JOSS (2023)
Jury
Jury: A Comprehensive Evaluation Toolkit - Published in JOSS (2024)
pactus
pactus: A Python framework for trajectory classification - Published in JOSS (2023)
lazyllm-llamafactory
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
decimer
DECIMER Image Transformer is a deep-learning-based tool designed for automated recognition of chemical structure images. Leveraging transformer architectures, the model converts chemical images into SMILES strings, enabling the digitization of chemical data from scanned documents, literature, and patents.
txtai
💡 All-in-one open-source AI framework for semantic search, LLM orchestration and language model workflows
farm-haystack
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
adapters
A Unified Library for Parameter-Efficient and Modular Transfer Learning
rwkv-lm
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
learn_prompting
Prompt Engineering, Generative AI, and LLM Guide by Learn Prompting | Join our discord for the largest Prompt Engineering learning community
chronos-forecasting
Chronos: Pretrained Models for Probabilistic Time Series Forecasting
transformers-interpret
Model explainability that works seamlessly with 🤗 transformers. Explain your transformers model in just 2 lines of code.
mlm-bias
Measuring Biases in Masked Language Models for PyTorch Transformers. Support for multiple social biases and evaluation measures.
alignment-handbook
Robust recipes to align language models with human and AI preferences
tokenizers
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
wiki-entity-summarization-preprocessor
Convert Wikidata and Wikipedia raw files to filterable formats with a focus of marking Wikidata as summaries based on their Wikipedia abstracts.
knn-transformers
PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including an implementation of kNN-LM and kNN-MT
maestro
streamline the fine-tuning process for multimodal models: PaliGemma 2, Florence-2, and Qwen2.5-VL
transformer-srl
Reimplementation of a BERT based model (Shi et al, 2019), currently the state-of-the-art for English SRL. This model implements also predicate disambiguation.
optimum
🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization tools
linear-relational
Linear Relational Embeddings (LREs) and Linear Relational Concepts (LRCs) for LLMs in PyTorch
turkish-question-generation
Automated question generation and question answering from Turkish texts using text-to-text transformers
https://github.com/alan-turing-institute/robots-in-disguise
Information and materials for the Turing's "robots-in-disguise" reading group on fundamental AI research.
transformers-tutorials
This repository contains demos I made with the Transformers library by HuggingFace.
huggingsound
HuggingSound: A toolkit for speech-related tasks based on Hugging Face's tools
nerpy
🌈 NERpy: Implementation of Named Entity Recognition using Python. 命名实体识别工具,支持BertSoftmax、BertSpan等模型,开箱即用。
https://github.com/betswish/cross-lingual-consistency
Easy-to-use framework for evaluating cross-lingual consistency of factual knowledge (Supported LLaMA, BLOOM, mT5, RoBERTa, etc.) Paper here: https://aclanthology.org/2023.emnlp-main.658/
hugsvision
HugsVision is a easy to use huggingface wrapper for state-of-the-art computer vision
stormtrooper
Zero/few shot learning components for scikit-learn pipelines with LLMs and transformers.
marqo-fashionclip
State-of-the-art CLIP/SigLIP embedding models finetuned for the fashion domain. +57% increase in evaluation metrics vs FashionCLIP 2.0.
https://github.com/google-research/scenic
Scenic: A Jax Library for Computer Vision Research and Beyond
logfire-callback
A callback for logging training events from Hugging Face's Transformers to Logfire 🤗
https://github.com/superduper-io/superduper
Superduper: End-to-end framework for building custom AI applications and agents.
efficient-task-transfer
Research code for "What to Pre-Train on? Efficient Intermediate Task Selection", EMNLP 2021
napolab
The Natural Portuguese Language Benchmark (Napolab). Stay up to date with the latest advancements in Portuguese language models and their performance across carefully curated Portuguese language tasks.
https://github.com/biomedsciai/biomed-multi-omic
Build foundation model on RNA or DNA data
astronet
Efficient Deep Learning for Real-time Classification of Astronomical Transients and Multivariate Time-series
spacy-wrap
spaCy-wrap is a wrapper library for spaCy for including fine-tuned transformers from Huggingface in your spaCy pipeline allowing you to include existing fine-tuned models within your SpaCy workflow.
https://github.com/chris-santiago/met
Reproducing the MET framework with PyTorch
https://github.com/bioinfomachinelearning/deepinteract
A geometric deep learning framework (Geometric Transformers) for predicting protein interface contacts. (ICLR 2022)
https://github.com/aisuko/notebooks
Implementation for the different ML tasks on Kaggle platform with GPUs.
https://github.com/beomi/transformers-language-modeling
Train 🤗transformers with DeepSpeed: ZeRO-2, ZeRO-3
https://github.com/dadananjesha/ai-text-humanizer-app
Transform AI-generated text into formal, human-like, and academic writing with ease, avoids AI detector!
https://github.com/compvis/geometry-free-view-synthesis
Is a geometric model required to synthesize novel views from a single image?
https://github.com/beomi/bitnet-transformers
0️⃣1️⃣🤗 BitNet-Transformers: Huggingface Transformers Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch with Llama(2) Architecture
https://github.com/rindow/rindow-neuralnetworks
Neural networks library for machine learning on PHP
https://github.com/bramvanroy/lt3-2019-transformer-trainer
Transformer trainer for variety of classification problems that has been used in-house at LT3 for different research topics.
https://github.com/buaadreamer/dlpy
Programming Language for Deep Learning in Python
backprop
Backprop makes it simple to use, finetune, and deploy state-of-the-art ML models.
gpl
Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: Generative Pseudo Labeling for Unsupervised Domain Adaptation of Dense Retrieval" https://arxiv.org/abs/2112.07577
fluence
A deep learning library based on Pytorch focussed on low resource language research and robustness
https://github.com/daniel-furman/sft-demos
Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.
https://github.com/beomi/gemma-easylm
Train GEMMA on TPU/GPU! (Codebase for training Gemma-Ko Series)
ml4prom
The implementation of the paper: "Explainable Artificial Intelligence for Improved Modeling of Processes".
https://github.com/cluebbers/nlp_deeplearning_spring2023
Implementing and fine-tuning BERT for sentiment analysis, paraphrase detection, and semantic textual similarity tasks. Includes code, data, and detailed results.
dpo-rlhf-paraphrase-types
Enhancing paraphrase-type generation using Direct Preference Optimization (DPO) and Reinforcement Learning from Human Feedback (RLHF), with large-scale HPC support. This project aligns model outputs to human-ranked data for robust, safety-focused NLP.
unified-graph-transformer
Unified Graph Transformer (UGT) is a novel Graph Transformer model specialised in preserving both local and global graph structures and developed by NS Lab @ CUK based on pure PyTorch backend.
pangoling
An R package for estimating the log-probabilities of words in a given context using transformer models.
linktransformer
A convenient way to link, deduplicate, aggregate and cluster data(frames) in Python using deep learning
stb-vmm
STB-VMM: Swin Transformer Based Video Motion Magnification (official repository)
transformers-explained
官方transformers源码解析。AI大模型时代,pytorch、transformer是新操作系统,其他都是运行在其上面的软件。
speechbrain
A PyTorch-based Speech Toolkit
biologicaltokenizers
Effect of tokenization on transformers for biological sequence
layoutreader
A Faster LayoutReader Model based on LayoutLMv3, Sort OCR bboxes to reading order.
layer_norm_expressivity_role
Code for the paper "On the Expressivity Role of LayerNorm in Transformers' Attention" (Findings of ACL'2023)
llm-confidentiality
Whispers in the Machine: Confidentiality in Agentic Systems
interlinear-translation
Morphology-enhanced neural models for Ancient Greek interlinear translation, achieving 35-38% BLEU improvements for English and Polish translations. Includes custom T5 implementations and training code. [LoResLM@COLING2025]
https://github.com/alan-turing-institute/prompto
An open source library for asynchronous querying of LLM endpoints
https://github.com/earthai-tech/fusionlab-learn
fusionlab-learn: Igniting Next-Gen Temporal Fusion Architectures
https://github.com/cyberagentailab/japanese-nli-model
This repository provides the code for Japanese NLI model, a fine-tuned masked language model.
tuned-lens
Tools for understanding how transformer predictions are built layer-by-layer
transformers-tf-finetune
Scripts to finetune huggingface transformers models with Tensorflow 2
transact-tf
An unofficial implementation of "TransAct: Transformer-based Realtime User Action Model for Recommendation at Pinterest" in Tensorflow
hate-speech-detection
Developed a hate speech detection model using PyTorch. Applied natural language processing techniques and deep learning algorithms to identify and classify offensive language. Achieved high accuracy in detecting hate speech, contributing to the development of more inclusive online communities.