fuxictr

A configurable, tunable, and reproducible library for CTR prediction https://fuxictr.github.io

https://github.com/reczoo/fuxictr

Science Score: 77.0%

This score indicates how likely this project is to be science-related based on various indicators:

  • CITATION.cff file
    Found CITATION.cff file
  • codemeta.json file
    Found codemeta.json file
  • .zenodo.json file
    Found .zenodo.json file
  • DOI references
    Found 9 DOI reference(s) in README
  • Academic publication links
    Links to: arxiv.org, acm.org
  • Committers with academic emails
    1 of 13 committers (7.7%) from academic institutions
  • Institutional organization owner
  • JOSS paper metadata
  • Scientific vocabulary similarity
    Low similarity (10.6%) to scientific vocabulary

Keywords

ctr ctr-prediction cvr pytorch recommender-systems

Keywords from Contributors

interpretability standardization hack
Last synced: 6 months ago · JSON representation ·

Repository

A configurable, tunable, and reproducible library for CTR prediction https://fuxictr.github.io

Basic Info
  • Host: GitHub
  • Owner: reczoo
  • License: apache-2.0
  • Language: Python
  • Default Branch: main
  • Homepage:
  • Size: 2.31 MB
Statistics
  • Stars: 1,253
  • Watchers: 14
  • Forks: 199
  • Open Issues: 13
  • Releases: 0
Topics
ctr ctr-prediction cvr pytorch recommender-systems
Created over 4 years ago · Last pushed 8 months ago
Metadata Files
Readme Changelog License Citation

README.md

Logo
Python version Pytorch version Pytorch version Pypi version Downloads License

Click-through rate (CTR) prediction is a critical task for various industrial applications such as online advertising, recommender systems, and sponsored search. FuxiCTR provides an open-source library for CTR prediction, with key features in configurability, tunability, and reproducibility. We hope this project could promote reproducible research and benefit both researchers and practitioners in this field.

Key Features

  • Configurable: Both data preprocessing and models are modularized and configurable.

  • Tunable: Models can be automatically tuned through easy configurations.

  • Reproducible: All the benchmarks can be easily reproduced.

  • Extensible: It can be easily extended to any new models, supporting both Pytorch and Tensorflow frameworks.

Model Zoo

| No | Publication | Model | Paper | Benchmark | Version | |:---:|:-----------------:|:----------------------------------------:|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |:---------------------------------------------------------------------------------------------------------------:|:-------------:| |:openfilefolder: Feature Interaction Models| | 1 | WWW'07 | LR | Predicting Clicks: Estimating the Click-Through Rate for New Ads :triangularflagonpost:Microsoft | :arrowupperright: | torch | | 2 | ICDM'10 | FM | Factorization Machines | :arrowupperright: | torch | | 3 | CIKM'13 | DSSM | Learning Deep Structured Semantic Models for Web Search using Clickthrough Data :triangularflagonpost:Microsoft | :arrowupperright: | torch | | 4 | CIKM'15 | CCPM | A Convolutional Click Prediction Model | :arrowupperright: | torch | | 5 | RecSys'16 | FFM | Field-aware Factorization Machines for CTR Prediction :triangularflagonpost:Criteo | :arrowupperright: | torch | | 6 | RecSys'16 | DNN | Deep Neural Networks for YouTube Recommendations :triangularflagonpost:Google | :arrowupperright: | torch, tf | | 7 | DLRS'16 | Wide&Deep | Wide & Deep Learning for Recommender Systems :triangularflagonpost:Google | :arrowupperright: | torch, tf | | 8 | ICDM'16 | PNN | Product-based Neural Networks for User Response Prediction | :arrowupperright: | torch | | 9 | KDD'16 | DeepCrossing | Deep Crossing: Web-Scale Modeling without Manually Crafted Combinatorial Features :triangularflagonpost:Microsoft | :arrowupperright: | torch | | 10 | NIPS'16 | HOFM | Higher-Order Factorization Machines | :arrowupperright: | torch | | 11 | IJCAI'17 | DeepFM | DeepFM: A Factorization-Machine based Neural Network for CTR Prediction :triangularflagonpost:Huawei | :arrowupperright: | torch, tf | | 12 | SIGIR'17 | NFM | Neural Factorization Machines for Sparse Predictive Analytics | :arrowupperright: | torch | | 13 | IJCAI'17 | AFM | Attentional Factorization Machines: Learning the Weight of Feature Interactions via Attention Networks | :arrowupperright: | torch | | 14 | ADKDD'17 | DCN | Deep & Cross Network for Ad Click Predictions :triangularflagonpost:Google | :arrowupperright: | torch, tf | | 15 | WWW'18 | FwFM | Field-weighted Factorization Machines for Click-Through Rate Prediction in Display Advertising :triangularflagonpost:Oath, TouchPal, LinkedIn, Alibaba | :arrowupperright: | torch | | 16 | KDD'18 | xDeepFM | xDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems :triangularflagonpost:Microsoft | :arrowupperright: | torch | | 17 | CIKM'19 | FiGNN | FiGNN: Modeling Feature Interactions via Graph Neural Networks for CTR Prediction | :arrowupperright: | torch | | 18 | CIKM'19 | AutoInt/AutoInt+ | AutoInt: Automatic Feature Interaction Learning via Self-Attentive Neural Networks | :arrowupperright: | torch | | 19 | RecSys'19 | FiBiNET | FiBiNET: Combining Feature Importance and Bilinear feature Interaction for Click-Through Rate Prediction :triangularflagonpost:Sina Weibo | :arrowupperright: | torch | | 20 | WWW'19 | FGCNN | Feature Generation by Convolutional Neural Network for Click-Through Rate Prediction :triangularflagonpost:Huawei | :arrowupperright: | torch | | 21 | AAAI'19 | HFM/HFM+ | Holographic Factorization Machines for Recommendation | :arrowupperright: | torch | | 22 | Arxiv'19 | DLRM | Deep Learning Recommendation Model for Personalization and Recommendation Systems :triangularflagonpost:Facebook | :arrowupperright: | torch | | 23 | NeuralNetworks'20 | ONN | Operation-aware Neural Networks for User Response Prediction | :arrowupperright: | torch, tf | | 24 | AAAI'20 | AFN/AFN+ | Adaptive Factorization Network: Learning Adaptive-Order Feature Interactions | :arrowupperright: | torch | | 25 | AAAI'20 | LorentzFM | Learning Feature Interactions with Lorentzian Factorization :triangularflagonpost:eBay | :arrowupperright: | torch | | 26 | WSDM'20 | InterHAt | Interpretable Click-through Rate Prediction through Hierarchical Attention :triangularflagonpost:NEC Labs, Google | :arrowupperright: | torch | | 27 | DLP-KDD'20 | FLEN | FLEN: Leveraging Field for Scalable CTR Prediction :triangularflagonpost:Tencent | :arrowupperright: | torch | | 28 | CIKM'20 | DeepIM | Deep Interaction Machine: A Simple but Effective Model for High-order Feature Interactions :triangularflagonpost:Alibaba, RealAI | :arrowupperright: | torch | | 29 | WWW'21 | FmFM | FM^2: Field-matrixed Factorization Machines for Recommender Systems :triangularflagonpost:Yahoo | :arrowupperright: | torch | | 30 | WWW'21 | DCN-V2 | DCN V2: Improved Deep & Cross Network and Practical Lessons for Web-scale Learning to Rank Systems :triangularflagonpost:Google | :arrowupperright: | torch | | 31 | CIKM'21 | DESTINE | Disentangled Self-Attentive Neural Networks for Click-Through Rate Prediction :triangularflagonpost:Alibaba | :arrowupperright: | torch | | 32 | CIKM'21 | EDCN | Enhancing Explicit and Implicit Feature Interactions via Information Sharing for Parallel Deep CTR Models :triangularflagonpost:Huawei | :arrowupperright: | torch | | 33 | DLP-KDD'21 | MaskNet | MaskNet: Introducing Feature-Wise Multiplication to CTR Ranking Models by Instance-Guided Mask :triangularflagonpost:Sina Weibo | :arrowupperright: | torch | | 34 | SIGIR'21 | SAM | Looking at CTR Prediction Again: Is Attention All You Need? :triangularflagonpost:BOSS Zhipin | :arrowupperright: | torch | | 35 | KDD'21 | AOANet | Architecture and Operation Adaptive Network for Online Recommendations :triangularflagonpost:Didi Chuxing | :arrowupperright: | torch | | 36 | AAAI'23 | FinalMLP | FinalMLP: An Enhanced Two-Stream MLP Model for CTR Prediction :triangularflagonpost:Huawei | :arrowupperright: | torch | | 37 | SIGIR'23 | FinalNet | FINAL: Factorized Interaction Layer for CTR Prediction :triangularflagonpost:Huawei | :arrowupperright: | torch | | 38 | SIGIR'23 | EulerNet | EulerNet: Adaptive Feature Interaction Learning via Euler's Formula for CTR Prediction :triangularflagonpost:Huawei | :arrowupperright: | torch | | 39 | CIKM'23 | GDCN | Towards Deeper, Lighter and Interpretable Cross Network for CTR Prediction :triangularflagonpost:Microsoft | | torch | | 40 | ICML'24 | WuKong | Wukong: Towards a Scaling Law for Large-Scale Recommendation :triangularflagonpost:Meta | :arrowupperright: | torch | |:openfilefolder: Behavior Sequence Modeling| | 42 | KDD'18 | DIN | Deep Interest Network for Click-Through Rate Prediction :triangularflagonpost:Alibaba | :arrowupperright: | torch | | 43 | AAAI'19 | DIEN | Deep Interest Evolution Network for Click-Through Rate Prediction :triangularflagonpost:Alibaba | :arrowupperright: | torch | | 44 | DLP-KDD'19 | BST | Behavior Sequence Transformer for E-commerce Recommendation in Alibaba :triangularflagonpost:Alibaba | :arrowupperright: | torch | | 45 | CIKM'20 | DMIN | Deep Multi-Interest Network for Click-through Rate Prediction :triangularflagonpost:Alibaba | :arrowupperright: | torch | | 46 | AAAI'20 | DMR | Deep Match to Rank Model for Personalized Click-Through Rate Prediction :triangularflagonpost:Alibaba | :arrowupperright: | torch | | 47 | KDD'23 | TransAct | TransAct: Transformer-based Realtime User Action Model for Recommendation at Pinterest :triangularflagonpost:Pinterest | :arrowupperright: | torch | |:openfilefolder: Long Sequence Modeling| | 48 | CIKM'20 | SIM | Search-based User Interest Modeling with Lifelong Sequential Behavior Data for Click-Through Rate Prediction :triangularflagonpost:Alibaba | | torch | | 49 | DLP-KDD'22 | ETA | Efficient Long Sequential User Data Modeling for Click-Through Rate Prediction :triangularflagonpost:Alibaba | | torch | | 50 | CIKM'22 | SDIM | Sampling Is All You Need on Modeling Long-Term User Behaviors for CTR Prediction :triangularflagonpost:Meituan | | torch | | 51 | KDD'23 | TWIN | TWIN: TWo-stage Interest Network for Lifelong User Behavior Modeling in CTR Prediction at Kuaishou :triangularflagonpost:KuaiShou | | torch | | 52 | KDD'25 | MIRRN | Multi-granularity Interest Retrieval and Refinement Network for Long-Term User Behavior Modeling in CTR Prediction :triangularflagonpost:Huawei | | torch | |:openfilefolder: Dynamic Weight Network| | 53 | NeurIPS'22 | APG | APG: Adaptive Parameter Generation Network for Click-Through Rate Prediction :triangularflagonpost:Alibaba | :arrowupperright: | torch | | 54 | KDD'23 | PPNet | PEPNet: Parameter and Embedding Personalized Network for Infusing with Personalized Prior Information :triangularflagonpost:KuaiShou | :arrowupperright: | torch | |:openfilefolder: Multi-Task Modeling| | 55 | Arxiv'17 | ShareBottom | An Overview of Multi-Task Learning in Deep Neural Networks | | torch | | 56 | KDD'18 | MMoE | Modeling Task Relationships in Multi-task Learning with Multi-Gate Mixture-of-Experts :triangularflagonpost:Google | | torch | | 57 | RecSys'20 | PLE | Progressive Layered Extraction (PLE): A Novel Multi-Task Learning (MTL) Model for Personalized Recommendations :triangularflagonpost:Tencent | | torch |

Benchmarking

We have benchmarked FuxiCTR models on a set of open datasets as follows:

Dependencies

FuxiCTR has the following dependencies:

  • python 3.9+
  • pytorch 1.10.0--2.1.2 (if using for torch models)
  • tensorflow 2.1 (if using for tensorflow models)

Please install other required packages via pip install -r requirements.txt.

Quick Start

  1. Run the demo examples

    Examples are provided in the demo directory to show some basic usage of FuxiCTR. Users can run the examples for quick start and to understand the workflow.

cd demo python example1_build_dataset_to_parquet.py python example2_DeepFM_with_parquet_input.py

  1. Run a model on tiny data

    Users can easily run each model in the model zoo following the commands below, which is a demo for running DCN. In addition, users can modify the dataset config and model config files to run on their own datasets or with new hyper-parameters. More details can be found in the README.

``` cd modelzoo/DCN/DCNtorch python runexpid.py --expid DCNtest --gpu 0

# Change MODEL according to the target model name cd modelzoo/MODEL python runexpid.py --expid MODEL_test --gpu 0 ```

  1. Run a model on benchmark datasets (e.g., Criteo)

Users can follow the benchmark section to get benchmark datasets and running steps for reproducing the existing results. Please see an example here: https://github.com/reczoo/BARS/tree/main/ranking/ctr/DCNv2/DCNv2criteox1

  1. Implement a new model

The FuxiCTR library is designed to be modularized, so that every component can be overwritten by users according to their needs. In many cases, only the model class needs to be implemented for a new customized model. If data preprocessing or data loader is not directly applicable, one can also overwrite a new one through the core APIs. We show a concrete example which implements our new model FinalMLP that has been recently published in AAAI 2023.

  1. Tune hyper-parameters of a model

FuxiCTR currently support fast grid search of hyper-parameters of a model using multiple GPUs. The following example shows the grid search of 8 experiments with 4 GPUs.

cd experiment python run_param_tuner.py --config config/DCN_tiny_parquet_tuner_config.yaml --gpu 0 1 2 3 0 1 2 3

🔥 Citation

If you use our code or benchmarks in your public research, please cite the following two papers.

🙋 Discussion

Welcome to join our WeChat group for any question and discussion. If you are interested in research and practice in recommender systems, please reach out via our WeChat group.

Scan QR code

Owner

  • Name: RECZOO
  • Login: reczoo
  • Kind: organization

Open Science by XUEPAI

Citation (CITATION)

@incollection{FuxiCTR,
  author    = {Jieming Zhu and
               Jinyang Liu and
               Shuai Yang and
               Qi Zhang and
               Xiuqiang He},
  title     = {Open Benchmarking for Click-Through Rate Prediction},
  booktitle = {The 30th {ACM} International Conference on Information
               and Knowledge Management (CIKM'21)},
  pages     = {2759--2769},
  year      = {2021}
}

@incollection{BARS,
  author    = {Jieming Zhu and
               Quanyu Dai and
               Liangcai Su and
               Rong Ma and
               Jinyang Liu and
               Guohao Cai and
               Xi Xiao and
               Rui Zhang},
  title     = {BARS: Towards Open Benchmarking for Recommender Systems},
  booktitle = {The 45th International ACM SIGIR Conference on Research 
               and Development in Information Retrieval (SIGIR'22)},
  year      = {2022}
}

GitHub Events

Total
  • Issues event: 60
  • Watch event: 298
  • Issue comment event: 39
  • Push event: 24
  • Pull request event: 7
  • Fork event: 43
  • Create event: 5
Last Year
  • Issues event: 60
  • Watch event: 298
  • Issue comment event: 39
  • Push event: 24
  • Pull request event: 7
  • Fork event: 43
  • Create event: 5

Committers

Last synced: 9 months ago

All Time
  • Total Commits: 126
  • Total Committers: 13
  • Avg Commits per committer: 9.692
  • Development Distribution Score (DDS): 0.302
Past Year
  • Commits: 31
  • Committers: 6
  • Avg Commits per committer: 5.167
  • Development Distribution Score (DDS): 0.258
Top Committers
Name Email Commits
xpai x****i 88
zhujiem z****m 22
LiangcaiSu 3****u 3
github-actions[bot] 4****] 2
XiaoLongtaoo 9****o 2
Honghao Li 7****2 2
乳酸君、 r****j@p****g 1
lu-minous 4****s 1
leesoojin l****2@n****m 1
ccfco 5****o 1
Tian Zhen 6****Z 1
Serdarcan Dilbaz s****z@g****m 1
Dansheng 3****g 1
Committer Domains (Top 20 + Academic)

Issues and Pull Requests

Last synced: 6 months ago

All Time
  • Total issues: 55
  • Total pull requests: 14
  • Average time to close issues: 7 days
  • Average time to close pull requests: 19 days
  • Total issue authors: 33
  • Total pull request authors: 11
  • Average comments per issue: 1.33
  • Average comments per pull request: 1.43
  • Merged pull requests: 7
  • Bot issues: 0
  • Bot pull requests: 0
Past Year
  • Issues: 38
  • Pull requests: 6
  • Average time to close issues: 5 days
  • Average time to close pull requests: 2 days
  • Issue authors: 24
  • Pull request authors: 4
  • Average comments per issue: 0.87
  • Average comments per pull request: 0.5
  • Merged pull requests: 0
  • Bot issues: 0
  • Bot pull requests: 0
Top Authors
Issue Authors
  • zhujiem (16)
  • Sharan123 (5)
  • ywangwxd (3)
  • clementineyyy (2)
  • byby221b (2)
  • xpai (1)
  • Lcy1-1 (1)
  • AIyumeng (1)
  • XudxDon (1)
  • AdamLTy (1)
  • ChengQianlong (1)
  • quency711 (1)
  • rsliu94 (1)
  • Hizkai (1)
  • houWenK (1)
Pull Request Authors
  • salmon1802 (4)
  • sdilbaz (2)
  • ccfco (2)
  • LiangcaiSu (2)
  • XiaoLongtaoo (2)
  • nilozdemir (2)
  • lu-minous (2)
  • rsliu94 (2)
  • Ethan-TZ (2)
  • w32zhong (1)
  • samins (1)
  • FuatOgme (1)
  • milein4u (1)
Top Labels
Issue Labels
bug (1)
Pull Request Labels
bug (2)

Packages

  • Total packages: 1
  • Total downloads:
    • pypi 646 last-month
  • Total dependent packages: 0
  • Total dependent repositories: 2
  • Total versions: 31
  • Total maintainers: 1
pypi.org: fuxictr

A configurable, tunable, and reproducible library for CTR prediction

  • Versions: 31
  • Dependent Packages: 0
  • Dependent Repositories: 2
  • Downloads: 646 Last month
Rankings
Stargazers count: 2.4%
Forks count: 4.3%
Average: 8.7%
Dependent packages count: 10.1%
Dependent repos count: 11.6%
Downloads: 15.1%
Maintainers (1)
Last synced: 6 months ago

Dependencies

requirements.txt pypi
  • PyYAML *
  • h5py *
  • keras_preprocessing *
  • numpy *
  • pandas *
  • scikit-learn *
  • torch *
  • tqdm *
setup.py pypi
  • pandas *
.github/workflows/jupyter-book.yml actions
.github/workflows/pypi.yml actions
  • pypa/gh-action-pypi-publish release/v1 composite