LargeBatchCTR aims to train CTR prediction models with large batch (~128k). The framework is based on DeepCTR. You can run the code on a V100 GPU to feel the fast training speed.
Adaptive Column-wise Clipping (CowClip) method from paper “CowClip: Reducing CTR Prediction Model Training
Time from 12 hours to 10 minutes on 1 GPU” is implemented in this repo.
Get Started
First, download dataset to the data folder. Use data_utils.py to preprocess the data for training.
Large Batch Training for CTR Prediction (CowClip)
LargeBatchCTR aims to train CTR prediction models with large batch (~128k). The framework is based on DeepCTR. You can run the code on a V100 GPU to feel the fast training speed.
Adaptive Column-wise Clipping (CowClip) method from paper “CowClip: Reducing CTR Prediction Model Training Time from 12 hours to 10 minutes on 1 GPU” is implemented in this repo.
Get Started
First, download dataset to the data folder. Use
data_utils.pyto preprocess the data for training.Then, use
train.pyto train the network.For large batch training with CowClip, do as follows:
CowClip Quick Look
Dataset List
train.txtindata/criteo_kaggle/trainindata/avazu/Hyperparameters
The meaning of hyperparameters in the command line is as follows:
The hyperparameters neet to be scaled are listed as follows. For Criteo dataset:
For Avazu dataset:
Model List
Requirements
Tensorflow 2.4.0
Tensorflow-Addons
Citation