Ad click fraud detection Christian Benson and Adam Thuvesen Problem - - PowerPoint PPT Presentation

ad click fraud detection
SMART_READER_LITE
LIVE PREVIEW

Ad click fraud detection Christian Benson and Adam Thuvesen Problem - - PowerPoint PPT Presentation

Ad click fraud detection Christian Benson and Adam Thuvesen Problem Ad click fraud Mobile Click fraud is a major issue for advertisers Pay per click ads The app creator (publisher) will profit from more clicks


slide-1
SLIDE 1

Christian Benson and Adam Thuvesen

Ad click fraud detection

slide-2
SLIDE 2

Problem

  • Ad click fraud

○ Mobile

  • Click fraud is a major issue for advertisers

○ Pay per click ads ■ The app creator (publisher) will profit from more clicks ○ Fraudulent automated clicks ■ The advertiser loses

slide-3
SLIDE 3

Problem

  • How to detect a fraudulent click in a mobile app?

○ Using data from ad clicks

slide-4
SLIDE 4

Dataset

  • Dataset from Kaggle
  • 7 features

○ ip (ip address) ○ app (mobile app) ○ device (type of device) ○

  • s (operating system)

○ channel (channel id of mobile ad publisher) ○ click time (ad was clicked) ○ attributed time (time of possible download) ○ is attributed (ad led to app download or not)

slide-5
SLIDE 5

Dataset

  • 187M entries
  • Very unbalanced

○ 99.8 % negative samples (not downloaded)

slide-6
SLIDE 6

Baseline

  • Dummy
  • k-NN
  • SVM
  • Logistic Regression
  • Decision Trees
  • Random Forest
  • Metric

○ ROC-AUC

slide-7
SLIDE 7

Architecture

Raw features Model Training Prediction

  • Raw data is used to train model
  • Using trained model to predict on test set

Test data Download: 0.01 Not download: 0.99 Training data

slide-8
SLIDE 8

Idea

  • Decision trees performed well
  • Research in the area supported various ensemble
  • f decision trees to be successful in similar

problems

  • Data preprocessing - extract new features
  • Gradient boosted trees

○ Frameworks ■ XGB popular ■ Microsofts LGBM newly gaining attention

  • Neural net
slide-9
SLIDE 9

How it works - Decision Trees

Ensemble of Decision Trees

slide-10
SLIDE 10

How it works - Gradient Boosted Trees

Gradient Boosted Trees

  • Error = bias + variance
slide-11
SLIDE 11
  • Data preprocessing - extract new features

○ Unique occurrences ○ Total count ○ Cumulative count ○ Variance ○ Mean ○ Aggregation ○ Previous/next click ○ Time

  • 23-30 features in total

Data preprocessing

slide-12
SLIDE 12

Training

  • Trained on 10M entries
  • Models

○ Neural net with embedding layer ○ LGBM ○ XGB

slide-13
SLIDE 13

Solution

  • Feature Engineering

○ Create new features from existing ones

  • Gradient Boosted Trees

○ XGB ○ LGBM

  • Ensemble of LGBM and XGB models
  • Neural net not performing quite as well
slide-14
SLIDE 14

Ensemble

  • Combining two or more models for better results
  • Can be done in several ways
  • Logarithmic average
slide-15
SLIDE 15

Solution architecture

Raw data Feature engineering LGBM model 2 training XGB model 1 training LGBM Prediction XGB Prediction Test data Ensemble prediction Training data LGBM model 1 training LGBM Prediction XGB model 2 training XGB Prediction

slide-16
SLIDE 16

Results

  • LGBM best model: 0.9784
  • XGB best model: 0.9733
  • Neural net best model: 0.9508
  • Logarithmic ensemble mix including the two

best LGBM and the two best XGB: 0.9787

slide-17
SLIDE 17

Thank you for listening!