hate speech detection is not as easy as you may think
play

Hate Speech Detection is Not as Easy as You May Think: A Closer Look - PowerPoint PPT Presentation

Hate Speech Detection is Not as Easy as You May Think: A Closer Look at Model Validation Aym Arango, Jorge Prez and Brbara Poblete UNDETECTED ALMOST PERFECT HATE SPEECH VS STATE-OF-THE-ART IN RESULTS SOCIAL MEDIA UNDETECTED HATE


  1. Hate Speech Detection is Not as Easy as You May Think: A Closer Look at Model Validation Aymé Arango, Jorge Pérez and Bárbara Poblete

  2. UNDETECTED ALMOST PERFECT HATE SPEECH VS STATE-OF-THE-ART IN RESULTS SOCIAL MEDIA

  3. UNDETECTED HATE SPEECH IN SOCIAL MEDIA

  4. 94% F1 [Agrawal and Awekar] ECIR 2018 93% F1 ALMOST PERFECT [Badjatiya et al.] WWW STATE-OF-THE-ART 2017 RESULTS 92% F1 [Zeerak Waseem] NAACL 2016

  5. Hate Speech Detection is Not as Easy as You May Think We show that state of the art results are highly overestimated due to experimental issues in the models: Including the testing set during training phase Oversampling the data before splitting User-biased datasets

  6. State-of-the-art replication User distribution Generalization

  7. State-of-the-art replication User distribution Generalization

  8. 94% F1 [Agrawal and Awekar] ECIR 2018 93% F1 ALMOST PERFECT [Badjatiya et al.] WWW STATE-OF-THE-ART 2017 RESULTS 92% F1 [Zeerak Waseem] NAACL 2016

  9. DATASET 1 [Waseem and Hovy] NAACL 2016 Tweet Label Hate Non-Hate

  10. Model 1 [Badjatiya et al.] 2017 DATASET 1 PHASE 1 PHASE 2 [Waseem and Hovy] NAACL Feature Extraction Classification Method 2016 93% F1

  11. PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method DATASET 1 [Waseem and Hovy] NAACL 2016 Embeddings LSTM Fully Connected Softmax Prediction

  12. PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method DATASET 1 [Waseem and Hovy] NAACL 2016 Embeddings LSTM Fully Connected Softmax Prediction

  13. PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method DATASET 1 [Waseem and Hovy] Splitting TRAIN TEST NAACL 2016 Embeddings Embeddings LSTM Fully Connected Softmax Prediction

  14. PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method DATASET 1 [Waseem and Hovy] Splitting TRAIN TEST NAACL 2016 93% F1 Embeddings AVG( Embeddings ) LSTM GBDT Prediction Fully Connected Softmax Prediction

  15. This looks great! But there is a problem.

  16. PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method DATASET 1 [Waseem and Hovy] Splitting TRAIN TEST NAACL 2016 TEST Embeddings AVG( Embeddings ) LSTM GBDT Prediction Fully Connected Softmax Prediction

  17. Let’s create the model only with the training set.

  18. PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method DATASET 1 [Waseem and Hovy] NAACL 2016

  19. New PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method TRAIN TEST Same Splitting TRAIN TEST

  20. New PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method TRAIN Same Splitting TRAIN TEST Embeddings LSTM Fully Connected Softmax Prediction

  21. New PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method TRAIN Same Splitting TRAIN TEST Embeddings Embeddings LSTM Fully Connected Softmax Prediction

  22. New PHASE 1 PHASE 2 Model 1 [Badjatiya et al.] 2017 Feature Extraction Classification Method TRAIN Same Splitting TRAIN TEST 73% F1 Embeddings AVG( Embeddings ) 93% F1 LSTM GBDT Prediction Fully Connected Softmax Prediction

  23. The result is overestimated due to the inclusion of the testing set during the training phase.

  24. Model 2 [Agrawal and Awekar] 2018 DATASET 1 Feature Extraction Oversampling [Waseem and Hovy] + Data NAACL Classification Method 2016 94% F1

  25. Model 2 [Agrawal and Awekar] 2018 DATASET 1 [Waseem and Hovy] NAACL 2016

  26. Model 2 [Agrawal and Awekar] 2018 TRAIN Oversampling Splitting TEST 94% F1 Embeddings LSTM Fully Connected Softmax Prediction

  27. This also looks great! But there is another problem.

  28. Model 2 [Agrawal and Awekar] 2018 DATASET 1 [Waseem and Hovy] NAACL 2016

  29. Model 2 [Agrawal and Awekar] 2018 TRAIN Oversampling Splitting TEST

  30. Model 2 [Agrawal and Awekar] 2018 Splitting Oversampling TEST 79% F1 Embeddings 94% F1 LSTM Fully Connected Softmax Prediction

  31. The result is overestimated due to the fact that the oversampling phase occurs before splitting the data.

  32. However, there is another issue to take into account.

  33. State-of-the-art replication User distribution Generalization

  34. % Tweets from the most prolific user per class 96 % 96% 44% 44 % 38% 25 % 25% Non-Hate Sexism Racism Hate

  35. DATASET 1 Splitting without [Waseem and Hovy] TRAIN TEST overlapped users NAACL 2016 Model 1 44% F1 73% F1 93% F1 [Badjatiya et al.] 2017 Model 2 35% F1 79% F1 94% F1 [Agrawal and Awekar] 2018

  36. What happens if we have a dataset with a better user distribution?

  37. DATASET 1 DATASET 2 DATASET 2 NEW 250 tweets [Davidson et al.] Hateful tweets DATASET per user ICWSM per class 2017

  38. NEW Splitting without TRAIN TEST DATASET overlapped users Model 1 78% F1 44% F1 73% F1 93% F1 [Badjatiya et al.] 2017 Model 2 76% F1 35% F1 79% F1 94% F1 [Agrawal and Awekar] 2018

  39. User distribution on datasets has an impact on the classification results.

  40. State-of-the-art replication User distribution Generalization

  41. TRAINING TESTING SET SET

  42. DATASET 3 TRAINING [Basile et al.] SET SemEval 2019

  43. DATASET 1 DATASET 3 47% F1 [Waseem and Hovy] [Basile et al.] NAACL SemEval 2016 2019 Model 1 [Badjatiya et al.] 2017 DATASET 3 NEW 51% F1 [Basile et al.] DATASET SemEval 2019 DATASET 1 DATASET 3 51% F1 [Waseem and Hovy] [Basile et al.] NAACL SemEval 2016 2019 Model 2 [Agrawal and Awekar] 2018 DATASET 3 NEW 54% F1 [Basile et al.] DATASET SemEval 2019

  44. Better user-distributed datasets lead to better generalization.

  45. Conclusions

  46. Hate Speech Detection is Not as Easy as You May Think We show that state of the art results are highly overestimated due to experimental issues in the models: Including the testing set during training phase Oversampling the data before splitting User-biased datasets

  47. Hate Speech Detection is Not as Easy as You May Think: A Closer Look at Model Validation Aymé Arango, Jorge Pérez and Bárbara Poblete

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend