Primary competition visual

Basic Needs Basic Rights Kenya - Tech4MentalHealth

Helping Kenya
$4 200 USD
Completed (over 5 years ago)
Classification
1244 joined
499 active
Starti
Apr 26, 20
Closei
Jul 05, 20
Reveali
Jul 05, 20
5th Place Solution
Notebooks · 28 Jul 2020, 18:28 · 4

Model - Single fine-tuned RoBERTa-base using Huggingface Transformers library with PyTorch.

NB: Not doing text preprocessing/cleaning performed better than doing so. All I had to do was remove duplicated texts.

Hyperparameters

  • 3 epochs
  • Learning rate of 5e-5
  • Batch Size of 8
  • 0.1 weight decay
  • 0.6 Multi-Sample Dropout
  • Learning rate scheduling (Linear)
  • Max text length of 35

Training

5 cv folds run 5 times with different seeds used in sampling data making a total of 25 runs. This was done in order to reduce variability in predictions as the data was very small. Test data predictions were done between folds, and later averaged in total.

Link to code

Discussion 4 answers

Awesome. Thanks for sharing.

28 Jul 2020, 18:37
Upvotes 0

Hi, thanks for sharing this. Amazing work.

28 Jul 2020, 18:51
Upvotes 0

Great approach!

28 Jul 2020, 19:50
Upvotes 0
User avatar
yakyo

Great work, thanks

28 Jul 2020, 20:39
Upvotes 0