Small Data, Big Problems? Limited training data can make models overfit, meaning they perform well on the competition set but struggle in the real world. One outlier in the test data could determine the winner - not ideal!
Validation! Validating models with small datasets is tough. Test data is covered different locations or scenarios compared to training data, making it hard to assess how well a model generalizes. Especially with RMSE as the metric, outliers can have a big impact.
Yes I call for the authors to change the metrics to MAE. Because as it stands, the winner will be the model with the luck of best performing on outliers.
Yes I call for the authors to change the metrics to MAE. Because as it stands, the winner will be the model with the luck of best performing on outliers.
you know, zindi is funny of RMSE
I came to the same conclusion
I hope the best solution will not be putted in trash