Primary competition visual

AI/ML for 5G-Energy Consumption Modelling by ITU AI/ML in 5G Challenge

20 000 CHF
Challenge completed ~2 years ago
Prediction
989 joined
278 active
Starti
Jul 26, 23
Closei
Oct 13, 23
Reveali
Oct 17, 23
Users who did not make a submission selection
Platform · 18 Oct 2023, 18:56 · 2

I'm thinking about point 3 from previous post, that the users who did not make a submission selection, should not assesment\present in leaderboard or should assessed by maximum score.

Many other participants who made submission selection had better scores in their other hundreds submissions. Therefore, the leaderboard should be scored on the minimum score basis from all submissions to ensure fairness and accurate representation of performance.

In this case participants from point 3 would be in equal conditions with the others ..... otherwise they got the advantage over others.

This is indirectly confirmed by the fact that some of the participants’ submissions were made 2 months ago. Therefore, maybe the best submissions were selected among all the results for these users, and according to the rules, the participant must himself determine 2 variants.

Discussion 2 answers
User avatar
Amy_Bray
Zindi

Hi Pavel,

As mentioned by Julius in the other thread, and lacking in my explanation.

For users who did not do submission selection, we only look at their public score, the bug is that it was choosing their worst public score, instead of their best public score. The correlating private score is revealed.

Oftentimes, the best public score is not the best private score.

Our rules recommend you do model selection for a couple of reasons, but users are welcome not to do submission selection.

The reasons to do model selection.

  1. It encourages experimentation and confidence in your model on unseen data.
  2. Before we had submission selection, code review was a problem as we found users weren't using version control or were not noting which versions they used to submit so top users were not reproducing their score. By encouraging choosing 2 submissions, we mitigate the chance of not reproducing.

For this particular competition, users in the top 20, will be required to submit a report and prizes will be awarded on:

In evaluating the final submission, both the quality of the report (weighted 50%) and the achieved model score (weighted 50%) will be considered.

P.S. Emails to top 20 should be sent out soon, please do not worry, we have not forgotten you.

19 Oct 2023, 03:51
Upvotes 0

Thanks for this complete explanation. Reproducibility of the result is indeed a separate big problem (for example, Catboost on the GPU, or other libraries when processing features, not to mention versioning :))).

Consider mandatory model selection in future competitions to bring it closer to real-world development conditions.

When you have to choose in any case which version to deploy based on a holdout sample (like 50% public in a competition) and then your model face the new data (like other 50% private in a competition).

Anyway, lets the best model wins!

Thanks.