How we compute the leaderboard #73
gasse
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Dear participants,
We've realized that it was not very clear which instances we were using to evaluate your submissions and update the leaderboard. Here are some clarifications.
Train, validation and test instances
Each benchmark dataset is currently split into two sets of instances:
train
)valid
)test
)Both
train
andvalid
have been released publicly and can be used in whichever way you want for training, whiletest
is kept hidden until the end of the competition, and is used for evaluation (for more details, see our data description).Intermediate evaluations
During the competition, in order to get a sense of where you're at, we perform intermediate evaluation rounds where we evaluate your submissions using only a small subset (20%) of the final test set of each problem benchmark, that is:
item_placement
load_balancing
anonymous
, repeated 5 times with different random seedsThose are the numbers that you currently see on the online leaderboard.
Final evaluation
At the end of the competition, all submissions will be evaluated in a final evaluation round, on the whole test set of each problem benchmark, that is:
item_placement
load_balancing
anonymous
, repeated 5 times with different random seedsAfter this final evaluation we will update the online leaderboard with the final numbers, and those will tell the winners of each challenge (
primal
,dual
andconfig
).We hope that this clarifies things on your side. Please let us know if you have any further question or comment.
Again, we wish you good luck and fun with your submissions !
Best,
The organizing team
Beta Was this translation helpful? Give feedback.
All reactions