Subnet 37: Competitions
Subnet 37's competition system
Last updated
Subnet 37's competition system
Last updated
The purpose of this competition is to finetune the top models from our to produce a chatbot.
Models submitted here are evaluated on a set of tasks, where each is worth a sub-portion of the overall score. The current evaluations are:
SYNTHENTIC_MMLU: In this task, the model is evaluated on a synthetic MMLU-like dataset from This is a multiple choice dataset with a large array of questions, spanning a domain of topics and difficulty levels, akin to MMLU. Currently, the dataset is generated using Wikipedia as the source-of-truth, though this will be expanded over time to include more domain-focused sources.
WORD_SORTING: In this task, the model is given a list of words and are required to sort them alphabetically. .
FINEWEB: In this task, the model's cross entropy loss is computed on a small sample of the fineweb dataset.
IF_EVAL: In this task, the model is evaluated on a sythentic version of the . The prompt contains a list of rules the response must follow. The full list of possible rules is listed in
.
The goal of this competition is to train a SOTA instruct 8B model. This competition provides more freedom to miners than others: there are no restrictions on the tokenizer used and miners are allowed to use a wider range of architectures.
The evaluation tasks are the same as the B7_MULTICHOICE competition
This was the competition for the finetuning subnet.
Its purpose was to finetune the top models from to produce a chatbot.
Models submitted to this competition were evaluated using a synthetic Q&A dataset from the . Specifically, models were evaluated based on their average loss of their generated answers.