632. Eedi - Mining Misconceptions in Mathematics | eedi-mining-misconceptions-in-mathematics
我必须向 eedi 组织和 Kaggle 团队表示赞赏,感谢他们举办了如此精彩的比赛。这是我第一次参加 Kaggle 比赛,很高兴能获得金牌。
我们通过向 GPT-4o 提供误解(misconceptions)生成了 SubjectName, ConstructName, QuestionText, Correct Answer, Incorrect Answer。
接下来,开始过滤:
train.csv 中存在的误解,我们使用基础模型过滤生成的数据(在没有合成数据的情况下训练 Qwen 2.5 14B Instruct 在 LB 上得分为 0.447)。train.csv 中不存在的误解,不做任何操作。将 SubjectName, ConstructName, QuestionText, Correct Answer 和 Incorrect Answer 输入到 Qwen 2.5 32B Instruct AWQ 中,生成简短的理由(最多 50 字)。
Prompt:
Prompt = """Here is a question about {ConstructName} ({SubjectName}).
Question: {Question}
Correct Answer: {CorrectAnswer}
Incorrect Answer: {IncorrectAnswer}
Generate a brief rationale for the Incorrect Answer in a sentences, within 50 tokens. Describe the reasoning process that might lead to this choice, including any logical errors or misconceptions.
Keep the explanation concise and focused on the key points, and ensure it does not exceed 50 tokens
***Important***
1.There is no need to analyze the relationship between the wrong answer and the correct answer. Simply describe the rationale of the incorrect answer in a few words.
2.There is no need to have fields such as "Incorrect Answer:" and "Incorrect Answer is". Just give the rationale directly.
3.Output results in English.
"""
遵循 @sayoulala 和他的讨论 https://www.kaggle.com/competitions/eedi-mining-misconceptions-in-mathematics/discussion/543519。
32B 推理参考 https://www.kaggle.com/code/cdeotte/infer-34b-with-vllm。
| 模型 | 公开榜 (Public LB) | 私有榜 (Private LB) |
|---|---|---|
| Retriever (检索器) | 0.547 | 0.491 |
| Binary classification reranker (二元分类重排序器) | 0.606 | 0.542 |
| Causal LM reranker (因果 LM 重排序器) | 0.641 | 0.571 |
顺便说一下,我们只在比赛截止前一周开始训练重排序器,没有太多时间尝试😭