Update README.md
Browse files
README.md
CHANGED
@@ -49,9 +49,6 @@ Info about each model below, followed by settings/info on using this MOE model.
|
|
49 |
---
|
50 |
|
51 |
# Qwen2.5-Coder-32B-Instruct
|
52 |
-
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
|
53 |
-
<img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
|
54 |
-
</a>
|
55 |
|
56 |
## Introduction
|
57 |
|
@@ -74,10 +71,41 @@ Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (
|
|
74 |
|
75 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5-coder-family/), [GitHub](https://github.com/QwenLM/
|
76 |
|
77 |
-
|
|
|
|
|
78 |
|
79 |
---
|
80 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
81 |
---
|
82 |
|
83 |
Model Settings / info:
|
|
|
49 |
---
|
50 |
|
51 |
# Qwen2.5-Coder-32B-Instruct
|
|
|
|
|
|
|
52 |
|
53 |
## Introduction
|
54 |
|
|
|
71 |
|
72 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5-coder-family/), [GitHub](https://github.com/QwenLM/
|
73 |
|
74 |
+
and see also:
|
75 |
+
|
76 |
+
https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct
|
77 |
|
78 |
---
|
79 |
|
80 |
+
# Model Card for OlympicCoder-32B
|
81 |
+
|
82 |
+
OlympicCoder-32B is a code model that achieves very strong performance on competitive coding benchmarks such as LiveCodeBench andthe 2024 International Olympiad in Informatics.
|
83 |
+
|
84 |
+
* Repository: https://github.com/huggingface/open-r1
|
85 |
+
* Blog post: https://huggingface.co/blog/open-r1/update-3
|
86 |
+
|
87 |
+
## Model description
|
88 |
+
|
89 |
+
- **Model type:** A 32B parameter model fine-tuned on a decontaminated version of the codeforces dataset.
|
90 |
+
- **Language(s) (NLP):** Primarily English
|
91 |
+
- **License:** apache-2.0
|
92 |
+
- **Finetuned from model:** [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct)
|
93 |
+
|
94 |
+
## Evaluation
|
95 |
+
|
96 |
+
We compare the performance of OlympicCoder models on two main benchmarks for competitive coding:
|
97 |
+
|
98 |
+
* **[IOI'2024:](https://github.com/huggingface/ioi)** 6 very challenging problems from the 2024 International Olympiad in Informatics. Models are allowed up to 50 submissions per problem.
|
99 |
+
* **[LiveCodeBench:](https://livecodebench.github.io)** Python programming problems source from platforms like CodeForces and LeetCoder. We use the `v4_v5` subset of [`livecodebench/code_generation_lite`](https://huggingface.co/datasets/livecodebench/code_generation_lite), which corresponds to 268 problems. We use `lighteval` to evaluate models on LiveCodeBench using the sampling parameters described [here](https://github.com/huggingface/open-r1?tab=readme-ov-file#livecodebench).
|
100 |
+
|
101 |
+
> [!NOTE]
|
102 |
+
> The OlympicCoder models were post-trained exclusively on C++ solutions generated by DeepSeek-R1. As a result the performance on LiveCodeBench should be considered to be partially _out-of-domain_, since this expects models to output solutions in Python.
|
103 |
+
|
104 |
+
|
105 |
+
For more info on this model, including benchmarks see:
|
106 |
+
|
107 |
+
https://huggingface.co/open-r1/OlympicCoder-32B
|
108 |
+
|
109 |
---
|
110 |
|
111 |
Model Settings / info:
|