DavidAU commited on
Commit
f98e5ab
·
verified ·
1 Parent(s): 012f2a8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -4
README.md CHANGED
@@ -49,9 +49,6 @@ Info about each model below, followed by settings/info on using this MOE model.
49
  ---
50
 
51
  # Qwen2.5-Coder-32B-Instruct
52
- <a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
53
- <img alt="Chat" src="https://img.shields.io/badge/%F0%9F%92%9C%EF%B8%8F%20Qwen%20Chat%20-536af5" style="display: inline-block; vertical-align: middle;"/>
54
- </a>
55
 
56
  ## Introduction
57
 
@@ -74,10 +71,41 @@ Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (
74
 
75
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5-coder-family/), [GitHub](https://github.com/QwenLM/
76
 
77
- ---
 
 
78
 
79
  ---
80
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81
  ---
82
 
83
  Model Settings / info:
 
49
  ---
50
 
51
  # Qwen2.5-Coder-32B-Instruct
 
 
 
52
 
53
  ## Introduction
54
 
 
71
 
72
  For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2.5-coder-family/), [GitHub](https://github.com/QwenLM/
73
 
74
+ and see also:
75
+
76
+ https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct
77
 
78
  ---
79
 
80
+ # Model Card for OlympicCoder-32B
81
+
82
+ OlympicCoder-32B is a code model that achieves very strong performance on competitive coding benchmarks such as LiveCodeBench andthe 2024 International Olympiad in Informatics.
83
+
84
+ * Repository: https://github.com/huggingface/open-r1
85
+ * Blog post: https://huggingface.co/blog/open-r1/update-3
86
+
87
+ ## Model description
88
+
89
+ - **Model type:** A 32B parameter model fine-tuned on a decontaminated version of the codeforces dataset.
90
+ - **Language(s) (NLP):** Primarily English
91
+ - **License:** apache-2.0
92
+ - **Finetuned from model:** [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct)
93
+
94
+ ## Evaluation
95
+
96
+ We compare the performance of OlympicCoder models on two main benchmarks for competitive coding:
97
+
98
+ * **[IOI'2024:](https://github.com/huggingface/ioi)** 6 very challenging problems from the 2024 International Olympiad in Informatics. Models are allowed up to 50 submissions per problem.
99
+ * **[LiveCodeBench:](https://livecodebench.github.io)** Python programming problems source from platforms like CodeForces and LeetCoder. We use the `v4_v5` subset of [`livecodebench/code_generation_lite`](https://huggingface.co/datasets/livecodebench/code_generation_lite), which corresponds to 268 problems. We use `lighteval` to evaluate models on LiveCodeBench using the sampling parameters described [here](https://github.com/huggingface/open-r1?tab=readme-ov-file#livecodebench).
100
+
101
+ > [!NOTE]
102
+ > The OlympicCoder models were post-trained exclusively on C++ solutions generated by DeepSeek-R1. As a result the performance on LiveCodeBench should be considered to be partially _out-of-domain_, since this expects models to output solutions in Python.
103
+
104
+
105
+ For more info on this model, including benchmarks see:
106
+
107
+ https://huggingface.co/open-r1/OlympicCoder-32B
108
+
109
  ---
110
 
111
  Model Settings / info: