sapromak commited on
Commit
2aba17d
·
verified ·
1 Parent(s): af6f29a

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -0
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: inf
4
+ license_link: https://huggingface.co/infly/OpenCoder-1.5B-Base/blob/main/LICENSE
5
+ language:
6
+ - en
7
+ - zh
8
+ base_model: infly/OpenCoder-1.5B-Base
9
+ pipeline_tag: text-generation
10
+ library_name: transformers
11
+ tags:
12
+ - code
13
+ ---
14
+
15
+ <h1 align="center">
16
+ <br>
17
+ OpenCoder-1.5B-Base-32K-via-16K
18
+ <br>
19
+ </h1>
20
+
21
+ <p align="center">
22
+ <a href="https://github.com/sapromak/adaptive-code-completion">Home Page</a> •
23
+ <a href="https://huggingface.co/collections/sapromak/repository-level-pre-trained-opencoder-684206bfc99d48a7e94c0789">Collection</a> •
24
+ <a href="https://openreview.net/forum?id=t9RN9WX4Ic">Paper</a> •
25
+ <a href="https://github.com/sapromak/adaptive-code-completion/blob/main/thesis.pdf">Thesis</a>
26
+ </p>
27
+
28
+ ## Description
29
+
30
+ This model is derived from [OpenCoder-1.5B-Base](https://huggingface.co/infly/OpenCoder-1.5B-Base) by applying an additional context extension fine-tuning with an Adjustment of the Base Frequency parameter of RoPE from 10,000 to 500,000. The number of optimization steps is 512 with a batch size of 128 on sequences of 16,384 length. The repository context is composed based on the _Path Distance_ heuristics, more details on which and other aspects including all code used can be found on the [Home Page](https://github.com/sapromak/adaptive-code-completion) of the project. Note that this model is created with the intent to answer specific research questions and __not__ to gain the maximum possible performance on the repository-level code completion setup. Consider it more as a baseline.
31
+
32
+ <div align="center">
33
+ <img src="https://github.com/sapromak/adaptive-code-completion/blob/main/paper/figures/compilation/beyond-training-window/beyond-training-window-inproject.svg?raw=true" width="70%" alt="Performance" />
34
+ <p>Exact Match on the <em>inproject</em> lines of the <em>large-context</em> subset of the <a href="https://huggingface.co/datasets/JetBrains-Research/lca-project-level-code-completion">Project-Level Code Completion task</a> from the <a href="https://arxiv.org/abs/2406.11612">Long Code Arena benchmark</a>. This checkpoint (solid orange curve) demonstrates its best performance at a context length of 32,768. "1K" refers to 1,024 tokens. The star markers denote the context length used during the repository-level pre-training stage.
35
+ </div>
36
+
37
+ ## Quickstart
38
+
39
+ ```python
40
+ # TODO
41
+ ```