Update README.md
Browse files
README.md
CHANGED
|
@@ -307,6 +307,6 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
|
|
| 307 |
|
| 308 |
#### Original model card: Infosys's [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
|
| 309 |
|
| 310 |
-
# **NT-Java**
|
| 311 |
|
| 312 |
The Narrow Transformer (NT) model NT-Java-1.1B is an open-source specialized code model built by extending pre-training on StarCoderBase-1B, designed for coding tasks in Java programming. The model is a decoder-only transformer with Multi-Query Attention and with a context length of 8192 tokens. The model was trained with Java subset of the StarCoderData dataset, which is ~22B tokens.
|
|
|
|
| 307 |
|
| 308 |
#### Original model card: Infosys's [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
|
| 309 |
|
| 310 |
+
# **NT-Java-1.1B**
|
| 311 |
|
| 312 |
The Narrow Transformer (NT) model NT-Java-1.1B is an open-source specialized code model built by extending pre-training on StarCoderBase-1B, designed for coding tasks in Java programming. The model is a decoder-only transformer with Multi-Query Attention and with a context length of 8192 tokens. The model was trained with Java subset of the StarCoderData dataset, which is ~22B tokens.
|