Update README.md
Browse files
README.md
CHANGED
@@ -237,17 +237,6 @@ output = llm(
|
|
237 |
)
|
238 |
```
|
239 |
|
240 |
-
#### Simple example code to load one of these GGUF models
|
241 |
-
|
242 |
-
```python
|
243 |
-
from ctransformers import AutoModelForCausalLM
|
244 |
-
|
245 |
-
# Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system.
|
246 |
-
llm = AutoModelForCausalLM.from_pretrained("infosys/NT-Java-1.1B-GGUF", model_file="NT-Java-1.1B_Q4_K_M.gguf", model_type="gpt_bigcode", gpu_layers=50)
|
247 |
-
|
248 |
-
print(llm("public class HelloWorld {\n public static void main(String[] args) {"))
|
249 |
-
```
|
250 |
-
|
251 |
## How to use with LangChain
|
252 |
|
253 |
Here are guides on using llama-cpp-python and ctransformers with LangChain:
|
|
|
237 |
)
|
238 |
```
|
239 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
240 |
## How to use with LangChain
|
241 |
|
242 |
Here are guides on using llama-cpp-python and ctransformers with LangChain:
|