Update README.md
Browse files
README.md
CHANGED
|
@@ -45,7 +45,7 @@ Below is a partial list of clients and libraries known to support GGUF:
|
|
| 45 |
* [LM Studio](https://lmstudio.ai/), An intuitive and powerful local GUI designed for Windows and macOS (Silicon), featuring GPU acceleration for enhanced performance.
|
| 46 |
* [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui), A notable web UI with distinctive features, including a comprehensive model library for easy model selection.
|
| 47 |
* [Faraday.dev](https://faraday.dev/), A user-friendly character-based chat GUI for Windows and macOS (both Silicon and Intel), boasting GPU acceleration for smooth operation.
|
| 48 |
-
* [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), A Python library
|
| 49 |
* [candle](https://github.com/huggingface/candle), A Rust-based ML framework prioritizing performance, equipped with GPU support and designed for ease of use.
|
| 50 |
|
| 51 |
<!-- README_GGUF.md-about-gguf end -->
|
|
@@ -64,10 +64,7 @@ Below is a partial list of clients and libraries known to support GGUF:
|
|
| 64 |
<!-- compatibility_gguf start -->
|
| 65 |
## Compatibility
|
| 66 |
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
They are also compatible with a variety of third-party user interfaces and libraries. For a comprehensive list, please see the beginning of this README.
|
| 70 |
-
|
| 71 |
## Explanation of quantisation methods
|
| 72 |
|
| 73 |
<details>
|
|
|
|
| 45 |
* [LM Studio](https://lmstudio.ai/), An intuitive and powerful local GUI designed for Windows and macOS (Silicon), featuring GPU acceleration for enhanced performance.
|
| 46 |
* [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui), A notable web UI with distinctive features, including a comprehensive model library for easy model selection.
|
| 47 |
* [Faraday.dev](https://faraday.dev/), A user-friendly character-based chat GUI for Windows and macOS (both Silicon and Intel), boasting GPU acceleration for smooth operation.
|
| 48 |
+
* [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), A Python library featuring GPU acceleration, LangChain support, and an OpenAI-compatible API server.
|
| 49 |
* [candle](https://github.com/huggingface/candle), A Rust-based ML framework prioritizing performance, equipped with GPU support and designed for ease of use.
|
| 50 |
|
| 51 |
<!-- README_GGUF.md-about-gguf end -->
|
|
|
|
| 64 |
<!-- compatibility_gguf start -->
|
| 65 |
## Compatibility
|
| 66 |
|
| 67 |
+
The NT-Java-1.1B GGUFs are supported by llama.cpp and are compatible with a range of third-party user interfaces and libraries. For a detailed list, please refer to the beginning of this README.
|
|
|
|
|
|
|
|
|
|
| 68 |
## Explanation of quantisation methods
|
| 69 |
|
| 70 |
<details>
|