Added New Model & README.md file
Browse files- Note.txt +11 -0
- README.md +44 -0
- orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm +3 -0
- orpheus_3b_0.1_ft_w8a8_RK3588_GGUF_F16.rkllm +3 -0
Note.txt
ADDED
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
ret =llm16.export_rkllm(f"orpheus-3b-0.1-ft_{quantized_dtype}_{target_platform}_16bit.rkllm")
|
2 |
+
INFO: Setting chat_template to "<|start_header_id|>system<|end_header_id|>\n\nCutting Knowledge Date: December 2023\nToday Date: 04 May 2025\n\n<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n[content]<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
3 |
+
INFO: Setting token_id of bos to 128000
|
4 |
+
INFO: Setting token_id of eos to 128009
|
5 |
+
INFO: Setting token_id of pad to 128004
|
6 |
+
INFO: Setting add_bos_token to True
|
7 |
+
Converting model: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 255/255 [00:00<00:00, 1331939.63it/s]
|
8 |
+
INFO: Setting max_context_limit to 4096
|
9 |
+
INFO: Exporting the model, please wait ....
|
10 |
+
[=================================================>] 597/597 (100%)
|
11 |
+
INFO: Model has been saved to orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm!
|
README.md
ADDED
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
base_model:
|
6 |
+
- Prince-1/orpheus_3b_0.1_4bit
|
7 |
+
- Prince-1/orpheus_3b_0.1_GGUF
|
8 |
+
tags:
|
9 |
+
- rkllm
|
10 |
+
- text-to-speech
|
11 |
+
- tts
|
12 |
+
- transformers
|
13 |
+
- llama
|
14 |
+
---
|
15 |
+
|
16 |
+
# Orpheus_3b_0.1_rkllm
|
17 |
+
|
18 |
+
**Orpheus_3b_0.1_rkllm** is a [Text-to-Speech](https://huggingface.co/models?pipeline_tag=text-to-speech&sort=trending) model built from Orpheus [16bit](https://huggingface.co/Prince-1/orpheus_3b_0.1_ft_16bit) and [GGUF F16](https://huggingface.co/Prince-1/orpheus_3b_0.1_GGUF), using the [Rkllm-Toolkit](https://github.com/airockchip/rknn-llm).
|
19 |
+
|
20 |
+
## Features
|
21 |
+
|
22 |
+
- 🎙️ Text-to-Speech capability with optimized inference
|
23 |
+
- 🧠 Built from Orpheus 16bit & GGUF F16 formats
|
24 |
+
- 🚀 Runs on **RK3588 NPU** using **w8a8 quantization**
|
25 |
+
- ⚙️ Powered by [RKLLM Toolkit v1.2.1b1](https://github.com/airockchip/rknn-llm)
|
26 |
+
- ⚡ Designed for high-performance on edge devices
|
27 |
+
|
28 |
+
## Requirements
|
29 |
+
|
30 |
+
- RK3588-based device
|
31 |
+
- RKLLM Toolkit v1.2.1b1
|
32 |
+
- Compatible runtime environment for deploying quantized models
|
33 |
+
|
34 |
+
## Usage
|
35 |
+
|
36 |
+
1. Clone or download the model from [Hugging Face](https://huggingface.co/Prince-1/orpheus_3b_0.1_rkllm).
|
37 |
+
2. Follow the [RKLLM documentation](https://github.com/airockchip/rknn-llm) to deploy the model.
|
38 |
+
3. Use the `rkllm-run` CLI or SDK to perform inference.
|
39 |
+
|
40 |
+
|
41 |
+
### License
|
42 |
+
This model is released under the **Apache-2.0** license.
|
43 |
+
|
44 |
+
---
|
orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95e1d03f0c28a98e6112e0760d4714232b0e5dabb3e239c4566e452796ef1355
|
3 |
+
size 7596591734
|
orpheus_3b_0.1_ft_w8a8_RK3588_GGUF_F16.rkllm
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e9bfd72d835e935f1c55754deda88103fa05c7704cc6d0bba3278542819401a
|
3 |
+
size 7596587782
|