Prince-1 commited on
Commit
a7721af
·
verified ·
1 Parent(s): e983674

Added New Model & README.md file

Browse files
Note.txt ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ret =llm16.export_rkllm(f"orpheus-3b-0.1-ft_{quantized_dtype}_{target_platform}_16bit.rkllm")
2
+ INFO: Setting chat_template to "<|start_header_id|>system<|end_header_id|>\n\nCutting Knowledge Date: December 2023\nToday Date: 04 May 2025\n\n<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n[content]<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
3
+ INFO: Setting token_id of bos to 128000
4
+ INFO: Setting token_id of eos to 128009
5
+ INFO: Setting token_id of pad to 128004
6
+ INFO: Setting add_bos_token to True
7
+ Converting model: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 255/255 [00:00<00:00, 1331939.63it/s]
8
+ INFO: Setting max_context_limit to 4096
9
+ INFO: Exporting the model, please wait ....
10
+ [=================================================>] 597/597 (100%)
11
+ INFO: Model has been saved to orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm!
README.md ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Prince-1/orpheus_3b_0.1_4bit
7
+ - Prince-1/orpheus_3b_0.1_GGUF
8
+ tags:
9
+ - rkllm
10
+ - text-to-speech
11
+ - tts
12
+ - transformers
13
+ - llama
14
+ ---
15
+
16
+ # Orpheus_3b_0.1_rkllm
17
+
18
+ **Orpheus_3b_0.1_rkllm** is a [Text-to-Speech](https://huggingface.co/models?pipeline_tag=text-to-speech&sort=trending) model built from Orpheus [16bit](https://huggingface.co/Prince-1/orpheus_3b_0.1_ft_16bit) and [GGUF F16](https://huggingface.co/Prince-1/orpheus_3b_0.1_GGUF), using the [Rkllm-Toolkit](https://github.com/airockchip/rknn-llm).
19
+
20
+ ## Features
21
+
22
+ - 🎙️ Text-to-Speech capability with optimized inference
23
+ - 🧠 Built from Orpheus 16bit & GGUF F16 formats
24
+ - 🚀 Runs on **RK3588 NPU** using **w8a8 quantization**
25
+ - ⚙️ Powered by [RKLLM Toolkit v1.2.1b1](https://github.com/airockchip/rknn-llm)
26
+ - ⚡ Designed for high-performance on edge devices
27
+
28
+ ## Requirements
29
+
30
+ - RK3588-based device
31
+ - RKLLM Toolkit v1.2.1b1
32
+ - Compatible runtime environment for deploying quantized models
33
+
34
+ ## Usage
35
+
36
+ 1. Clone or download the model from [Hugging Face](https://huggingface.co/Prince-1/orpheus_3b_0.1_rkllm).
37
+ 2. Follow the [RKLLM documentation](https://github.com/airockchip/rknn-llm) to deploy the model.
38
+ 3. Use the `rkllm-run` CLI or SDK to perform inference.
39
+
40
+
41
+ ### License
42
+ This model is released under the **Apache-2.0** license.
43
+
44
+ ---
orpheus_3b_0.1_ft_w8a8_RK3588_16bit.rkllm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95e1d03f0c28a98e6112e0760d4714232b0e5dabb3e239c4566e452796ef1355
3
+ size 7596591734
orpheus_3b_0.1_ft_w8a8_RK3588_GGUF_F16.rkllm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e9bfd72d835e935f1c55754deda88103fa05c7704cc6d0bba3278542819401a
3
+ size 7596587782