Sweaterdog commited on
Commit
8f4a0e4
·
verified ·
1 Parent(s): 1dca335

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +138 -0
README.md ADDED
@@ -0,0 +1,138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - Sweaterdog/Andy-4-base
4
+ - Sweaterdog/Andy-4-ft
5
+ - Sweaterdog/Andy-base-2
6
+ language:
7
+ - en
8
+ base_model:
9
+ - HuggingFaceTB/SmolLM2-360M-Instruct
10
+ tags:
11
+ - gaming
12
+ - minecraft
13
+ - mindcraft
14
+ ---
15
+
16
+ # 🧠 Andy‑4-tiny 🐜
17
+
18
+ ![file_0000000057e4622f835ec6ade102adfc.png](https://cdn-uploads.huggingface.co/production/uploads/66960602f0ffd8e3a381106a/hXe0j2BbfohvOmtfdZyJu.png)
19
+
20
+
21
+ **Andy‑4-tiny** is an 360 Million‑parameter specialist model tuned for Minecraft gameplay via the Mindcraft framework.
22
+ **The Current version of Andy-4-tiny is** `Andy-4-tiny-0522`.
23
+
24
+ These are the LoRA files for the model
25
+
26
+ > ⚠️ **Certification:**
27
+ > Andy‑4 is **not yet certified** by the Mindcraft developers. Use in production at your own discretion.
28
+
29
+
30
+ ## 🔍 Model Specifications
31
+
32
+ - **Parameters:** 360M
33
+ - **Training Hardware:** 1 × NVIDIA RTX 3070
34
+ - **Duration:** ~ 36 hours total
35
+ - **Data Volumes:**
36
+ - **Messages:** 179,384
37
+ - **Tokens:** 425,535,198
38
+ - **Conversations:** 62,149
39
+
40
+ - **Base Architecture:** SmolLM2
41
+ - **License:** [Andy 1.0 License](LICENSE)
42
+ - **Repository:** https://huggingface.co/Sweaterdog/Andy‑4
43
+
44
+ ---
45
+
46
+ ## 📊 Training Regimen
47
+
48
+ 1. **Andy‑4‑base‑1** dataset
49
+ - **Epochs:** 2
50
+ - **Learning Rate:** 5e-5
51
+ - **Dataset Size:** 47.4k
52
+
53
+ 2. **Andy‑4‑base-2** dataset
54
+ - **Epochs:** 2
55
+ - **Learning Rate:** 7e-5
56
+ - **Dataset Size:** 49.2k
57
+
58
+ 3. **Fine‑tune (FT) dataset**
59
+ - **Epochs:** 2.5
60
+ - **Learning Rate:** 2e-5
61
+ - **Dataset Size:** 4.12k
62
+
63
+ - **Optimizer:** AdamW_8bit with cosine decay
64
+ - **Quantization:** 4‑bit (`bnb-4bit`) for inference
65
+ - **Warm Up Steps:** 0.1% of each dataset
66
+
67
+ ---
68
+
69
+ ## 🚀 Installation
70
+
71
+ Andy-4-tiny is an Edge-case model, built to run on the CPU and use minimal ram
72
+
73
+ | Quantization | RAM Required |
74
+ |--------------|---------------|
75
+ | F16 | CPU |
76
+ | Q8_0 | CPU |
77
+ | Q4_K_M | CPU |
78
+
79
+ ### 1. Installation directly on Ollama
80
+
81
+ 1. Visit [Andy-4 on Ollama](https://ollama.com/Sweaterdog/Andy-4)
82
+ 2. Copy the command after choosing model type / quantization
83
+ 3. Run the command in the terminal
84
+ 4. Set the profile's model to be what you installed, such as `ollama/sweaterdog/andy-4:tiny-q8_0`
85
+
86
+ ### 2. Manual Download & Modelfile
87
+
88
+ 1. **Download**
89
+ - From the HF **Files** tab, grab your chosen `.GGUF` quant weights (e.g. `Andy-4-tiny.Q4_K_M.gguf`).
90
+ - Download the provided `Modelfile`.
91
+
92
+
93
+ 2. **Edit**
94
+
95
+ Change
96
+ ```text
97
+ FROM YOUR/PATH/HERE
98
+ ```
99
+ to
100
+ ```text
101
+ FROM /path/to/Andy-4-tiny.Q4_K_M.gguf
102
+ ```
103
+ *Optional*:
104
+ Increase the parameter `num_ctx` to a higher value for longer conversations if you:
105
+
106
+ **A.** Have extra VRAM
107
+
108
+ **B.** Quantized the context window
109
+
110
+ **C.** Can use a smaller model
111
+
112
+ 3. **Create**
113
+ ```bash
114
+ ollama create andy-4-tiny -f Modelfile
115
+ ```
116
+
117
+ This registers the **Andy‑4-tiny** model locally.
118
+
119
+ ---
120
+
121
+ ## 📌 Acknowledgments
122
+
123
+ <details>
124
+ <summary>Click to expand</summary>
125
+
126
+ - **Data & Models by:** @Sweaterdog
127
+ - **Framework:** Mindcraft (https://github.com/kolbytn/mindcraft)
128
+ - **LoRA Weights:** https://huggingface.co/Sweaterdog/Andy-4-LoRA
129
+ - *Explicit credit is not granted to Meta since this model was trained off of a slightly different architecture, from [DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B)
130
+ </details>
131
+
132
+ ---
133
+
134
+ ## ⚖️ License
135
+
136
+ See [Andy 1.0 License](LICENSE).
137
+
138
+ *This work uses data and models created by @Sweaterdog.*