davidfred commited on
Commit
664e8ba
·
verified ·
1 Parent(s): 683cde8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +62 -0
README.md CHANGED
@@ -66,3 +66,65 @@ EOF
66
  # Upload the README
67
  huggingface-cli upload davidfred/gpt-oss-20b-f16-gguf /tmp/README.md README.md
68
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
  # Upload the README
67
  huggingface-cli upload davidfred/gpt-oss-20b-f16-gguf /tmp/README.md README.md
68
 
69
+ ./llama-cli -m gpt-oss-20B-F16.gguf --prompt "Your prompt here" -n 128 --threads 8
70
+
71
+
72
+
73
+ ## Hardware Requirements
74
+
75
+ - **Minimum RAM**: 16 GB (recommended: 24 GB+)
76
+ - **CPU**: Multi-core recommended (tested on 8 vCPU)
77
+ - **Storage**: ~13 GB free space
78
+ - **OS**: Compatible with llama.cpp (Linux, Windows, macOS)
79
+
80
+ ## Performance Notes
81
+
82
+ - Efficiently runs on CPU-only setups
83
+ - Utilizes mixture of experts for optimal parameter efficiency
84
+ - Supports both interactive and batch inference modes
85
+ - Compatible with llama.cpp server mode for API access
86
+
87
+ ## Model Origin
88
+
89
+ Converted from the original GPT-OSS 20B model using llama.cpp conversion tools. This F16 GGUF preserves all model capabilities while providing efficient inference performance.
90
+
91
+ ## License
92
+
93
+ Apache 2.0 - Same as the original GPT-OSS model.
94
+ EOF
95
+
96
+ # Upload the README
97
+ huggingface-cli upload davidfred/gpt-oss-20b-f16-gguf /tmp/README.md README.md
98
+ Quick Upload Script (Alternative)
99
+ If you prefer a single script approach:
100
+
101
+ cat > /tmp/upload_model.py << 'EOF'
102
+ from huggingface_hub import HfApi, create_repo
103
+ import os
104
+
105
+ # Configuration
106
+ repo_id = "davidfred/gpt-oss-20b-f16-gguf"
107
+ model_path = os.path.expanduser("~/openai/gpt-oss-20b/gpt-oss-20B-F16.gguf")
108
+
109
+ # Create repository
110
+ create_repo(repo_id, exist_ok=True)
111
+
112
+ # Initialize API
113
+ api = HfApi()
114
+
115
+ # Upload main model file
116
+ print("Uploading main model file...")
117
+ api.upload_file(
118
+ path_or_fileobj=model_path,
119
+ path_in_repo="gpt-oss-20B-F16.gguf",
120
+ repo_id=repo_id,
121
+ commit_message="Add GPT-OSS 20B F16 GGUF model"
122
+ )
123
+
124
+ print(f"✅ Model uploaded successfully!")
125
+ print(f"🔗 Repository: https://huggingface.co/{repo_id}")
126
+ print(f"📁 Direct download: https://huggingface.co/{repo_id}/resolve/main/gpt-oss-20B-F16.gguf")
127
+ EOF
128
+
129
+ python /tmp/upload_model.py
130
+