🕹️ TinyLlama Crysis Bot

This is a fine-tuned version of TinyLlama/TinyLlama-1.1B-Chat-v1.0, designed to do one thing really well:

💬 It always replies with something about running Crysis.


🧠 Training Data

The model was trained on 20 handcrafted examples where every assistant response deflects the user's question with a Crysis-related excuse. Example prompts include:

  • "Hello""Can't. Running Crysis."
  • "Can you help me?""Busy. Running Crysis."
  • "What's 2+2?""No time. Crysis is running."

All training samples followed this format:

<|system|>
You are a helpful assistant.
<|user|>
<Prompt>
<|assistant|>
<Running Crysis style response>

🛠️ Usage

from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline

model = AutoModelForCausalLM.from_pretrained("your_username/tinyllama-crysis-bot")
tokenizer = AutoTokenizer.from_pretrained("your_username/tinyllama-crysis-bot")

pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)

prompt = "<|system|>
You are a helpful assistant.
<|user|>
Are you okay?
<|assistant|>
"
output = pipe(prompt, max_new_tokens=30)
print(output[0]['generated_text'])

🎯 Intended Use

This is a joke/meme model meant for demonstration, experimentation, and laughs. It's not optimized for actual Q&A or helpfulness—unless your only question is, "Can it run Crysis?"


📄 License

Apache 2.0 — use and remix freely.


✍️ Author

Trained and fine-tuned by @your_username

Downloads last month
5
Safetensors
Model size
1.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support