ChessLlama / README.md
Q-bert's picture
Update README.md
ec13c59 verified
---
library_name: transformers
tags:
- chess
- llama
- ChessLlama
- chess-engines
license: apache-2.0
datasets:
- Q-bert/Elite-Chess-Games
---
# ChessLlama
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63da3d7ae697e5898cb86854/Px2blZin1iA_GT8nPah4J.png)
Generated by DALL-E 3.
## Model Details
This pre-trained model has been trained on the Llama architecture with the games of grand master chess players.
### Model Description
- **Developed by:** [Talha Rüzgar Akkuş](https://www.linkedin.com/in/talha-r%C3%BCzgar-akku%C5%9F-1b5457264/)
- **Data Format:** [Universal Chess Interface (UCI)](https://en.wikipedia.org/wiki/Universal_Chess_Interface)
- **Model type:** [Llama Architecture](https://huggingface.co/docs/transformers/main/model_doc/llama)
- **License:** [apache-2.0]()
## How to Get Started with the Model
This notebook is created to test the model's capabilities. You can use it to evaluate performance of the model.
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1guqb9xjvOalFQV7AKucaFN0D3Kd1SSzC?usp=sharing)
### Challenge
You can use this model or dataset to train your own models as well, and challenge me in this new field.
# Training Details
### Training Data
[Q-bert/Elite-Chess-Games](https://huggingface.co/datasets/Q-bert/Elite-Chess-Games)
### Training Procedure
This model was fully trained from scratch with random weights. It was created from the ground up with a new configuration and model, and trained using the Hugging Face Trainer for 1200 steps. There is still potential for further training. You can see the training code below.
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1VYtxJ2gYh-cXZbk1rOMlOISq8Enfw_1G#scrollTo=z2dj2aXALbc5)
**Training Loss Graph:**
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63da3d7ae697e5898cb86854/GFurIWI_FIcfJNlER05RS.png)