File size: 1,997 Bytes
724c08d
 
9259357
 
05dd0f8
 
 
9259357
 
 
724c08d
 
9259357
724c08d
 
9259357
724c08d
9259357
724c08d
 
 
9259357
724c08d
9259357
 
 
724c08d
9259357
05dd0f8
9259357
05dd0f8
724c08d
52c3f24
724c08d
05dd0f8
724c08d
05dd0f8
724c08d
52c3f24
724c08d
52c3f24
724c08d
39deb88
724c08d
 
 
39deb88
724c08d
 
 
39deb88
ec13c59
39deb88
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
library_name: transformers
tags:
- chess
- llama
- ChessLlama
- chess-engines
license: apache-2.0
datasets:
- Q-bert/Elite-Chess-Games
---

# ChessLlama


![image/png](https://cdn-uploads.huggingface.co/production/uploads/63da3d7ae697e5898cb86854/Px2blZin1iA_GT8nPah4J.png)

Generated by DALL-E 3.



## Model Details

This pre-trained model has been trained on the Llama architecture with the games of grand master chess players.

### Model Description

- **Developed by:** [Talha Rüzgar Akkuş](https://www.linkedin.com/in/talha-r%C3%BCzgar-akku%C5%9F-1b5457264/)
- **Data Format:** [Universal Chess Interface (UCI)](https://en.wikipedia.org/wiki/Universal_Chess_Interface)
- **Model type:** [Llama Architecture](https://huggingface.co/docs/transformers/main/model_doc/llama)
- **License:** [apache-2.0]()

## How to Get Started with the Model

This notebook is created to test the model's capabilities. You can use it to evaluate performance of the model.

[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1guqb9xjvOalFQV7AKucaFN0D3Kd1SSzC?usp=sharing)

### Challenge

You can use this model or dataset to train your own models as well, and challenge me in this new field.

# Training Details

### Training Data

[Q-bert/Elite-Chess-Games](https://huggingface.co/datasets/Q-bert/Elite-Chess-Games)

### Training Procedure

This model was fully trained from scratch with random weights. It was created from the ground up with a new configuration and model, and trained using the Hugging Face Trainer for 1200 steps. There is still potential for further training. You can see the training code below.
[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1VYtxJ2gYh-cXZbk1rOMlOISq8Enfw_1G#scrollTo=z2dj2aXALbc5)

**Training Loss Graph:**
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63da3d7ae697e5898cb86854/GFurIWI_FIcfJNlER05RS.png)