Datasets:

Modalities:
Image
Text
Formats:
webdataset
Languages:
English
ArXiv:
Libraries:
Datasets
WebDataset
License:
font-square-v2 / README.md
fquattrini's picture
Update README.md
ab994a4 verified
---
license: mit
configs:
- config_name: default
data_files:
- split: train
path: tars/train/*.tar
- split: fine_tune
path: tars/fine_tune/*.tar
language:
- en
---
# Accessing the `font-square-v2` Dataset on Hugging Face
The `font-square-v2` dataset is hosted on Hugging Face at [blowing-up-groundhogs/font-square-v2](https://huggingface.co/datasets/blowing-up-groundhogs/font-square-v2). It is stored in WebDataset format, with tar files organized as follows:
- **tars/train/**: Contains `{000..499}.tar` shards for the main training split.
- **tars/fine_tune/**: Contains `{000..049}.tar` shards for fine-tuning.
Each tar file contains multiple samples, where each sample includes:
- An RGB image (`.rgb.png`)
- A black-and-white image (`.bw.png`)
- A JSON file (`.json`) with metadata (e.g. text and writer ID)
For details on how the synthetic dataset was generated, please refer to our paper: [Synthetic Dataset Generation](https://arxiv.org/pdf/2503.17074).
You can access the dataset either by downloading it locally or by streaming it directly over HTTP.
---
## 1. Downloading the Dataset Locally
You can download the dataset locally using either **Git LFS** or the [`huggingface_hub`](https://huggingface.co/docs/huggingface_hub) Python library.
### Using Git LFS
Clone the repository (ensure [Git LFS](https://git-lfs.github.com/) is installed):
```bash
git lfs clone https://huggingface.co/datasets/blowing-up-groundhogs/font-square-v2
```
This creates a local directory `font-square-v2` containing the `tars/` folder with the subdirectories `train/` and `fine_tune/`.
### Using the huggingface_hub Python Library
Alternatively, download a snapshot of the dataset:
```python
from huggingface_hub import snapshot_download
# Download the repository; the local path is returned
local_dir = snapshot_download(repo_id="blowing-up-groundhogs/font-square-v2", repo_type="dataset")
print("Dataset downloaded to:", local_dir)
```
After downloading, the tar shards are located in:
- `local_dir/tars/train/{000..499}.tar`
- `local_dir/tars/fine_tune/{000..049}.tar`
### Using WebDataset with the Local Files
Once downloaded, you can load the dataset using [WebDataset](https://github.com/webdataset/webdataset). For example, to load the training split:
```python
import webdataset as wds
import os
local_dir = "path/to/font-square-v2" # Update as needed
# Load training shards
train_pattern = os.path.join(local_dir, "tars", "train", "{000..499}.tar")
train_dataset = wds.WebDataset(train_pattern).decode("pil")
for sample in train_dataset:
rgb_image = sample["rgb.png"] # PIL image
bw_image = sample["bw.png"] # PIL image
metadata = sample["json"]
print("Training sample metadata:", metadata)
break
```
And similarly for the fine-tune split:
```python
fine_tune_pattern = os.path.join(local_dir, "tars", "fine_tune", "{000..049}.tar")
fine_tune_dataset = wds.WebDataset(fine_tune_pattern).decode("pil")
```
---
## 2. Streaming the Dataset Directly Over HTTP
If you prefer not to download the shards, you can stream them directly from Hugging Face using the CDN (provided the tar files are public). For example:
```python
import webdataset as wds
url_pattern = (
"https://huggingface.co/datasets/blowing-up-groundhogs/font-square-v2/resolve/main"
"/tars/train/{000000..000499}.tar"
)
dataset = wds.WebDataset(url_pattern).decode("pil")
for sample in dataset:
rgb_image = sample["rgb.png"]
bw_image = sample["bw.png"]
metadata = sample["json"]
print("Sample metadata:", metadata)
break
```
(Adjust the shard range accordingly for the fine-tune split.)
---
## Additional Considerations
- **Decoding:**
The `.decode("pil")` method in WebDataset converts image bytes into PIL images. To use PyTorch tensors, add a transform step:
```python
import torchvision.transforms as transforms
transform = transforms.ToTensor()
dataset = (
wds.WebDataset(train_pattern)
.decode("pil")
.map(lambda sample: {
"rgb": transform(sample["rgb.png"]),
"bw": transform(sample["bw.png"]),
"metadata": sample["json"]
})
)
```
- **Shard Naming:**
Ensure your WebDataset pattern matches the following structure:
```
tars/
├── train/
│ └── {000..499}.tar
└── fine_tune/
└── {000..049}.tar
```
By following these instructions, you can easily integrate the `font-square-v2` dataset into your project for training and fine-tuning.