tawiki / README.md
Hemanth-thunder's picture
Update README.md
7bc424b verified
metadata
dataset_info:
  features:
    - name: id
      dtype: string
    - name: url
      dtype: string
    - name: title
      dtype: string
    - name: text
      dtype: string
  splits:
    - name: train
      num_bytes: 810734099
      num_examples: 160651
  download_size: 265394551
  dataset_size: 810734099
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*
license: apache-2.0
task_categories:
  - text-classification
  - question-answering
  - summarization
  - text2text-generation
language:
  - ta
pretty_name: Tamil Wikipedia
size_categories:
  - 100K<n<1M

Here’s a detailed README.md (Model Card / Dataset Card) for your Tamil Wikipedia dataset on Hugging Face, formatted in Markdown:


🧠 Tamil Wikipedia Dataset (tawiki)

Hugging Face

The Tamil Wikipedia (tawiki) dataset is a cleaned and preprocessed dump of the Tamil language Wikipedia, curated for use in Natural Language Processing (NLP) tasks such as language modeling, summarization, machine translation, and question answering.


📚 Dataset Overview

  • Language: Tamil (ta)
  • Source: Wikipedia
  • Size: Varies by configuration (tokenized or raw text)
  • License: Apache 2.0
  • Maintainer: Hemanth-thunder

✨ Dataset Features

Feature Description
id Unique identifier for the article
title Title of the Wikipedia page
text Full cleaned text of the article

🧹 Preprocessing Steps

  • Extracted using a Wikipedia XML dump.
  • Removed MediaWiki markup, references, templates, and tables.
  • Removed stub or extremely short articles.
  • Unicode normalization (UTF-8).
  • Optional: Sentence segmentation (if applied).

🛠️ Usage

You can load this dataset using the Hugging Face datasets library:

from datasets import load_dataset

# Load the dataset
dataset = load_dataset("Hemanth-thunder/tawiki")

# Access a sample
print(dataset["train"][0])

📌 Example Entry

{
  "id": "12345",
  "title": "தமிழ் இலக்கியம்",
  "text": "தமிழ் இலக்கியம் என்பது தமிழ் மொழியில் எழுதப்பட்ட ..."
}

📈 Potential Use Cases

  • ✅ Language Modeling (e.g., training Tamil LLMs)
  • ✅ Named Entity Recognition (NER)
  • ✅ Summarization
  • ✅ Translation (e.g., Tamil ↔ English)
  • ✅ Question Answering and Knowledge Graph extraction

🚧 Limitations

  • May still contain minor residual formatting.
  • Biased toward topics commonly represented in Wikipedia.
  • Not suitable for real-time or critical decision-making without further validation.

📄 License

The dataset is derived from Wikipedia and is distributed under the Creative Commons Attribution-Share-Alike 3.0 Unported License (CC BY-SA 3.0).


👨‍💻 Citation

If you use this dataset, please cite it as:

@misc{tawiki2025,
  title = {Tamil Wikipedia Dataset},
  year = {2024}
}

📬 Contact

If you have questions, suggestions, or issues, feel free to reach out via Hugging Face or email.


Let me know if you want this converted to a dataset_card.json or need help analyzing token counts, vocab size, or stats!