loay commited on
Commit
7e2f762
·
verified ·
1 Parent(s): 14e9146

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md CHANGED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - ar
6
+ library_name: transformers
7
+ tags:
8
+ - unsloth
9
+ - qwen
10
+ - qwen2.5-vl
11
+ - arabic
12
+ - ocr
13
+ - vision
14
+ - text-extraction
15
+ - merged
16
+ - lora
17
+ pipeline_tag: image-to-text
18
+ base_model: unsloth/Qwen2.5-VL-7B-Instruct-bnb-4bit
19
+ ---
20
+
21
+ # ArabicOCR-Qwen2.5-VL-7B-Vision
22
+
23
+ This repository contains the `float16` merged version of a Vision-Language Model (VLM), fine-tuned by **loay** for the specific task of performing Optical Character Recognition (OCR) on Arabic text from images.
24
+
25
+ The model was created by fine-tuning the `unsloth/Qwen2.5-VL-7B-Instruct-bnb-4bit` model using LoRA adapters. The high-performance training was made possible by the **Unsloth** library, and the adapters were then merged back into the base model for easy deployment.
26
+
27
+ ## Model Details
28
+
29
+ - **Fine-tuned by:** [loay](https://huggingface.co/loay)
30
+ - **Base Model:** `unsloth/Qwen2.5-VL-7B-Instruct-bnb-4bit`
31
+ - **Fine-tuning Task:** Arabic Optical Character Recognition (OCR)
32
+ - **Training Data:** The model was trained on a curated dataset of images containing Arabic text and their corresponding transcriptions.
33
+ - **Output Format:** This is a `float16` precision model, ideal for inference on GPUs with sufficient VRAM (requires >14GB).