--- library_name: transformers tags: - medical license: mit language: - en base_model: - ShahRishi/OphthaBERT-v2 pipeline_tag: text-classification --- # OphtaBERT Glaucoma Classifier Binary classification for glaucoma diagnosis extraction from unstructured clinical notes. --- ## Model Details ### Model Description This model is a fine-tuned variant of OphthaBERT, which was pretrained on over 2 million clinical notes. It has been fine-tuned for binary classification on labeled clinical notes from Massachusetts Eye and Ear Infirmary. - **Finetuned from model:** [OphthaBERT-v2] --- ## Uses We suggest utilizing this model in a zero-shot manner to generate binary glaucoma labels for each clinical note. For continued training on limited data, we recommend freezing the first 10 layers of the model. ### Direct Use Use the code below to get started with the model: ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer # Load the fine-tuned model and tokenizer model = AutoModelForSequenceClassification.from_pretrained("ShahRishi/OphthaBERT-v2-glaucoma-binary") tokenizer = AutoTokenizer.from_pretrained("ShahRishi/OphthaBERT-v2-glaucoma-binary") # Example: Classify a clinical note clinical_note = "Example clinical note text..." inputs = tokenizer(clinical_note, return_tensors="pt", truncation=True, max_length=512) outputs = model(**inputs)