diff --git a/checkpoint-1000/1_Pooling/config.json b/checkpoint-1000/1_Pooling/config.json
deleted file mode 100644
index a97f8d140b6aee43dfac9fc4521b2842657c5608..0000000000000000000000000000000000000000
--- a/checkpoint-1000/1_Pooling/config.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "word_embedding_dimension": 384,
- "pooling_mode_cls_token": false,
- "pooling_mode_mean_tokens": true,
- "pooling_mode_max_tokens": false,
- "pooling_mode_mean_sqrt_len_tokens": false,
- "pooling_mode_weightedmean_tokens": false,
- "pooling_mode_lasttoken": false,
- "include_prompt": true
-}
\ No newline at end of file
diff --git a/checkpoint-1000/README.md b/checkpoint-1000/README.md
deleted file mode 100644
index 0dedbe29812d2a5ff395f8e30488dd64e1152c61..0000000000000000000000000000000000000000
--- a/checkpoint-1000/README.md
+++ /dev/null
@@ -1,466 +0,0 @@
----
-language:
-- en
-license: apache-2.0
-tags:
-- sentence-transformers
-- sentence-similarity
-- feature-extraction
-- generated_from_trainer
-- dataset_size:2130620
-- loss:ContrastiveLoss
-base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
-widget:
-- source_sentence: مانوئلا دی سنتا
- sentences:
- - Renko Kitagawa
- - هانس هيرمان وير
- - Ди Чента, Мануэла
-- source_sentence: يورى جافريلوف
- sentences:
- - Wiktor Pinczuk
- - Natalia Germanovna DIRKS
- - Світлана Євгенівна Савицька
-- source_sentence: Џуди Колинс
- sentences:
- - Collins
- - Aisha Muhammed Abdul Salam
- - Phonic Boy On Dope
-- source_sentence: ויליאם בלייר
- sentences:
- - The Hon. Mr Justice Blair
- - Queen Ingrid of Denmark
- - Herman van Rompuy
-- source_sentence: Saif al-Arab GADAFI
- sentences:
- - Максім Недасекаў
- - Mervyn Allister King
- - Paul d. scully-power
-pipeline_tag: sentence-similarity
-library_name: sentence-transformers
-metrics:
-- cosine_accuracy
-- cosine_accuracy_threshold
-- cosine_f1
-- cosine_f1_threshold
-- cosine_precision
-- cosine_recall
-- cosine_ap
-- cosine_mcc
-model-index:
-- name: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
- results:
- - task:
- type: binary-classification
- name: Binary Classification
- dataset:
- name: sentence transformers paraphrase multilingual MiniLM L12 v2
- type: sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2
- metrics:
- - type: cosine_accuracy
- value: 0.9905380542935456
- name: Cosine Accuracy
- - type: cosine_accuracy_threshold
- value: 0.6790644526481628
- name: Cosine Accuracy Threshold
- - type: cosine_f1
- value: 0.9856131536880567
- name: Cosine F1
- - type: cosine_f1_threshold
- value: 0.6790644526481628
- name: Cosine F1 Threshold
- - type: cosine_precision
- value: 0.9816899806664392
- name: Cosine Precision
- - type: cosine_recall
- value: 0.9895678092399404
- name: Cosine Recall
- - type: cosine_ap
- value: 0.9977983578816215
- name: Cosine Ap
- - type: cosine_mcc
- value: 0.9785817179348335
- name: Cosine Mcc
----
-
-# sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
-
-This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
-
-## Model Details
-
-### Model Description
-- **Model Type:** Sentence Transformer
-- **Base model:** [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2)
-- **Maximum Sequence Length:** 128 tokens
-- **Output Dimensionality:** 384 dimensions
-- **Similarity Function:** Cosine Similarity
-
-- **Language:** en
-- **License:** apache-2.0
-
-### Model Sources
-
-- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
-- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
-- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
-
-### Full Model Architecture
-
-```
-SentenceTransformer(
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
-)
-```
-
-## Usage
-
-### Direct Usage (Sentence Transformers)
-
-First install the Sentence Transformers library:
-
-```bash
-pip install -U sentence-transformers
-```
-
-Then you can load this model and run inference.
-```python
-from sentence_transformers import SentenceTransformer
-
-# Download from the 🤗 Hub
-model = SentenceTransformer("sentence_transformers_model_id")
-# Run inference
-sentences = [
- 'Saif al-Arab GADAFI',
- 'Максім Недасекаў',
- 'Mervyn Allister King',
-]
-embeddings = model.encode(sentences)
-print(embeddings.shape)
-# [3, 384]
-
-# Get the similarity scores for the embeddings
-similarities = model.similarity(embeddings, embeddings)
-print(similarities.shape)
-# [3, 3]
-```
-
-
-
-
-
-
-
-## Evaluation
-
-### Metrics
-
-#### Binary Classification
-
-* Dataset: `sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2`
-* Evaluated with [BinaryClassificationEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator)
-
-| Metric | Value |
-|:--------------------------|:-----------|
-| cosine_accuracy | 0.9905 |
-| cosine_accuracy_threshold | 0.6791 |
-| cosine_f1 | 0.9856 |
-| cosine_f1_threshold | 0.6791 |
-| cosine_precision | 0.9817 |
-| cosine_recall | 0.9896 |
-| **cosine_ap** | **0.9978** |
-| cosine_mcc | 0.9786 |
-
-
-
-
-
-## Training Details
-
-### Training Dataset
-
-#### Unnamed Dataset
-
-* Size: 2,130,620 training samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details |
- min: 3 tokens
- mean: 9.28 tokens
- max: 57 tokens
| - min: 3 tokens
- mean: 9.11 tokens
- max: 65 tokens
| - min: 0.0
- mean: 0.34
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:----------------------------|:-------------------------------|:-----------------|
- | ג'ק וייט
| Jack White
| 1.0
|
- | Абдуллоҳ Гул
| Савицкая Светлана
| 0.0
|
- | ショーン・ジャスティン・ペン
| شان پن
| 1.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Evaluation Dataset
-
-#### Unnamed Dataset
-
-* Size: 266,328 evaluation samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.27 tokens
- max: 79 tokens
| - min: 3 tokens
- mean: 8.99 tokens
- max: 61 tokens
| - min: 0.0
- mean: 0.32
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:---------------------------------------------|:-----------------------------------------------|:-----------------|
- | Анатолий Николаевич Герасимов
| Anatoli Nikolajewitsch Gerassimow
| 1.0
|
- | Igor Stanislavovitsj Prokopenko
| Angelo Lauricella
| 0.0
|
- | Кофе, Линда
| Святлана Яўгенаўна Савіцкая
| 0.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Training Hyperparameters
-#### Non-Default Hyperparameters
-
-- `eval_strategy`: steps
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `gradient_accumulation_steps`: 4
-- `weight_decay`: 0.02
-- `num_train_epochs`: 10
-- `warmup_ratio`: 0.1
-- `fp16`: True
-- `load_best_model_at_end`: True
-- `optim`: adafactor
-- `gradient_checkpointing`: True
-
-#### All Hyperparameters
-Click to expand
-
-- `overwrite_output_dir`: False
-- `do_predict`: False
-- `eval_strategy`: steps
-- `prediction_loss_only`: True
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `per_gpu_train_batch_size`: None
-- `per_gpu_eval_batch_size`: None
-- `gradient_accumulation_steps`: 4
-- `eval_accumulation_steps`: None
-- `torch_empty_cache_steps`: None
-- `learning_rate`: 5e-05
-- `weight_decay`: 0.02
-- `adam_beta1`: 0.9
-- `adam_beta2`: 0.999
-- `adam_epsilon`: 1e-08
-- `max_grad_norm`: 1.0
-- `num_train_epochs`: 10
-- `max_steps`: -1
-- `lr_scheduler_type`: linear
-- `lr_scheduler_kwargs`: {}
-- `warmup_ratio`: 0.1
-- `warmup_steps`: 0
-- `log_level`: passive
-- `log_level_replica`: warning
-- `log_on_each_node`: True
-- `logging_nan_inf_filter`: True
-- `save_safetensors`: True
-- `save_on_each_node`: False
-- `save_only_model`: False
-- `restore_callback_states_from_checkpoint`: False
-- `no_cuda`: False
-- `use_cpu`: False
-- `use_mps_device`: False
-- `seed`: 42
-- `data_seed`: None
-- `jit_mode_eval`: False
-- `use_ipex`: False
-- `bf16`: False
-- `fp16`: True
-- `fp16_opt_level`: O1
-- `half_precision_backend`: auto
-- `bf16_full_eval`: False
-- `fp16_full_eval`: False
-- `tf32`: None
-- `local_rank`: 0
-- `ddp_backend`: None
-- `tpu_num_cores`: None
-- `tpu_metrics_debug`: False
-- `debug`: []
-- `dataloader_drop_last`: False
-- `dataloader_num_workers`: 0
-- `dataloader_prefetch_factor`: None
-- `past_index`: -1
-- `disable_tqdm`: False
-- `remove_unused_columns`: True
-- `label_names`: None
-- `load_best_model_at_end`: True
-- `ignore_data_skip`: False
-- `fsdp`: []
-- `fsdp_min_num_params`: 0
-- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
-- `tp_size`: 0
-- `fsdp_transformer_layer_cls_to_wrap`: None
-- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
-- `deepspeed`: None
-- `label_smoothing_factor`: 0.0
-- `optim`: adafactor
-- `optim_args`: None
-- `adafactor`: False
-- `group_by_length`: False
-- `length_column_name`: length
-- `ddp_find_unused_parameters`: None
-- `ddp_bucket_cap_mb`: None
-- `ddp_broadcast_buffers`: False
-- `dataloader_pin_memory`: True
-- `dataloader_persistent_workers`: False
-- `skip_memory_metrics`: True
-- `use_legacy_prediction_loop`: False
-- `push_to_hub`: False
-- `resume_from_checkpoint`: None
-- `hub_model_id`: None
-- `hub_strategy`: every_save
-- `hub_private_repo`: None
-- `hub_always_push`: False
-- `gradient_checkpointing`: True
-- `gradient_checkpointing_kwargs`: None
-- `include_inputs_for_metrics`: False
-- `include_for_metrics`: []
-- `eval_do_concat_batches`: True
-- `fp16_backend`: auto
-- `push_to_hub_model_id`: None
-- `push_to_hub_organization`: None
-- `mp_parameters`:
-- `auto_find_batch_size`: False
-- `full_determinism`: False
-- `torchdynamo`: None
-- `ray_scope`: last
-- `ddp_timeout`: 1800
-- `torch_compile`: False
-- `torch_compile_backend`: None
-- `torch_compile_mode`: None
-- `include_tokens_per_second`: False
-- `include_num_input_tokens_seen`: False
-- `neftune_noise_alpha`: None
-- `optim_target_modules`: None
-- `batch_eval_metrics`: False
-- `eval_on_start`: False
-- `use_liger_kernel`: False
-- `eval_use_gather_object`: False
-- `average_tokens_across_devices`: False
-- `prompts`: None
-- `batch_sampler`: batch_sampler
-- `multi_dataset_batch_sampler`: proportional
-
-
-
-### Training Logs
-| Epoch | Step | Training Loss | Validation Loss | sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap |
-|:------:|:----:|:-------------:|:---------------:|:---------------------------------------------------------------------:|
-| -1 | -1 | - | - | 0.7195 |
-| 0.9368 | 100 | - | 0.0083 | 0.9597 |
-| 1.8712 | 200 | - | 0.0043 | 0.9877 |
-| 2.8056 | 300 | - | 0.0028 | 0.9936 |
-| 3.7400 | 400 | - | 0.0021 | 0.9954 |
-| 4.6745 | 500 | 0.0224 | 0.0016 | 0.9964 |
-| 5.6089 | 600 | - | 0.0015 | 0.9970 |
-| 6.5433 | 700 | - | 0.0014 | 0.9974 |
-| 7.4778 | 800 | - | 0.0013 | 0.9975 |
-| 8.4122 | 900 | - | 0.0013 | 0.9977 |
-| 9.3466 | 1000 | 0.0052 | 0.0012 | 0.9978 |
-
-
-### Framework Versions
-- Python: 3.12.9
-- Sentence Transformers: 3.4.1
-- Transformers: 4.51.3
-- PyTorch: 2.7.0+cu126
-- Accelerate: 1.6.0
-- Datasets: 3.6.0
-- Tokenizers: 0.21.1
-
-## Citation
-
-### BibTeX
-
-#### Sentence Transformers
-```bibtex
-@inproceedings{reimers-2019-sentence-bert,
- title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
- author = "Reimers, Nils and Gurevych, Iryna",
- booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
- month = "11",
- year = "2019",
- publisher = "Association for Computational Linguistics",
- url = "https://arxiv.org/abs/1908.10084",
-}
-```
-
-#### ContrastiveLoss
-```bibtex
-@inproceedings{hadsell2006dimensionality,
- author={Hadsell, R. and Chopra, S. and LeCun, Y.},
- booktitle={2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)},
- title={Dimensionality Reduction by Learning an Invariant Mapping},
- year={2006},
- volume={2},
- number={},
- pages={1735-1742},
- doi={10.1109/CVPR.2006.100}
-}
-```
-
-
-
-
-
-
\ No newline at end of file
diff --git a/checkpoint-1000/config.json b/checkpoint-1000/config.json
deleted file mode 100644
index 26e48501fdf44110239e00ad4d438aee8679504a..0000000000000000000000000000000000000000
--- a/checkpoint-1000/config.json
+++ /dev/null
@@ -1,25 +0,0 @@
-{
- "architectures": [
- "BertModel"
- ],
- "attention_probs_dropout_prob": 0.1,
- "classifier_dropout": null,
- "gradient_checkpointing": false,
- "hidden_act": "gelu",
- "hidden_dropout_prob": 0.1,
- "hidden_size": 384,
- "initializer_range": 0.02,
- "intermediate_size": 1536,
- "layer_norm_eps": 1e-12,
- "max_position_embeddings": 512,
- "model_type": "bert",
- "num_attention_heads": 12,
- "num_hidden_layers": 12,
- "pad_token_id": 0,
- "position_embedding_type": "absolute",
- "torch_dtype": "float32",
- "transformers_version": "4.51.3",
- "type_vocab_size": 2,
- "use_cache": true,
- "vocab_size": 250037
-}
diff --git a/checkpoint-1000/config_sentence_transformers.json b/checkpoint-1000/config_sentence_transformers.json
deleted file mode 100644
index dcf436801f55bd22a257de2aad7eef5cfd06efaa..0000000000000000000000000000000000000000
--- a/checkpoint-1000/config_sentence_transformers.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "__version__": {
- "sentence_transformers": "3.4.1",
- "transformers": "4.51.3",
- "pytorch": "2.7.0+cu126"
- },
- "prompts": {},
- "default_prompt_name": null,
- "similarity_fn_name": "cosine"
-}
\ No newline at end of file
diff --git a/checkpoint-1000/model.safetensors b/checkpoint-1000/model.safetensors
deleted file mode 100644
index 89f9b51779c36bf737f8aa0a3f8c108b9baf5fb7..0000000000000000000000000000000000000000
--- a/checkpoint-1000/model.safetensors
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:4f4d32a2eafc839cb2ab10b136bf98c4d30bdad7f85e5f55ceafdf3a54a9e859
-size 470637416
diff --git a/checkpoint-1000/modules.json b/checkpoint-1000/modules.json
deleted file mode 100644
index f7640f94e81bb7f4f04daf1668850b38763a13d9..0000000000000000000000000000000000000000
--- a/checkpoint-1000/modules.json
+++ /dev/null
@@ -1,14 +0,0 @@
-[
- {
- "idx": 0,
- "name": "0",
- "path": "",
- "type": "sentence_transformers.models.Transformer"
- },
- {
- "idx": 1,
- "name": "1",
- "path": "1_Pooling",
- "type": "sentence_transformers.models.Pooling"
- }
-]
\ No newline at end of file
diff --git a/checkpoint-1000/optimizer.pt b/checkpoint-1000/optimizer.pt
deleted file mode 100644
index 8c25d5d6c24c5862bf6a8eae4479a6740f67861b..0000000000000000000000000000000000000000
--- a/checkpoint-1000/optimizer.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:c78ab330430a2994c0f6654ccbded4a6558ca0f6cfe08b4fa75960dd3563c6fa
-size 1715019
diff --git a/checkpoint-1000/rng_state.pth b/checkpoint-1000/rng_state.pth
deleted file mode 100644
index 757611e57b74afb17615e80ee85e9d2bc7187ffa..0000000000000000000000000000000000000000
--- a/checkpoint-1000/rng_state.pth
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:f3ba9bdd5b024d60bbe4f0e967c35c5c47d5c8b8d992558db327a2aae780abce
-size 14645
diff --git a/checkpoint-1000/scaler.pt b/checkpoint-1000/scaler.pt
deleted file mode 100644
index 0405b11307032f43fcf2212da19c52823cc01ac4..0000000000000000000000000000000000000000
--- a/checkpoint-1000/scaler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:14ae2a2128444abab378aa06c09a61a84665f758fcc19fc46f5789b0bc1b5665
-size 1383
diff --git a/checkpoint-1000/scheduler.pt b/checkpoint-1000/scheduler.pt
deleted file mode 100644
index e82ede86519166544ed65f7b97a42963278283f0..0000000000000000000000000000000000000000
--- a/checkpoint-1000/scheduler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:c6cb8ef30177a351efb4472a87f05bae30f484d65c37957d4e17a50e58c3b3e9
-size 1465
diff --git a/checkpoint-1000/sentence_bert_config.json b/checkpoint-1000/sentence_bert_config.json
deleted file mode 100644
index 5fd10429389515d3e5cccdeda08cae5fea1ae82e..0000000000000000000000000000000000000000
--- a/checkpoint-1000/sentence_bert_config.json
+++ /dev/null
@@ -1,4 +0,0 @@
-{
- "max_seq_length": 128,
- "do_lower_case": false
-}
\ No newline at end of file
diff --git a/checkpoint-1000/special_tokens_map.json b/checkpoint-1000/special_tokens_map.json
deleted file mode 100644
index b1879d702821e753ffe4245048eee415d54a9385..0000000000000000000000000000000000000000
--- a/checkpoint-1000/special_tokens_map.json
+++ /dev/null
@@ -1,51 +0,0 @@
-{
- "bos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "cls_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "eos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "mask_token": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "pad_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "sep_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "unk_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- }
-}
diff --git a/checkpoint-1000/tokenizer.json b/checkpoint-1000/tokenizer.json
deleted file mode 100644
index e3420945e193cc0791136cdc6e5cd69801c838af..0000000000000000000000000000000000000000
--- a/checkpoint-1000/tokenizer.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:cad551d5600a84242d0973327029452a1e3672ba6313c2a3c3d69c4310e12719
-size 17082987
diff --git a/checkpoint-1000/tokenizer_config.json b/checkpoint-1000/tokenizer_config.json
deleted file mode 100644
index facf4436a8f11c26085c16a14f4e576853927a9e..0000000000000000000000000000000000000000
--- a/checkpoint-1000/tokenizer_config.json
+++ /dev/null
@@ -1,65 +0,0 @@
-{
- "added_tokens_decoder": {
- "0": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "1": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "2": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "3": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "250001": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- }
- },
- "bos_token": "",
- "clean_up_tokenization_spaces": false,
- "cls_token": "",
- "do_lower_case": true,
- "eos_token": "",
- "extra_special_tokens": {},
- "mask_token": "",
- "max_length": 128,
- "model_max_length": 128,
- "pad_to_multiple_of": null,
- "pad_token": "",
- "pad_token_type_id": 0,
- "padding_side": "right",
- "sep_token": "",
- "stride": 0,
- "strip_accents": null,
- "tokenize_chinese_chars": true,
- "tokenizer_class": "BertTokenizer",
- "truncation_side": "right",
- "truncation_strategy": "longest_first",
- "unk_token": ""
-}
diff --git a/checkpoint-1000/trainer_state.json b/checkpoint-1000/trainer_state.json
deleted file mode 100644
index b7a362d66d20c2e80d50fa0d9449e80ba6a50d63..0000000000000000000000000000000000000000
--- a/checkpoint-1000/trainer_state.json
+++ /dev/null
@@ -1,217 +0,0 @@
-{
- "best_global_step": 1000,
- "best_metric": 0.0012360884575173259,
- "best_model_checkpoint": "data/fine-tuned-sbert-sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2-original-adafactor/checkpoint-1000",
- "epoch": 9.346604215456674,
- "eval_steps": 100,
- "global_step": 1000,
- "is_hyper_param_search": false,
- "is_local_process_zero": true,
- "is_world_process_zero": true,
- "log_history": [
- {
- "epoch": 0.936768149882904,
- "eval_loss": 0.008251233026385307,
- "eval_runtime": 117.4457,
- "eval_samples_per_second": 2267.669,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9330529793864755,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6639679670333862,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9596591982248662,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.8990018609372358,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6536919474601746,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.8488676021429209,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.8846836847946726,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.913791126905881,
- "eval_steps_per_second": 0.46,
- "step": 100
- },
- {
- "epoch": 1.8711943793911008,
- "eval_loss": 0.004326523281633854,
- "eval_runtime": 118.308,
- "eval_samples_per_second": 2251.141,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9683099913640971,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.987669070948898,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9520018198362147,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9284143244509058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9445886468795847,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9595322710076808,
- "eval_steps_per_second": 0.456,
- "step": 200
- },
- {
- "epoch": 2.8056206088992974,
- "eval_loss": 0.002782753435894847,
- "eval_runtime": 117.8399,
- "eval_samples_per_second": 2260.083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9790110013892539,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7040826678276062,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9935758649482886,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9680662667809197,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7029732465744019,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9524469797852624,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9648143930767479,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9713401352745615,
- "eval_steps_per_second": 0.458,
- "step": 300
- },
- {
- "epoch": 3.740046838407494,
- "eval_loss": 0.0020659712608903646,
- "eval_runtime": 116.8077,
- "eval_samples_per_second": 2280.056,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9837419742424811,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7114190459251404,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9954100421733855,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.975348704810703,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6966520547866821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.963270232791414,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9687853426826509,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9820016049524246,
- "eval_steps_per_second": 0.462,
- "step": 400
- },
- {
- "epoch": 4.674473067915691,
- "grad_norm": 0.07067500799894333,
- "learning_rate": 2.9402515723270442e-05,
- "loss": 0.0224,
- "step": 500
- },
- {
- "epoch": 4.674473067915691,
- "eval_loss": 0.0016409169184044003,
- "eval_runtime": 117.7739,
- "eval_samples_per_second": 2261.35,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.986370292494274,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.996439193909599,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9792820044518008,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9691467317957321,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.975107979086156,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9834919179181474,
- "eval_steps_per_second": 0.459,
- "step": 500
- },
- {
- "epoch": 5.608899297423887,
- "eval_loss": 0.0014551315689459443,
- "eval_runtime": 117.5801,
- "eval_samples_per_second": 2265.077,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9884729470957083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7460525035858154,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9969945004512654,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9824360661365067,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7435637712478638,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9738614226726382,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9805847418912745,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9842943941304597,
- "eval_steps_per_second": 0.459,
- "step": 600
- },
- {
- "epoch": 6.543325526932084,
- "eval_loss": 0.0013776659034192562,
- "eval_runtime": 117.6764,
- "eval_samples_per_second": 2263.223,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9893740847820374,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.997357375070481,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9838035826704058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9758996171607873,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9822857142857143,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9853261492605755,
- "eval_steps_per_second": 0.459,
- "step": 700
- },
- {
- "epoch": 7.477751756440281,
- "eval_loss": 0.0013444514479488134,
- "eval_runtime": 117.3408,
- "eval_samples_per_second": 2269.696,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9898246536252018,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7261425852775574,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9975494130839752,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9844654628833477,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7227741479873657,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9769000718683564,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9845218986470993,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9844090335893615,
- "eval_steps_per_second": 0.46,
- "step": 800
- },
- {
- "epoch": 8.412177985948478,
- "eval_loss": 0.0012511691311374307,
- "eval_runtime": 117.668,
- "eval_samples_per_second": 2263.385,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9902752224683663,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.685534656047821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9977460917001926,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9852413242919824,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6582455635070801,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9780277137066985,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9794924087922049,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9910581222056631,
- "eval_steps_per_second": 0.459,
- "step": 900
- },
- {
- "epoch": 9.346604215456674,
- "grad_norm": 0.018028028309345245,
- "learning_rate": 3.1970649895178203e-06,
- "loss": 0.0052,
- "step": 1000
- },
- {
- "epoch": 9.346604215456674,
- "eval_loss": 0.0012360884575173259,
- "eval_runtime": 117.4598,
- "eval_samples_per_second": 2267.396,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9905380542935456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6790644526481628,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9977983578816215,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9856131536880567,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6790644526481628,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9785817179348335,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9816899806664392,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9895678092399404,
- "eval_steps_per_second": 0.46,
- "step": 1000
- }
- ],
- "logging_steps": 500,
- "max_steps": 1060,
- "num_input_tokens_seen": 0,
- "num_train_epochs": 10,
- "save_steps": 100,
- "stateful_callbacks": {
- "EarlyStoppingCallback": {
- "args": {
- "early_stopping_patience": 2,
- "early_stopping_threshold": 0.0
- },
- "attributes": {
- "early_stopping_patience_counter": 0
- }
- },
- "TrainerControl": {
- "args": {
- "should_epoch_stop": false,
- "should_evaluate": false,
- "should_log": false,
- "should_save": true,
- "should_training_stop": false
- },
- "attributes": {}
- }
- },
- "total_flos": 0.0,
- "train_batch_size": 5000,
- "trial_name": null,
- "trial_params": null
-}
diff --git a/checkpoint-1000/training_args.bin b/checkpoint-1000/training_args.bin
deleted file mode 100644
index d9dc73de16e341766a62f00cd26c21c6f69c3391..0000000000000000000000000000000000000000
--- a/checkpoint-1000/training_args.bin
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:319aaa354e613c6db82c6bb78290f3da04198ef2c7a75b61b314fa305ed33c45
-size 6033
diff --git a/checkpoint-1000/unigram.json b/checkpoint-1000/unigram.json
deleted file mode 100644
index 2faa9ec874108d53a017ff2c7ab98d155fb21a82..0000000000000000000000000000000000000000
--- a/checkpoint-1000/unigram.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:da145b5e7700ae40f16691ec32a0b1fdc1ee3298db22a31ea55f57a966c4a65d
-size 14763260
diff --git a/checkpoint-1060/1_Pooling/config.json b/checkpoint-1060/1_Pooling/config.json
deleted file mode 100644
index a97f8d140b6aee43dfac9fc4521b2842657c5608..0000000000000000000000000000000000000000
--- a/checkpoint-1060/1_Pooling/config.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "word_embedding_dimension": 384,
- "pooling_mode_cls_token": false,
- "pooling_mode_mean_tokens": true,
- "pooling_mode_max_tokens": false,
- "pooling_mode_mean_sqrt_len_tokens": false,
- "pooling_mode_weightedmean_tokens": false,
- "pooling_mode_lasttoken": false,
- "include_prompt": true
-}
\ No newline at end of file
diff --git a/checkpoint-1060/README.md b/checkpoint-1060/README.md
deleted file mode 100644
index 0dedbe29812d2a5ff395f8e30488dd64e1152c61..0000000000000000000000000000000000000000
--- a/checkpoint-1060/README.md
+++ /dev/null
@@ -1,466 +0,0 @@
----
-language:
-- en
-license: apache-2.0
-tags:
-- sentence-transformers
-- sentence-similarity
-- feature-extraction
-- generated_from_trainer
-- dataset_size:2130620
-- loss:ContrastiveLoss
-base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
-widget:
-- source_sentence: مانوئلا دی سنتا
- sentences:
- - Renko Kitagawa
- - هانس هيرمان وير
- - Ди Чента, Мануэла
-- source_sentence: يورى جافريلوف
- sentences:
- - Wiktor Pinczuk
- - Natalia Germanovna DIRKS
- - Світлана Євгенівна Савицька
-- source_sentence: Џуди Колинс
- sentences:
- - Collins
- - Aisha Muhammed Abdul Salam
- - Phonic Boy On Dope
-- source_sentence: ויליאם בלייר
- sentences:
- - The Hon. Mr Justice Blair
- - Queen Ingrid of Denmark
- - Herman van Rompuy
-- source_sentence: Saif al-Arab GADAFI
- sentences:
- - Максім Недасекаў
- - Mervyn Allister King
- - Paul d. scully-power
-pipeline_tag: sentence-similarity
-library_name: sentence-transformers
-metrics:
-- cosine_accuracy
-- cosine_accuracy_threshold
-- cosine_f1
-- cosine_f1_threshold
-- cosine_precision
-- cosine_recall
-- cosine_ap
-- cosine_mcc
-model-index:
-- name: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
- results:
- - task:
- type: binary-classification
- name: Binary Classification
- dataset:
- name: sentence transformers paraphrase multilingual MiniLM L12 v2
- type: sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2
- metrics:
- - type: cosine_accuracy
- value: 0.9905380542935456
- name: Cosine Accuracy
- - type: cosine_accuracy_threshold
- value: 0.6790644526481628
- name: Cosine Accuracy Threshold
- - type: cosine_f1
- value: 0.9856131536880567
- name: Cosine F1
- - type: cosine_f1_threshold
- value: 0.6790644526481628
- name: Cosine F1 Threshold
- - type: cosine_precision
- value: 0.9816899806664392
- name: Cosine Precision
- - type: cosine_recall
- value: 0.9895678092399404
- name: Cosine Recall
- - type: cosine_ap
- value: 0.9977983578816215
- name: Cosine Ap
- - type: cosine_mcc
- value: 0.9785817179348335
- name: Cosine Mcc
----
-
-# sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
-
-This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
-
-## Model Details
-
-### Model Description
-- **Model Type:** Sentence Transformer
-- **Base model:** [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2)
-- **Maximum Sequence Length:** 128 tokens
-- **Output Dimensionality:** 384 dimensions
-- **Similarity Function:** Cosine Similarity
-
-- **Language:** en
-- **License:** apache-2.0
-
-### Model Sources
-
-- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
-- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
-- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
-
-### Full Model Architecture
-
-```
-SentenceTransformer(
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
-)
-```
-
-## Usage
-
-### Direct Usage (Sentence Transformers)
-
-First install the Sentence Transformers library:
-
-```bash
-pip install -U sentence-transformers
-```
-
-Then you can load this model and run inference.
-```python
-from sentence_transformers import SentenceTransformer
-
-# Download from the 🤗 Hub
-model = SentenceTransformer("sentence_transformers_model_id")
-# Run inference
-sentences = [
- 'Saif al-Arab GADAFI',
- 'Максім Недасекаў',
- 'Mervyn Allister King',
-]
-embeddings = model.encode(sentences)
-print(embeddings.shape)
-# [3, 384]
-
-# Get the similarity scores for the embeddings
-similarities = model.similarity(embeddings, embeddings)
-print(similarities.shape)
-# [3, 3]
-```
-
-
-
-
-
-
-
-## Evaluation
-
-### Metrics
-
-#### Binary Classification
-
-* Dataset: `sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2`
-* Evaluated with [BinaryClassificationEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator)
-
-| Metric | Value |
-|:--------------------------|:-----------|
-| cosine_accuracy | 0.9905 |
-| cosine_accuracy_threshold | 0.6791 |
-| cosine_f1 | 0.9856 |
-| cosine_f1_threshold | 0.6791 |
-| cosine_precision | 0.9817 |
-| cosine_recall | 0.9896 |
-| **cosine_ap** | **0.9978** |
-| cosine_mcc | 0.9786 |
-
-
-
-
-
-## Training Details
-
-### Training Dataset
-
-#### Unnamed Dataset
-
-* Size: 2,130,620 training samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.28 tokens
- max: 57 tokens
| - min: 3 tokens
- mean: 9.11 tokens
- max: 65 tokens
| - min: 0.0
- mean: 0.34
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:----------------------------|:-------------------------------|:-----------------|
- | ג'ק וייט
| Jack White
| 1.0
|
- | Абдуллоҳ Гул
| Савицкая Светлана
| 0.0
|
- | ショーン・ジャスティン・ペン
| شان پن
| 1.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Evaluation Dataset
-
-#### Unnamed Dataset
-
-* Size: 266,328 evaluation samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.27 tokens
- max: 79 tokens
| - min: 3 tokens
- mean: 8.99 tokens
- max: 61 tokens
| - min: 0.0
- mean: 0.32
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:---------------------------------------------|:-----------------------------------------------|:-----------------|
- | Анатолий Николаевич Герасимов
| Anatoli Nikolajewitsch Gerassimow
| 1.0
|
- | Igor Stanislavovitsj Prokopenko
| Angelo Lauricella
| 0.0
|
- | Кофе, Линда
| Святлана Яўгенаўна Савіцкая
| 0.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Training Hyperparameters
-#### Non-Default Hyperparameters
-
-- `eval_strategy`: steps
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `gradient_accumulation_steps`: 4
-- `weight_decay`: 0.02
-- `num_train_epochs`: 10
-- `warmup_ratio`: 0.1
-- `fp16`: True
-- `load_best_model_at_end`: True
-- `optim`: adafactor
-- `gradient_checkpointing`: True
-
-#### All Hyperparameters
-Click to expand
-
-- `overwrite_output_dir`: False
-- `do_predict`: False
-- `eval_strategy`: steps
-- `prediction_loss_only`: True
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `per_gpu_train_batch_size`: None
-- `per_gpu_eval_batch_size`: None
-- `gradient_accumulation_steps`: 4
-- `eval_accumulation_steps`: None
-- `torch_empty_cache_steps`: None
-- `learning_rate`: 5e-05
-- `weight_decay`: 0.02
-- `adam_beta1`: 0.9
-- `adam_beta2`: 0.999
-- `adam_epsilon`: 1e-08
-- `max_grad_norm`: 1.0
-- `num_train_epochs`: 10
-- `max_steps`: -1
-- `lr_scheduler_type`: linear
-- `lr_scheduler_kwargs`: {}
-- `warmup_ratio`: 0.1
-- `warmup_steps`: 0
-- `log_level`: passive
-- `log_level_replica`: warning
-- `log_on_each_node`: True
-- `logging_nan_inf_filter`: True
-- `save_safetensors`: True
-- `save_on_each_node`: False
-- `save_only_model`: False
-- `restore_callback_states_from_checkpoint`: False
-- `no_cuda`: False
-- `use_cpu`: False
-- `use_mps_device`: False
-- `seed`: 42
-- `data_seed`: None
-- `jit_mode_eval`: False
-- `use_ipex`: False
-- `bf16`: False
-- `fp16`: True
-- `fp16_opt_level`: O1
-- `half_precision_backend`: auto
-- `bf16_full_eval`: False
-- `fp16_full_eval`: False
-- `tf32`: None
-- `local_rank`: 0
-- `ddp_backend`: None
-- `tpu_num_cores`: None
-- `tpu_metrics_debug`: False
-- `debug`: []
-- `dataloader_drop_last`: False
-- `dataloader_num_workers`: 0
-- `dataloader_prefetch_factor`: None
-- `past_index`: -1
-- `disable_tqdm`: False
-- `remove_unused_columns`: True
-- `label_names`: None
-- `load_best_model_at_end`: True
-- `ignore_data_skip`: False
-- `fsdp`: []
-- `fsdp_min_num_params`: 0
-- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
-- `tp_size`: 0
-- `fsdp_transformer_layer_cls_to_wrap`: None
-- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
-- `deepspeed`: None
-- `label_smoothing_factor`: 0.0
-- `optim`: adafactor
-- `optim_args`: None
-- `adafactor`: False
-- `group_by_length`: False
-- `length_column_name`: length
-- `ddp_find_unused_parameters`: None
-- `ddp_bucket_cap_mb`: None
-- `ddp_broadcast_buffers`: False
-- `dataloader_pin_memory`: True
-- `dataloader_persistent_workers`: False
-- `skip_memory_metrics`: True
-- `use_legacy_prediction_loop`: False
-- `push_to_hub`: False
-- `resume_from_checkpoint`: None
-- `hub_model_id`: None
-- `hub_strategy`: every_save
-- `hub_private_repo`: None
-- `hub_always_push`: False
-- `gradient_checkpointing`: True
-- `gradient_checkpointing_kwargs`: None
-- `include_inputs_for_metrics`: False
-- `include_for_metrics`: []
-- `eval_do_concat_batches`: True
-- `fp16_backend`: auto
-- `push_to_hub_model_id`: None
-- `push_to_hub_organization`: None
-- `mp_parameters`:
-- `auto_find_batch_size`: False
-- `full_determinism`: False
-- `torchdynamo`: None
-- `ray_scope`: last
-- `ddp_timeout`: 1800
-- `torch_compile`: False
-- `torch_compile_backend`: None
-- `torch_compile_mode`: None
-- `include_tokens_per_second`: False
-- `include_num_input_tokens_seen`: False
-- `neftune_noise_alpha`: None
-- `optim_target_modules`: None
-- `batch_eval_metrics`: False
-- `eval_on_start`: False
-- `use_liger_kernel`: False
-- `eval_use_gather_object`: False
-- `average_tokens_across_devices`: False
-- `prompts`: None
-- `batch_sampler`: batch_sampler
-- `multi_dataset_batch_sampler`: proportional
-
-
-
-### Training Logs
-| Epoch | Step | Training Loss | Validation Loss | sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap |
-|:------:|:----:|:-------------:|:---------------:|:---------------------------------------------------------------------:|
-| -1 | -1 | - | - | 0.7195 |
-| 0.9368 | 100 | - | 0.0083 | 0.9597 |
-| 1.8712 | 200 | - | 0.0043 | 0.9877 |
-| 2.8056 | 300 | - | 0.0028 | 0.9936 |
-| 3.7400 | 400 | - | 0.0021 | 0.9954 |
-| 4.6745 | 500 | 0.0224 | 0.0016 | 0.9964 |
-| 5.6089 | 600 | - | 0.0015 | 0.9970 |
-| 6.5433 | 700 | - | 0.0014 | 0.9974 |
-| 7.4778 | 800 | - | 0.0013 | 0.9975 |
-| 8.4122 | 900 | - | 0.0013 | 0.9977 |
-| 9.3466 | 1000 | 0.0052 | 0.0012 | 0.9978 |
-
-
-### Framework Versions
-- Python: 3.12.9
-- Sentence Transformers: 3.4.1
-- Transformers: 4.51.3
-- PyTorch: 2.7.0+cu126
-- Accelerate: 1.6.0
-- Datasets: 3.6.0
-- Tokenizers: 0.21.1
-
-## Citation
-
-### BibTeX
-
-#### Sentence Transformers
-```bibtex
-@inproceedings{reimers-2019-sentence-bert,
- title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
- author = "Reimers, Nils and Gurevych, Iryna",
- booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
- month = "11",
- year = "2019",
- publisher = "Association for Computational Linguistics",
- url = "https://arxiv.org/abs/1908.10084",
-}
-```
-
-#### ContrastiveLoss
-```bibtex
-@inproceedings{hadsell2006dimensionality,
- author={Hadsell, R. and Chopra, S. and LeCun, Y.},
- booktitle={2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)},
- title={Dimensionality Reduction by Learning an Invariant Mapping},
- year={2006},
- volume={2},
- number={},
- pages={1735-1742},
- doi={10.1109/CVPR.2006.100}
-}
-```
-
-
-
-
-
-
\ No newline at end of file
diff --git a/checkpoint-1060/config.json b/checkpoint-1060/config.json
deleted file mode 100644
index 26e48501fdf44110239e00ad4d438aee8679504a..0000000000000000000000000000000000000000
--- a/checkpoint-1060/config.json
+++ /dev/null
@@ -1,25 +0,0 @@
-{
- "architectures": [
- "BertModel"
- ],
- "attention_probs_dropout_prob": 0.1,
- "classifier_dropout": null,
- "gradient_checkpointing": false,
- "hidden_act": "gelu",
- "hidden_dropout_prob": 0.1,
- "hidden_size": 384,
- "initializer_range": 0.02,
- "intermediate_size": 1536,
- "layer_norm_eps": 1e-12,
- "max_position_embeddings": 512,
- "model_type": "bert",
- "num_attention_heads": 12,
- "num_hidden_layers": 12,
- "pad_token_id": 0,
- "position_embedding_type": "absolute",
- "torch_dtype": "float32",
- "transformers_version": "4.51.3",
- "type_vocab_size": 2,
- "use_cache": true,
- "vocab_size": 250037
-}
diff --git a/checkpoint-1060/config_sentence_transformers.json b/checkpoint-1060/config_sentence_transformers.json
deleted file mode 100644
index dcf436801f55bd22a257de2aad7eef5cfd06efaa..0000000000000000000000000000000000000000
--- a/checkpoint-1060/config_sentence_transformers.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "__version__": {
- "sentence_transformers": "3.4.1",
- "transformers": "4.51.3",
- "pytorch": "2.7.0+cu126"
- },
- "prompts": {},
- "default_prompt_name": null,
- "similarity_fn_name": "cosine"
-}
\ No newline at end of file
diff --git a/checkpoint-1060/model.safetensors b/checkpoint-1060/model.safetensors
deleted file mode 100644
index d9354d3df85376d0bf7aa2b3caf3192a88836dc8..0000000000000000000000000000000000000000
--- a/checkpoint-1060/model.safetensors
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:6d56240a57c4c07d9788a1faa198fd34bca9ea0a1e5a26691b1d009dcae94358
-size 470637416
diff --git a/checkpoint-1060/modules.json b/checkpoint-1060/modules.json
deleted file mode 100644
index f7640f94e81bb7f4f04daf1668850b38763a13d9..0000000000000000000000000000000000000000
--- a/checkpoint-1060/modules.json
+++ /dev/null
@@ -1,14 +0,0 @@
-[
- {
- "idx": 0,
- "name": "0",
- "path": "",
- "type": "sentence_transformers.models.Transformer"
- },
- {
- "idx": 1,
- "name": "1",
- "path": "1_Pooling",
- "type": "sentence_transformers.models.Pooling"
- }
-]
\ No newline at end of file
diff --git a/checkpoint-1060/optimizer.pt b/checkpoint-1060/optimizer.pt
deleted file mode 100644
index 2cc47d46647c5fbf6f1cd4b19a7285502494edbb..0000000000000000000000000000000000000000
--- a/checkpoint-1060/optimizer.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:9b64d5492cb2e0518735a1a0dc7b3b7826a2b4f5d195b44246fdc70db2a64017
-size 1715019
diff --git a/checkpoint-1060/rng_state.pth b/checkpoint-1060/rng_state.pth
deleted file mode 100644
index 287cf7c99824d1f2394b0448f189905f59f73dcc..0000000000000000000000000000000000000000
--- a/checkpoint-1060/rng_state.pth
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:8897bcd95f20279aabd5aac16966d704f565763d9f133ce3e3009c72d02b6438
-size 14645
diff --git a/checkpoint-1060/scaler.pt b/checkpoint-1060/scaler.pt
deleted file mode 100644
index e35e2ac1826f09ddb8e398d029ce19e7b4e9e866..0000000000000000000000000000000000000000
--- a/checkpoint-1060/scaler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:d3bfa35dd520299a41189dc520e331d371a8b9b17d9abff7077c34c5e038a3b0
-size 1383
diff --git a/checkpoint-1060/scheduler.pt b/checkpoint-1060/scheduler.pt
deleted file mode 100644
index 5c8cdb0070f1420eb4eb24b61e319e4e6f814330..0000000000000000000000000000000000000000
--- a/checkpoint-1060/scheduler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:d60e33f28b0c6c27be860020c700ca71a97176bb114f84fcae7c353e227c8a2e
-size 1465
diff --git a/checkpoint-1060/sentence_bert_config.json b/checkpoint-1060/sentence_bert_config.json
deleted file mode 100644
index 5fd10429389515d3e5cccdeda08cae5fea1ae82e..0000000000000000000000000000000000000000
--- a/checkpoint-1060/sentence_bert_config.json
+++ /dev/null
@@ -1,4 +0,0 @@
-{
- "max_seq_length": 128,
- "do_lower_case": false
-}
\ No newline at end of file
diff --git a/checkpoint-1060/special_tokens_map.json b/checkpoint-1060/special_tokens_map.json
deleted file mode 100644
index b1879d702821e753ffe4245048eee415d54a9385..0000000000000000000000000000000000000000
--- a/checkpoint-1060/special_tokens_map.json
+++ /dev/null
@@ -1,51 +0,0 @@
-{
- "bos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "cls_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "eos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "mask_token": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "pad_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "sep_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "unk_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- }
-}
diff --git a/checkpoint-1060/tokenizer.json b/checkpoint-1060/tokenizer.json
deleted file mode 100644
index e3420945e193cc0791136cdc6e5cd69801c838af..0000000000000000000000000000000000000000
--- a/checkpoint-1060/tokenizer.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:cad551d5600a84242d0973327029452a1e3672ba6313c2a3c3d69c4310e12719
-size 17082987
diff --git a/checkpoint-1060/tokenizer_config.json b/checkpoint-1060/tokenizer_config.json
deleted file mode 100644
index facf4436a8f11c26085c16a14f4e576853927a9e..0000000000000000000000000000000000000000
--- a/checkpoint-1060/tokenizer_config.json
+++ /dev/null
@@ -1,65 +0,0 @@
-{
- "added_tokens_decoder": {
- "0": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "1": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "2": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "3": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "250001": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- }
- },
- "bos_token": "",
- "clean_up_tokenization_spaces": false,
- "cls_token": "",
- "do_lower_case": true,
- "eos_token": "",
- "extra_special_tokens": {},
- "mask_token": "",
- "max_length": 128,
- "model_max_length": 128,
- "pad_to_multiple_of": null,
- "pad_token": "",
- "pad_token_type_id": 0,
- "padding_side": "right",
- "sep_token": "",
- "stride": 0,
- "strip_accents": null,
- "tokenize_chinese_chars": true,
- "tokenizer_class": "BertTokenizer",
- "truncation_side": "right",
- "truncation_strategy": "longest_first",
- "unk_token": ""
-}
diff --git a/checkpoint-1060/trainer_state.json b/checkpoint-1060/trainer_state.json
deleted file mode 100644
index 658a4e7b9b83b9574d5fe00d4a931233242a29e3..0000000000000000000000000000000000000000
--- a/checkpoint-1060/trainer_state.json
+++ /dev/null
@@ -1,217 +0,0 @@
-{
- "best_global_step": 1000,
- "best_metric": 0.0012360884575173259,
- "best_model_checkpoint": "data/fine-tuned-sbert-sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2-original-adafactor/checkpoint-1000",
- "epoch": 9.908665105386417,
- "eval_steps": 100,
- "global_step": 1060,
- "is_hyper_param_search": false,
- "is_local_process_zero": true,
- "is_world_process_zero": true,
- "log_history": [
- {
- "epoch": 0.936768149882904,
- "eval_loss": 0.008251233026385307,
- "eval_runtime": 117.4457,
- "eval_samples_per_second": 2267.669,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9330529793864755,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6639679670333862,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9596591982248662,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.8990018609372358,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6536919474601746,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.8488676021429209,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.8846836847946726,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.913791126905881,
- "eval_steps_per_second": 0.46,
- "step": 100
- },
- {
- "epoch": 1.8711943793911008,
- "eval_loss": 0.004326523281633854,
- "eval_runtime": 118.308,
- "eval_samples_per_second": 2251.141,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9683099913640971,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.987669070948898,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9520018198362147,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9284143244509058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9445886468795847,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9595322710076808,
- "eval_steps_per_second": 0.456,
- "step": 200
- },
- {
- "epoch": 2.8056206088992974,
- "eval_loss": 0.002782753435894847,
- "eval_runtime": 117.8399,
- "eval_samples_per_second": 2260.083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9790110013892539,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7040826678276062,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9935758649482886,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9680662667809197,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7029732465744019,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9524469797852624,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9648143930767479,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9713401352745615,
- "eval_steps_per_second": 0.458,
- "step": 300
- },
- {
- "epoch": 3.740046838407494,
- "eval_loss": 0.0020659712608903646,
- "eval_runtime": 116.8077,
- "eval_samples_per_second": 2280.056,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9837419742424811,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7114190459251404,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9954100421733855,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.975348704810703,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6966520547866821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.963270232791414,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9687853426826509,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9820016049524246,
- "eval_steps_per_second": 0.462,
- "step": 400
- },
- {
- "epoch": 4.674473067915691,
- "grad_norm": 0.07067500799894333,
- "learning_rate": 2.9402515723270442e-05,
- "loss": 0.0224,
- "step": 500
- },
- {
- "epoch": 4.674473067915691,
- "eval_loss": 0.0016409169184044003,
- "eval_runtime": 117.7739,
- "eval_samples_per_second": 2261.35,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.986370292494274,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.996439193909599,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9792820044518008,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9691467317957321,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.975107979086156,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9834919179181474,
- "eval_steps_per_second": 0.459,
- "step": 500
- },
- {
- "epoch": 5.608899297423887,
- "eval_loss": 0.0014551315689459443,
- "eval_runtime": 117.5801,
- "eval_samples_per_second": 2265.077,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9884729470957083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7460525035858154,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9969945004512654,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9824360661365067,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7435637712478638,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9738614226726382,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9805847418912745,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9842943941304597,
- "eval_steps_per_second": 0.459,
- "step": 600
- },
- {
- "epoch": 6.543325526932084,
- "eval_loss": 0.0013776659034192562,
- "eval_runtime": 117.6764,
- "eval_samples_per_second": 2263.223,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9893740847820374,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.997357375070481,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9838035826704058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9758996171607873,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9822857142857143,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9853261492605755,
- "eval_steps_per_second": 0.459,
- "step": 700
- },
- {
- "epoch": 7.477751756440281,
- "eval_loss": 0.0013444514479488134,
- "eval_runtime": 117.3408,
- "eval_samples_per_second": 2269.696,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9898246536252018,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7261425852775574,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9975494130839752,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9844654628833477,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7227741479873657,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9769000718683564,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9845218986470993,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9844090335893615,
- "eval_steps_per_second": 0.46,
- "step": 800
- },
- {
- "epoch": 8.412177985948478,
- "eval_loss": 0.0012511691311374307,
- "eval_runtime": 117.668,
- "eval_samples_per_second": 2263.385,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9902752224683663,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.685534656047821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9977460917001926,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9852413242919824,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6582455635070801,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9780277137066985,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9794924087922049,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9910581222056631,
- "eval_steps_per_second": 0.459,
- "step": 900
- },
- {
- "epoch": 9.346604215456674,
- "grad_norm": 0.018028028309345245,
- "learning_rate": 3.1970649895178203e-06,
- "loss": 0.0052,
- "step": 1000
- },
- {
- "epoch": 9.346604215456674,
- "eval_loss": 0.0012360884575173259,
- "eval_runtime": 117.4598,
- "eval_samples_per_second": 2267.396,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9905380542935456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6790644526481628,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9977983578816215,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9856131536880567,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6790644526481628,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9785817179348335,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9816899806664392,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9895678092399404,
- "eval_steps_per_second": 0.46,
- "step": 1000
- }
- ],
- "logging_steps": 500,
- "max_steps": 1060,
- "num_input_tokens_seen": 0,
- "num_train_epochs": 10,
- "save_steps": 100,
- "stateful_callbacks": {
- "EarlyStoppingCallback": {
- "args": {
- "early_stopping_patience": 2,
- "early_stopping_threshold": 0.0
- },
- "attributes": {
- "early_stopping_patience_counter": 0
- }
- },
- "TrainerControl": {
- "args": {
- "should_epoch_stop": false,
- "should_evaluate": false,
- "should_log": false,
- "should_save": true,
- "should_training_stop": true
- },
- "attributes": {}
- }
- },
- "total_flos": 0.0,
- "train_batch_size": 5000,
- "trial_name": null,
- "trial_params": null
-}
diff --git a/checkpoint-1060/training_args.bin b/checkpoint-1060/training_args.bin
deleted file mode 100644
index d9dc73de16e341766a62f00cd26c21c6f69c3391..0000000000000000000000000000000000000000
--- a/checkpoint-1060/training_args.bin
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:319aaa354e613c6db82c6bb78290f3da04198ef2c7a75b61b314fa305ed33c45
-size 6033
diff --git a/checkpoint-1060/unigram.json b/checkpoint-1060/unigram.json
deleted file mode 100644
index 2faa9ec874108d53a017ff2c7ab98d155fb21a82..0000000000000000000000000000000000000000
--- a/checkpoint-1060/unigram.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:da145b5e7700ae40f16691ec32a0b1fdc1ee3298db22a31ea55f57a966c4a65d
-size 14763260
diff --git a/checkpoint-700/1_Pooling/config.json b/checkpoint-700/1_Pooling/config.json
deleted file mode 100644
index a97f8d140b6aee43dfac9fc4521b2842657c5608..0000000000000000000000000000000000000000
--- a/checkpoint-700/1_Pooling/config.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "word_embedding_dimension": 384,
- "pooling_mode_cls_token": false,
- "pooling_mode_mean_tokens": true,
- "pooling_mode_max_tokens": false,
- "pooling_mode_mean_sqrt_len_tokens": false,
- "pooling_mode_weightedmean_tokens": false,
- "pooling_mode_lasttoken": false,
- "include_prompt": true
-}
\ No newline at end of file
diff --git a/checkpoint-700/README.md b/checkpoint-700/README.md
deleted file mode 100644
index 6d2149402975e689e5fe0d57d6463c3f3ab5e0ed..0000000000000000000000000000000000000000
--- a/checkpoint-700/README.md
+++ /dev/null
@@ -1,463 +0,0 @@
----
-language:
-- en
-license: apache-2.0
-tags:
-- sentence-transformers
-- sentence-similarity
-- feature-extraction
-- generated_from_trainer
-- dataset_size:2130620
-- loss:ContrastiveLoss
-base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
-widget:
-- source_sentence: مانوئلا دی سنتا
- sentences:
- - Renko Kitagawa
- - هانس هيرمان وير
- - Ди Чента, Мануэла
-- source_sentence: يورى جافريلوف
- sentences:
- - Wiktor Pinczuk
- - Natalia Germanovna DIRKS
- - Світлана Євгенівна Савицька
-- source_sentence: Џуди Колинс
- sentences:
- - Collins
- - Aisha Muhammed Abdul Salam
- - Phonic Boy On Dope
-- source_sentence: ויליאם בלייר
- sentences:
- - The Hon. Mr Justice Blair
- - Queen Ingrid of Denmark
- - Herman van Rompuy
-- source_sentence: Saif al-Arab GADAFI
- sentences:
- - Максім Недасекаў
- - Mervyn Allister King
- - Paul d. scully-power
-pipeline_tag: sentence-similarity
-library_name: sentence-transformers
-metrics:
-- cosine_accuracy
-- cosine_accuracy_threshold
-- cosine_f1
-- cosine_f1_threshold
-- cosine_precision
-- cosine_recall
-- cosine_ap
-- cosine_mcc
-model-index:
-- name: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
- results:
- - task:
- type: binary-classification
- name: Binary Classification
- dataset:
- name: sentence transformers paraphrase multilingual MiniLM L12 v2
- type: sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2
- metrics:
- - type: cosine_accuracy
- value: 0.9893740847820374
- name: Cosine Accuracy
- - type: cosine_accuracy_threshold
- value: 0.7209540009498596
- name: Cosine Accuracy Threshold
- - type: cosine_f1
- value: 0.9838035826704058
- name: Cosine F1
- - type: cosine_f1_threshold
- value: 0.7209540009498596
- name: Cosine F1 Threshold
- - type: cosine_precision
- value: 0.9822857142857143
- name: Cosine Precision
- - type: cosine_recall
- value: 0.9853261492605755
- name: Cosine Recall
- - type: cosine_ap
- value: 0.997357375070481
- name: Cosine Ap
- - type: cosine_mcc
- value: 0.9758996171607873
- name: Cosine Mcc
----
-
-# sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
-
-This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
-
-## Model Details
-
-### Model Description
-- **Model Type:** Sentence Transformer
-- **Base model:** [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2)
-- **Maximum Sequence Length:** 128 tokens
-- **Output Dimensionality:** 384 dimensions
-- **Similarity Function:** Cosine Similarity
-
-- **Language:** en
-- **License:** apache-2.0
-
-### Model Sources
-
-- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
-- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
-- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
-
-### Full Model Architecture
-
-```
-SentenceTransformer(
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
-)
-```
-
-## Usage
-
-### Direct Usage (Sentence Transformers)
-
-First install the Sentence Transformers library:
-
-```bash
-pip install -U sentence-transformers
-```
-
-Then you can load this model and run inference.
-```python
-from sentence_transformers import SentenceTransformer
-
-# Download from the 🤗 Hub
-model = SentenceTransformer("sentence_transformers_model_id")
-# Run inference
-sentences = [
- 'Saif al-Arab GADAFI',
- 'Максім Недасекаў',
- 'Mervyn Allister King',
-]
-embeddings = model.encode(sentences)
-print(embeddings.shape)
-# [3, 384]
-
-# Get the similarity scores for the embeddings
-similarities = model.similarity(embeddings, embeddings)
-print(similarities.shape)
-# [3, 3]
-```
-
-
-
-
-
-
-
-## Evaluation
-
-### Metrics
-
-#### Binary Classification
-
-* Dataset: `sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2`
-* Evaluated with [BinaryClassificationEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator)
-
-| Metric | Value |
-|:--------------------------|:-----------|
-| cosine_accuracy | 0.9894 |
-| cosine_accuracy_threshold | 0.721 |
-| cosine_f1 | 0.9838 |
-| cosine_f1_threshold | 0.721 |
-| cosine_precision | 0.9823 |
-| cosine_recall | 0.9853 |
-| **cosine_ap** | **0.9974** |
-| cosine_mcc | 0.9759 |
-
-
-
-
-
-## Training Details
-
-### Training Dataset
-
-#### Unnamed Dataset
-
-* Size: 2,130,620 training samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.28 tokens
- max: 57 tokens
| - min: 3 tokens
- mean: 9.11 tokens
- max: 65 tokens
| - min: 0.0
- mean: 0.34
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:----------------------------|:-------------------------------|:-----------------|
- | ג'ק וייט
| Jack White
| 1.0
|
- | Абдуллоҳ Гул
| Савицкая Светлана
| 0.0
|
- | ショーン・ジャスティン・ペン
| شان پن
| 1.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Evaluation Dataset
-
-#### Unnamed Dataset
-
-* Size: 266,328 evaluation samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.27 tokens
- max: 79 tokens
| - min: 3 tokens
- mean: 8.99 tokens
- max: 61 tokens
| - min: 0.0
- mean: 0.32
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:---------------------------------------------|:-----------------------------------------------|:-----------------|
- | Анатолий Николаевич Герасимов
| Anatoli Nikolajewitsch Gerassimow
| 1.0
|
- | Igor Stanislavovitsj Prokopenko
| Angelo Lauricella
| 0.0
|
- | Кофе, Линда
| Святлана Яўгенаўна Савіцкая
| 0.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Training Hyperparameters
-#### Non-Default Hyperparameters
-
-- `eval_strategy`: steps
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `gradient_accumulation_steps`: 4
-- `weight_decay`: 0.02
-- `num_train_epochs`: 10
-- `warmup_ratio`: 0.1
-- `fp16`: True
-- `load_best_model_at_end`: True
-- `optim`: adafactor
-- `gradient_checkpointing`: True
-
-#### All Hyperparameters
-Click to expand
-
-- `overwrite_output_dir`: False
-- `do_predict`: False
-- `eval_strategy`: steps
-- `prediction_loss_only`: True
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `per_gpu_train_batch_size`: None
-- `per_gpu_eval_batch_size`: None
-- `gradient_accumulation_steps`: 4
-- `eval_accumulation_steps`: None
-- `torch_empty_cache_steps`: None
-- `learning_rate`: 5e-05
-- `weight_decay`: 0.02
-- `adam_beta1`: 0.9
-- `adam_beta2`: 0.999
-- `adam_epsilon`: 1e-08
-- `max_grad_norm`: 1.0
-- `num_train_epochs`: 10
-- `max_steps`: -1
-- `lr_scheduler_type`: linear
-- `lr_scheduler_kwargs`: {}
-- `warmup_ratio`: 0.1
-- `warmup_steps`: 0
-- `log_level`: passive
-- `log_level_replica`: warning
-- `log_on_each_node`: True
-- `logging_nan_inf_filter`: True
-- `save_safetensors`: True
-- `save_on_each_node`: False
-- `save_only_model`: False
-- `restore_callback_states_from_checkpoint`: False
-- `no_cuda`: False
-- `use_cpu`: False
-- `use_mps_device`: False
-- `seed`: 42
-- `data_seed`: None
-- `jit_mode_eval`: False
-- `use_ipex`: False
-- `bf16`: False
-- `fp16`: True
-- `fp16_opt_level`: O1
-- `half_precision_backend`: auto
-- `bf16_full_eval`: False
-- `fp16_full_eval`: False
-- `tf32`: None
-- `local_rank`: 0
-- `ddp_backend`: None
-- `tpu_num_cores`: None
-- `tpu_metrics_debug`: False
-- `debug`: []
-- `dataloader_drop_last`: False
-- `dataloader_num_workers`: 0
-- `dataloader_prefetch_factor`: None
-- `past_index`: -1
-- `disable_tqdm`: False
-- `remove_unused_columns`: True
-- `label_names`: None
-- `load_best_model_at_end`: True
-- `ignore_data_skip`: False
-- `fsdp`: []
-- `fsdp_min_num_params`: 0
-- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
-- `tp_size`: 0
-- `fsdp_transformer_layer_cls_to_wrap`: None
-- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
-- `deepspeed`: None
-- `label_smoothing_factor`: 0.0
-- `optim`: adafactor
-- `optim_args`: None
-- `adafactor`: False
-- `group_by_length`: False
-- `length_column_name`: length
-- `ddp_find_unused_parameters`: None
-- `ddp_bucket_cap_mb`: None
-- `ddp_broadcast_buffers`: False
-- `dataloader_pin_memory`: True
-- `dataloader_persistent_workers`: False
-- `skip_memory_metrics`: True
-- `use_legacy_prediction_loop`: False
-- `push_to_hub`: False
-- `resume_from_checkpoint`: None
-- `hub_model_id`: None
-- `hub_strategy`: every_save
-- `hub_private_repo`: None
-- `hub_always_push`: False
-- `gradient_checkpointing`: True
-- `gradient_checkpointing_kwargs`: None
-- `include_inputs_for_metrics`: False
-- `include_for_metrics`: []
-- `eval_do_concat_batches`: True
-- `fp16_backend`: auto
-- `push_to_hub_model_id`: None
-- `push_to_hub_organization`: None
-- `mp_parameters`:
-- `auto_find_batch_size`: False
-- `full_determinism`: False
-- `torchdynamo`: None
-- `ray_scope`: last
-- `ddp_timeout`: 1800
-- `torch_compile`: False
-- `torch_compile_backend`: None
-- `torch_compile_mode`: None
-- `include_tokens_per_second`: False
-- `include_num_input_tokens_seen`: False
-- `neftune_noise_alpha`: None
-- `optim_target_modules`: None
-- `batch_eval_metrics`: False
-- `eval_on_start`: False
-- `use_liger_kernel`: False
-- `eval_use_gather_object`: False
-- `average_tokens_across_devices`: False
-- `prompts`: None
-- `batch_sampler`: batch_sampler
-- `multi_dataset_batch_sampler`: proportional
-
-
-
-### Training Logs
-| Epoch | Step | Training Loss | Validation Loss | sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap |
-|:------:|:----:|:-------------:|:---------------:|:---------------------------------------------------------------------:|
-| -1 | -1 | - | - | 0.7195 |
-| 0.9368 | 100 | - | 0.0083 | 0.9597 |
-| 1.8712 | 200 | - | 0.0043 | 0.9877 |
-| 2.8056 | 300 | - | 0.0028 | 0.9936 |
-| 3.7400 | 400 | - | 0.0021 | 0.9954 |
-| 4.6745 | 500 | 0.0224 | 0.0016 | 0.9964 |
-| 5.6089 | 600 | - | 0.0015 | 0.9970 |
-| 6.5433 | 700 | - | 0.0014 | 0.9974 |
-
-
-### Framework Versions
-- Python: 3.12.9
-- Sentence Transformers: 3.4.1
-- Transformers: 4.51.3
-- PyTorch: 2.7.0+cu126
-- Accelerate: 1.6.0
-- Datasets: 3.6.0
-- Tokenizers: 0.21.1
-
-## Citation
-
-### BibTeX
-
-#### Sentence Transformers
-```bibtex
-@inproceedings{reimers-2019-sentence-bert,
- title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
- author = "Reimers, Nils and Gurevych, Iryna",
- booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
- month = "11",
- year = "2019",
- publisher = "Association for Computational Linguistics",
- url = "https://arxiv.org/abs/1908.10084",
-}
-```
-
-#### ContrastiveLoss
-```bibtex
-@inproceedings{hadsell2006dimensionality,
- author={Hadsell, R. and Chopra, S. and LeCun, Y.},
- booktitle={2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)},
- title={Dimensionality Reduction by Learning an Invariant Mapping},
- year={2006},
- volume={2},
- number={},
- pages={1735-1742},
- doi={10.1109/CVPR.2006.100}
-}
-```
-
-
-
-
-
-
\ No newline at end of file
diff --git a/checkpoint-700/config.json b/checkpoint-700/config.json
deleted file mode 100644
index 26e48501fdf44110239e00ad4d438aee8679504a..0000000000000000000000000000000000000000
--- a/checkpoint-700/config.json
+++ /dev/null
@@ -1,25 +0,0 @@
-{
- "architectures": [
- "BertModel"
- ],
- "attention_probs_dropout_prob": 0.1,
- "classifier_dropout": null,
- "gradient_checkpointing": false,
- "hidden_act": "gelu",
- "hidden_dropout_prob": 0.1,
- "hidden_size": 384,
- "initializer_range": 0.02,
- "intermediate_size": 1536,
- "layer_norm_eps": 1e-12,
- "max_position_embeddings": 512,
- "model_type": "bert",
- "num_attention_heads": 12,
- "num_hidden_layers": 12,
- "pad_token_id": 0,
- "position_embedding_type": "absolute",
- "torch_dtype": "float32",
- "transformers_version": "4.51.3",
- "type_vocab_size": 2,
- "use_cache": true,
- "vocab_size": 250037
-}
diff --git a/checkpoint-700/config_sentence_transformers.json b/checkpoint-700/config_sentence_transformers.json
deleted file mode 100644
index dcf436801f55bd22a257de2aad7eef5cfd06efaa..0000000000000000000000000000000000000000
--- a/checkpoint-700/config_sentence_transformers.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "__version__": {
- "sentence_transformers": "3.4.1",
- "transformers": "4.51.3",
- "pytorch": "2.7.0+cu126"
- },
- "prompts": {},
- "default_prompt_name": null,
- "similarity_fn_name": "cosine"
-}
\ No newline at end of file
diff --git a/checkpoint-700/model.safetensors b/checkpoint-700/model.safetensors
deleted file mode 100644
index 52693cb740849a2f7a1c733385bbd5cdb0f8dc2a..0000000000000000000000000000000000000000
--- a/checkpoint-700/model.safetensors
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:000c7957828311658198adf8e84fe33c1a660836e1cb7b256504f04b8cc770aa
-size 470637416
diff --git a/checkpoint-700/modules.json b/checkpoint-700/modules.json
deleted file mode 100644
index f7640f94e81bb7f4f04daf1668850b38763a13d9..0000000000000000000000000000000000000000
--- a/checkpoint-700/modules.json
+++ /dev/null
@@ -1,14 +0,0 @@
-[
- {
- "idx": 0,
- "name": "0",
- "path": "",
- "type": "sentence_transformers.models.Transformer"
- },
- {
- "idx": 1,
- "name": "1",
- "path": "1_Pooling",
- "type": "sentence_transformers.models.Pooling"
- }
-]
\ No newline at end of file
diff --git a/checkpoint-700/optimizer.pt b/checkpoint-700/optimizer.pt
deleted file mode 100644
index 91d2b908484584d3301f52c5e35521c8e8f4bc1f..0000000000000000000000000000000000000000
--- a/checkpoint-700/optimizer.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:e4d52b430164992014c645f921eb89cb4f11af746bb4925e980955f54650d62b
-size 1715019
diff --git a/checkpoint-700/rng_state.pth b/checkpoint-700/rng_state.pth
deleted file mode 100644
index 6833f675ca3ef16b7f927cbf94b66a586d75fbde..0000000000000000000000000000000000000000
--- a/checkpoint-700/rng_state.pth
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:2d56fff20bbc2f130ed2e293f289ea71c316a57e902789c67ac719e6a30c1b4e
-size 14645
diff --git a/checkpoint-700/scaler.pt b/checkpoint-700/scaler.pt
deleted file mode 100644
index 08873e1b922bcedaea9e8ed84f83e0fe850ad40e..0000000000000000000000000000000000000000
--- a/checkpoint-700/scaler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:850c3d909f8a0af6f9b431fac5a25833ab1658c39f899825e3b347b6af8a490b
-size 1383
diff --git a/checkpoint-700/scheduler.pt b/checkpoint-700/scheduler.pt
deleted file mode 100644
index 5bf6c616eb1a6c50f7da1eea57727265738d698b..0000000000000000000000000000000000000000
--- a/checkpoint-700/scheduler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:d4c2cfbe9b0a118af0de30855464c364252cb3147a7ab4ad3d16c608263feebb
-size 1465
diff --git a/checkpoint-700/sentence_bert_config.json b/checkpoint-700/sentence_bert_config.json
deleted file mode 100644
index 5fd10429389515d3e5cccdeda08cae5fea1ae82e..0000000000000000000000000000000000000000
--- a/checkpoint-700/sentence_bert_config.json
+++ /dev/null
@@ -1,4 +0,0 @@
-{
- "max_seq_length": 128,
- "do_lower_case": false
-}
\ No newline at end of file
diff --git a/checkpoint-700/special_tokens_map.json b/checkpoint-700/special_tokens_map.json
deleted file mode 100644
index b1879d702821e753ffe4245048eee415d54a9385..0000000000000000000000000000000000000000
--- a/checkpoint-700/special_tokens_map.json
+++ /dev/null
@@ -1,51 +0,0 @@
-{
- "bos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "cls_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "eos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "mask_token": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "pad_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "sep_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "unk_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- }
-}
diff --git a/checkpoint-700/tokenizer.json b/checkpoint-700/tokenizer.json
deleted file mode 100644
index e3420945e193cc0791136cdc6e5cd69801c838af..0000000000000000000000000000000000000000
--- a/checkpoint-700/tokenizer.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:cad551d5600a84242d0973327029452a1e3672ba6313c2a3c3d69c4310e12719
-size 17082987
diff --git a/checkpoint-700/tokenizer_config.json b/checkpoint-700/tokenizer_config.json
deleted file mode 100644
index facf4436a8f11c26085c16a14f4e576853927a9e..0000000000000000000000000000000000000000
--- a/checkpoint-700/tokenizer_config.json
+++ /dev/null
@@ -1,65 +0,0 @@
-{
- "added_tokens_decoder": {
- "0": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "1": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "2": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "3": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "250001": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- }
- },
- "bos_token": "",
- "clean_up_tokenization_spaces": false,
- "cls_token": "",
- "do_lower_case": true,
- "eos_token": "",
- "extra_special_tokens": {},
- "mask_token": "",
- "max_length": 128,
- "model_max_length": 128,
- "pad_to_multiple_of": null,
- "pad_token": "",
- "pad_token_type_id": 0,
- "padding_side": "right",
- "sep_token": "",
- "stride": 0,
- "strip_accents": null,
- "tokenize_chinese_chars": true,
- "tokenizer_class": "BertTokenizer",
- "truncation_side": "right",
- "truncation_strategy": "longest_first",
- "unk_token": ""
-}
diff --git a/checkpoint-700/trainer_state.json b/checkpoint-700/trainer_state.json
deleted file mode 100644
index a9f4a1abde75937a2f8acf4b776d2eb68f899e53..0000000000000000000000000000000000000000
--- a/checkpoint-700/trainer_state.json
+++ /dev/null
@@ -1,162 +0,0 @@
-{
- "best_global_step": 700,
- "best_metric": 0.0013776659034192562,
- "best_model_checkpoint": "data/fine-tuned-sbert-sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2-original-adafactor/checkpoint-700",
- "epoch": 6.543325526932084,
- "eval_steps": 100,
- "global_step": 700,
- "is_hyper_param_search": false,
- "is_local_process_zero": true,
- "is_world_process_zero": true,
- "log_history": [
- {
- "epoch": 0.936768149882904,
- "eval_loss": 0.008251233026385307,
- "eval_runtime": 117.4457,
- "eval_samples_per_second": 2267.669,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9330529793864755,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6639679670333862,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9596591982248662,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.8990018609372358,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6536919474601746,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.8488676021429209,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.8846836847946726,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.913791126905881,
- "eval_steps_per_second": 0.46,
- "step": 100
- },
- {
- "epoch": 1.8711943793911008,
- "eval_loss": 0.004326523281633854,
- "eval_runtime": 118.308,
- "eval_samples_per_second": 2251.141,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9683099913640971,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.987669070948898,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9520018198362147,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9284143244509058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9445886468795847,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9595322710076808,
- "eval_steps_per_second": 0.456,
- "step": 200
- },
- {
- "epoch": 2.8056206088992974,
- "eval_loss": 0.002782753435894847,
- "eval_runtime": 117.8399,
- "eval_samples_per_second": 2260.083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9790110013892539,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7040826678276062,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9935758649482886,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9680662667809197,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7029732465744019,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9524469797852624,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9648143930767479,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9713401352745615,
- "eval_steps_per_second": 0.458,
- "step": 300
- },
- {
- "epoch": 3.740046838407494,
- "eval_loss": 0.0020659712608903646,
- "eval_runtime": 116.8077,
- "eval_samples_per_second": 2280.056,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9837419742424811,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7114190459251404,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9954100421733855,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.975348704810703,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6966520547866821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.963270232791414,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9687853426826509,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9820016049524246,
- "eval_steps_per_second": 0.462,
- "step": 400
- },
- {
- "epoch": 4.674473067915691,
- "grad_norm": 0.07067500799894333,
- "learning_rate": 2.9402515723270442e-05,
- "loss": 0.0224,
- "step": 500
- },
- {
- "epoch": 4.674473067915691,
- "eval_loss": 0.0016409169184044003,
- "eval_runtime": 117.7739,
- "eval_samples_per_second": 2261.35,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.986370292494274,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.996439193909599,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9792820044518008,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9691467317957321,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.975107979086156,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9834919179181474,
- "eval_steps_per_second": 0.459,
- "step": 500
- },
- {
- "epoch": 5.608899297423887,
- "eval_loss": 0.0014551315689459443,
- "eval_runtime": 117.5801,
- "eval_samples_per_second": 2265.077,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9884729470957083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7460525035858154,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9969945004512654,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9824360661365067,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7435637712478638,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9738614226726382,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9805847418912745,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9842943941304597,
- "eval_steps_per_second": 0.459,
- "step": 600
- },
- {
- "epoch": 6.543325526932084,
- "eval_loss": 0.0013776659034192562,
- "eval_runtime": 117.6764,
- "eval_samples_per_second": 2263.223,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9893740847820374,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.997357375070481,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9838035826704058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9758996171607873,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9822857142857143,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9853261492605755,
- "eval_steps_per_second": 0.459,
- "step": 700
- }
- ],
- "logging_steps": 500,
- "max_steps": 1060,
- "num_input_tokens_seen": 0,
- "num_train_epochs": 10,
- "save_steps": 100,
- "stateful_callbacks": {
- "EarlyStoppingCallback": {
- "args": {
- "early_stopping_patience": 2,
- "early_stopping_threshold": 0.0
- },
- "attributes": {
- "early_stopping_patience_counter": 0
- }
- },
- "TrainerControl": {
- "args": {
- "should_epoch_stop": false,
- "should_evaluate": false,
- "should_log": false,
- "should_save": true,
- "should_training_stop": false
- },
- "attributes": {}
- }
- },
- "total_flos": 0.0,
- "train_batch_size": 5000,
- "trial_name": null,
- "trial_params": null
-}
diff --git a/checkpoint-700/training_args.bin b/checkpoint-700/training_args.bin
deleted file mode 100644
index d9dc73de16e341766a62f00cd26c21c6f69c3391..0000000000000000000000000000000000000000
--- a/checkpoint-700/training_args.bin
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:319aaa354e613c6db82c6bb78290f3da04198ef2c7a75b61b314fa305ed33c45
-size 6033
diff --git a/checkpoint-700/unigram.json b/checkpoint-700/unigram.json
deleted file mode 100644
index 2faa9ec874108d53a017ff2c7ab98d155fb21a82..0000000000000000000000000000000000000000
--- a/checkpoint-700/unigram.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:da145b5e7700ae40f16691ec32a0b1fdc1ee3298db22a31ea55f57a966c4a65d
-size 14763260
diff --git a/checkpoint-800/1_Pooling/config.json b/checkpoint-800/1_Pooling/config.json
deleted file mode 100644
index a97f8d140b6aee43dfac9fc4521b2842657c5608..0000000000000000000000000000000000000000
--- a/checkpoint-800/1_Pooling/config.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "word_embedding_dimension": 384,
- "pooling_mode_cls_token": false,
- "pooling_mode_mean_tokens": true,
- "pooling_mode_max_tokens": false,
- "pooling_mode_mean_sqrt_len_tokens": false,
- "pooling_mode_weightedmean_tokens": false,
- "pooling_mode_lasttoken": false,
- "include_prompt": true
-}
\ No newline at end of file
diff --git a/checkpoint-800/README.md b/checkpoint-800/README.md
deleted file mode 100644
index cd313b3261e9ba67277caf33566b39b227c4c9e0..0000000000000000000000000000000000000000
--- a/checkpoint-800/README.md
+++ /dev/null
@@ -1,464 +0,0 @@
----
-language:
-- en
-license: apache-2.0
-tags:
-- sentence-transformers
-- sentence-similarity
-- feature-extraction
-- generated_from_trainer
-- dataset_size:2130620
-- loss:ContrastiveLoss
-base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
-widget:
-- source_sentence: مانوئلا دی سنتا
- sentences:
- - Renko Kitagawa
- - هانس هيرمان وير
- - Ди Чента, Мануэла
-- source_sentence: يورى جافريلوف
- sentences:
- - Wiktor Pinczuk
- - Natalia Germanovna DIRKS
- - Світлана Євгенівна Савицька
-- source_sentence: Џуди Колинс
- sentences:
- - Collins
- - Aisha Muhammed Abdul Salam
- - Phonic Boy On Dope
-- source_sentence: ויליאם בלייר
- sentences:
- - The Hon. Mr Justice Blair
- - Queen Ingrid of Denmark
- - Herman van Rompuy
-- source_sentence: Saif al-Arab GADAFI
- sentences:
- - Максім Недасекаў
- - Mervyn Allister King
- - Paul d. scully-power
-pipeline_tag: sentence-similarity
-library_name: sentence-transformers
-metrics:
-- cosine_accuracy
-- cosine_accuracy_threshold
-- cosine_f1
-- cosine_f1_threshold
-- cosine_precision
-- cosine_recall
-- cosine_ap
-- cosine_mcc
-model-index:
-- name: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
- results:
- - task:
- type: binary-classification
- name: Binary Classification
- dataset:
- name: sentence transformers paraphrase multilingual MiniLM L12 v2
- type: sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2
- metrics:
- - type: cosine_accuracy
- value: 0.9898246536252018
- name: Cosine Accuracy
- - type: cosine_accuracy_threshold
- value: 0.7261425852775574
- name: Cosine Accuracy Threshold
- - type: cosine_f1
- value: 0.9844654628833477
- name: Cosine F1
- - type: cosine_f1_threshold
- value: 0.7227741479873657
- name: Cosine F1 Threshold
- - type: cosine_precision
- value: 0.9845218986470993
- name: Cosine Precision
- - type: cosine_recall
- value: 0.9844090335893615
- name: Cosine Recall
- - type: cosine_ap
- value: 0.9975494130839752
- name: Cosine Ap
- - type: cosine_mcc
- value: 0.9769000718683564
- name: Cosine Mcc
----
-
-# sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
-
-This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
-
-## Model Details
-
-### Model Description
-- **Model Type:** Sentence Transformer
-- **Base model:** [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2)
-- **Maximum Sequence Length:** 128 tokens
-- **Output Dimensionality:** 384 dimensions
-- **Similarity Function:** Cosine Similarity
-
-- **Language:** en
-- **License:** apache-2.0
-
-### Model Sources
-
-- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
-- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
-- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
-
-### Full Model Architecture
-
-```
-SentenceTransformer(
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
-)
-```
-
-## Usage
-
-### Direct Usage (Sentence Transformers)
-
-First install the Sentence Transformers library:
-
-```bash
-pip install -U sentence-transformers
-```
-
-Then you can load this model and run inference.
-```python
-from sentence_transformers import SentenceTransformer
-
-# Download from the 🤗 Hub
-model = SentenceTransformer("sentence_transformers_model_id")
-# Run inference
-sentences = [
- 'Saif al-Arab GADAFI',
- 'Максім Недасекаў',
- 'Mervyn Allister King',
-]
-embeddings = model.encode(sentences)
-print(embeddings.shape)
-# [3, 384]
-
-# Get the similarity scores for the embeddings
-similarities = model.similarity(embeddings, embeddings)
-print(similarities.shape)
-# [3, 3]
-```
-
-
-
-
-
-
-
-## Evaluation
-
-### Metrics
-
-#### Binary Classification
-
-* Dataset: `sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2`
-* Evaluated with [BinaryClassificationEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator)
-
-| Metric | Value |
-|:--------------------------|:-----------|
-| cosine_accuracy | 0.9898 |
-| cosine_accuracy_threshold | 0.7261 |
-| cosine_f1 | 0.9845 |
-| cosine_f1_threshold | 0.7228 |
-| cosine_precision | 0.9845 |
-| cosine_recall | 0.9844 |
-| **cosine_ap** | **0.9975** |
-| cosine_mcc | 0.9769 |
-
-
-
-
-
-## Training Details
-
-### Training Dataset
-
-#### Unnamed Dataset
-
-* Size: 2,130,620 training samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.28 tokens
- max: 57 tokens
| - min: 3 tokens
- mean: 9.11 tokens
- max: 65 tokens
| - min: 0.0
- mean: 0.34
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:----------------------------|:-------------------------------|:-----------------|
- | ג'ק וייט
| Jack White
| 1.0
|
- | Абдуллоҳ Гул
| Савицкая Светлана
| 0.0
|
- | ショーン・ジャスティン・ペン
| شان پن
| 1.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Evaluation Dataset
-
-#### Unnamed Dataset
-
-* Size: 266,328 evaluation samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.27 tokens
- max: 79 tokens
| - min: 3 tokens
- mean: 8.99 tokens
- max: 61 tokens
| - min: 0.0
- mean: 0.32
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:---------------------------------------------|:-----------------------------------------------|:-----------------|
- | Анатолий Николаевич Герасимов
| Anatoli Nikolajewitsch Gerassimow
| 1.0
|
- | Igor Stanislavovitsj Prokopenko
| Angelo Lauricella
| 0.0
|
- | Кофе, Линда
| Святлана Яўгенаўна Савіцкая
| 0.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Training Hyperparameters
-#### Non-Default Hyperparameters
-
-- `eval_strategy`: steps
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `gradient_accumulation_steps`: 4
-- `weight_decay`: 0.02
-- `num_train_epochs`: 10
-- `warmup_ratio`: 0.1
-- `fp16`: True
-- `load_best_model_at_end`: True
-- `optim`: adafactor
-- `gradient_checkpointing`: True
-
-#### All Hyperparameters
-Click to expand
-
-- `overwrite_output_dir`: False
-- `do_predict`: False
-- `eval_strategy`: steps
-- `prediction_loss_only`: True
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `per_gpu_train_batch_size`: None
-- `per_gpu_eval_batch_size`: None
-- `gradient_accumulation_steps`: 4
-- `eval_accumulation_steps`: None
-- `torch_empty_cache_steps`: None
-- `learning_rate`: 5e-05
-- `weight_decay`: 0.02
-- `adam_beta1`: 0.9
-- `adam_beta2`: 0.999
-- `adam_epsilon`: 1e-08
-- `max_grad_norm`: 1.0
-- `num_train_epochs`: 10
-- `max_steps`: -1
-- `lr_scheduler_type`: linear
-- `lr_scheduler_kwargs`: {}
-- `warmup_ratio`: 0.1
-- `warmup_steps`: 0
-- `log_level`: passive
-- `log_level_replica`: warning
-- `log_on_each_node`: True
-- `logging_nan_inf_filter`: True
-- `save_safetensors`: True
-- `save_on_each_node`: False
-- `save_only_model`: False
-- `restore_callback_states_from_checkpoint`: False
-- `no_cuda`: False
-- `use_cpu`: False
-- `use_mps_device`: False
-- `seed`: 42
-- `data_seed`: None
-- `jit_mode_eval`: False
-- `use_ipex`: False
-- `bf16`: False
-- `fp16`: True
-- `fp16_opt_level`: O1
-- `half_precision_backend`: auto
-- `bf16_full_eval`: False
-- `fp16_full_eval`: False
-- `tf32`: None
-- `local_rank`: 0
-- `ddp_backend`: None
-- `tpu_num_cores`: None
-- `tpu_metrics_debug`: False
-- `debug`: []
-- `dataloader_drop_last`: False
-- `dataloader_num_workers`: 0
-- `dataloader_prefetch_factor`: None
-- `past_index`: -1
-- `disable_tqdm`: False
-- `remove_unused_columns`: True
-- `label_names`: None
-- `load_best_model_at_end`: True
-- `ignore_data_skip`: False
-- `fsdp`: []
-- `fsdp_min_num_params`: 0
-- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
-- `tp_size`: 0
-- `fsdp_transformer_layer_cls_to_wrap`: None
-- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
-- `deepspeed`: None
-- `label_smoothing_factor`: 0.0
-- `optim`: adafactor
-- `optim_args`: None
-- `adafactor`: False
-- `group_by_length`: False
-- `length_column_name`: length
-- `ddp_find_unused_parameters`: None
-- `ddp_bucket_cap_mb`: None
-- `ddp_broadcast_buffers`: False
-- `dataloader_pin_memory`: True
-- `dataloader_persistent_workers`: False
-- `skip_memory_metrics`: True
-- `use_legacy_prediction_loop`: False
-- `push_to_hub`: False
-- `resume_from_checkpoint`: None
-- `hub_model_id`: None
-- `hub_strategy`: every_save
-- `hub_private_repo`: None
-- `hub_always_push`: False
-- `gradient_checkpointing`: True
-- `gradient_checkpointing_kwargs`: None
-- `include_inputs_for_metrics`: False
-- `include_for_metrics`: []
-- `eval_do_concat_batches`: True
-- `fp16_backend`: auto
-- `push_to_hub_model_id`: None
-- `push_to_hub_organization`: None
-- `mp_parameters`:
-- `auto_find_batch_size`: False
-- `full_determinism`: False
-- `torchdynamo`: None
-- `ray_scope`: last
-- `ddp_timeout`: 1800
-- `torch_compile`: False
-- `torch_compile_backend`: None
-- `torch_compile_mode`: None
-- `include_tokens_per_second`: False
-- `include_num_input_tokens_seen`: False
-- `neftune_noise_alpha`: None
-- `optim_target_modules`: None
-- `batch_eval_metrics`: False
-- `eval_on_start`: False
-- `use_liger_kernel`: False
-- `eval_use_gather_object`: False
-- `average_tokens_across_devices`: False
-- `prompts`: None
-- `batch_sampler`: batch_sampler
-- `multi_dataset_batch_sampler`: proportional
-
-
-
-### Training Logs
-| Epoch | Step | Training Loss | Validation Loss | sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap |
-|:------:|:----:|:-------------:|:---------------:|:---------------------------------------------------------------------:|
-| -1 | -1 | - | - | 0.7195 |
-| 0.9368 | 100 | - | 0.0083 | 0.9597 |
-| 1.8712 | 200 | - | 0.0043 | 0.9877 |
-| 2.8056 | 300 | - | 0.0028 | 0.9936 |
-| 3.7400 | 400 | - | 0.0021 | 0.9954 |
-| 4.6745 | 500 | 0.0224 | 0.0016 | 0.9964 |
-| 5.6089 | 600 | - | 0.0015 | 0.9970 |
-| 6.5433 | 700 | - | 0.0014 | 0.9974 |
-| 7.4778 | 800 | - | 0.0013 | 0.9975 |
-
-
-### Framework Versions
-- Python: 3.12.9
-- Sentence Transformers: 3.4.1
-- Transformers: 4.51.3
-- PyTorch: 2.7.0+cu126
-- Accelerate: 1.6.0
-- Datasets: 3.6.0
-- Tokenizers: 0.21.1
-
-## Citation
-
-### BibTeX
-
-#### Sentence Transformers
-```bibtex
-@inproceedings{reimers-2019-sentence-bert,
- title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
- author = "Reimers, Nils and Gurevych, Iryna",
- booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
- month = "11",
- year = "2019",
- publisher = "Association for Computational Linguistics",
- url = "https://arxiv.org/abs/1908.10084",
-}
-```
-
-#### ContrastiveLoss
-```bibtex
-@inproceedings{hadsell2006dimensionality,
- author={Hadsell, R. and Chopra, S. and LeCun, Y.},
- booktitle={2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)},
- title={Dimensionality Reduction by Learning an Invariant Mapping},
- year={2006},
- volume={2},
- number={},
- pages={1735-1742},
- doi={10.1109/CVPR.2006.100}
-}
-```
-
-
-
-
-
-
\ No newline at end of file
diff --git a/checkpoint-800/config.json b/checkpoint-800/config.json
deleted file mode 100644
index 26e48501fdf44110239e00ad4d438aee8679504a..0000000000000000000000000000000000000000
--- a/checkpoint-800/config.json
+++ /dev/null
@@ -1,25 +0,0 @@
-{
- "architectures": [
- "BertModel"
- ],
- "attention_probs_dropout_prob": 0.1,
- "classifier_dropout": null,
- "gradient_checkpointing": false,
- "hidden_act": "gelu",
- "hidden_dropout_prob": 0.1,
- "hidden_size": 384,
- "initializer_range": 0.02,
- "intermediate_size": 1536,
- "layer_norm_eps": 1e-12,
- "max_position_embeddings": 512,
- "model_type": "bert",
- "num_attention_heads": 12,
- "num_hidden_layers": 12,
- "pad_token_id": 0,
- "position_embedding_type": "absolute",
- "torch_dtype": "float32",
- "transformers_version": "4.51.3",
- "type_vocab_size": 2,
- "use_cache": true,
- "vocab_size": 250037
-}
diff --git a/checkpoint-800/config_sentence_transformers.json b/checkpoint-800/config_sentence_transformers.json
deleted file mode 100644
index dcf436801f55bd22a257de2aad7eef5cfd06efaa..0000000000000000000000000000000000000000
--- a/checkpoint-800/config_sentence_transformers.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "__version__": {
- "sentence_transformers": "3.4.1",
- "transformers": "4.51.3",
- "pytorch": "2.7.0+cu126"
- },
- "prompts": {},
- "default_prompt_name": null,
- "similarity_fn_name": "cosine"
-}
\ No newline at end of file
diff --git a/checkpoint-800/model.safetensors b/checkpoint-800/model.safetensors
deleted file mode 100644
index 4cf13bf32dd147b016d9eb9cf982715bd7bb55d7..0000000000000000000000000000000000000000
--- a/checkpoint-800/model.safetensors
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:b66d1d585dfebc08863614ccdebb38299de51ab7d25f409fb291039d5ea02eb9
-size 470637416
diff --git a/checkpoint-800/modules.json b/checkpoint-800/modules.json
deleted file mode 100644
index f7640f94e81bb7f4f04daf1668850b38763a13d9..0000000000000000000000000000000000000000
--- a/checkpoint-800/modules.json
+++ /dev/null
@@ -1,14 +0,0 @@
-[
- {
- "idx": 0,
- "name": "0",
- "path": "",
- "type": "sentence_transformers.models.Transformer"
- },
- {
- "idx": 1,
- "name": "1",
- "path": "1_Pooling",
- "type": "sentence_transformers.models.Pooling"
- }
-]
\ No newline at end of file
diff --git a/checkpoint-800/optimizer.pt b/checkpoint-800/optimizer.pt
deleted file mode 100644
index 2dd984789da5208968a802d468614f3ea293008c..0000000000000000000000000000000000000000
--- a/checkpoint-800/optimizer.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:d869574956c72feded8b2ea03c2cf068d8a9bbecaa9d6860da120a9c05d7a0d5
-size 1715019
diff --git a/checkpoint-800/rng_state.pth b/checkpoint-800/rng_state.pth
deleted file mode 100644
index 7d9c83944c7a6b2ab4dd70d0d5a86ebc32ad15c5..0000000000000000000000000000000000000000
--- a/checkpoint-800/rng_state.pth
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:a99e3b70fe30f03da06ddbe6481de15749c3ce69220b4bcbaee9bb85c0a740e2
-size 14645
diff --git a/checkpoint-800/scaler.pt b/checkpoint-800/scaler.pt
deleted file mode 100644
index d00b69a7cf9d97a806097c96fd3a5101140e2b7d..0000000000000000000000000000000000000000
--- a/checkpoint-800/scaler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:6fef122931c86c2d2736773be787da21ac6460d41580735381e953556fb410be
-size 1383
diff --git a/checkpoint-800/scheduler.pt b/checkpoint-800/scheduler.pt
deleted file mode 100644
index 6e7cc18271006a75a6cb5c925082f593573110da..0000000000000000000000000000000000000000
--- a/checkpoint-800/scheduler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:43f7b03f3c993bc2d17ba83f67a64574681014d2507fe5ebfcbb06ee79166e03
-size 1465
diff --git a/checkpoint-800/sentence_bert_config.json b/checkpoint-800/sentence_bert_config.json
deleted file mode 100644
index 5fd10429389515d3e5cccdeda08cae5fea1ae82e..0000000000000000000000000000000000000000
--- a/checkpoint-800/sentence_bert_config.json
+++ /dev/null
@@ -1,4 +0,0 @@
-{
- "max_seq_length": 128,
- "do_lower_case": false
-}
\ No newline at end of file
diff --git a/checkpoint-800/special_tokens_map.json b/checkpoint-800/special_tokens_map.json
deleted file mode 100644
index b1879d702821e753ffe4245048eee415d54a9385..0000000000000000000000000000000000000000
--- a/checkpoint-800/special_tokens_map.json
+++ /dev/null
@@ -1,51 +0,0 @@
-{
- "bos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "cls_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "eos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "mask_token": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "pad_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "sep_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "unk_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- }
-}
diff --git a/checkpoint-800/tokenizer.json b/checkpoint-800/tokenizer.json
deleted file mode 100644
index e3420945e193cc0791136cdc6e5cd69801c838af..0000000000000000000000000000000000000000
--- a/checkpoint-800/tokenizer.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:cad551d5600a84242d0973327029452a1e3672ba6313c2a3c3d69c4310e12719
-size 17082987
diff --git a/checkpoint-800/tokenizer_config.json b/checkpoint-800/tokenizer_config.json
deleted file mode 100644
index facf4436a8f11c26085c16a14f4e576853927a9e..0000000000000000000000000000000000000000
--- a/checkpoint-800/tokenizer_config.json
+++ /dev/null
@@ -1,65 +0,0 @@
-{
- "added_tokens_decoder": {
- "0": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "1": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "2": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "3": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "250001": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- }
- },
- "bos_token": "",
- "clean_up_tokenization_spaces": false,
- "cls_token": "",
- "do_lower_case": true,
- "eos_token": "",
- "extra_special_tokens": {},
- "mask_token": "",
- "max_length": 128,
- "model_max_length": 128,
- "pad_to_multiple_of": null,
- "pad_token": "",
- "pad_token_type_id": 0,
- "padding_side": "right",
- "sep_token": "",
- "stride": 0,
- "strip_accents": null,
- "tokenize_chinese_chars": true,
- "tokenizer_class": "BertTokenizer",
- "truncation_side": "right",
- "truncation_strategy": "longest_first",
- "unk_token": ""
-}
diff --git a/checkpoint-800/trainer_state.json b/checkpoint-800/trainer_state.json
deleted file mode 100644
index 9390fb849c156a3cfff25c1ca1f807d01b21b49a..0000000000000000000000000000000000000000
--- a/checkpoint-800/trainer_state.json
+++ /dev/null
@@ -1,178 +0,0 @@
-{
- "best_global_step": 800,
- "best_metric": 0.0013444514479488134,
- "best_model_checkpoint": "data/fine-tuned-sbert-sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2-original-adafactor/checkpoint-800",
- "epoch": 7.477751756440281,
- "eval_steps": 100,
- "global_step": 800,
- "is_hyper_param_search": false,
- "is_local_process_zero": true,
- "is_world_process_zero": true,
- "log_history": [
- {
- "epoch": 0.936768149882904,
- "eval_loss": 0.008251233026385307,
- "eval_runtime": 117.4457,
- "eval_samples_per_second": 2267.669,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9330529793864755,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6639679670333862,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9596591982248662,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.8990018609372358,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6536919474601746,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.8488676021429209,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.8846836847946726,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.913791126905881,
- "eval_steps_per_second": 0.46,
- "step": 100
- },
- {
- "epoch": 1.8711943793911008,
- "eval_loss": 0.004326523281633854,
- "eval_runtime": 118.308,
- "eval_samples_per_second": 2251.141,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9683099913640971,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.987669070948898,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9520018198362147,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9284143244509058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9445886468795847,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9595322710076808,
- "eval_steps_per_second": 0.456,
- "step": 200
- },
- {
- "epoch": 2.8056206088992974,
- "eval_loss": 0.002782753435894847,
- "eval_runtime": 117.8399,
- "eval_samples_per_second": 2260.083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9790110013892539,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7040826678276062,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9935758649482886,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9680662667809197,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7029732465744019,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9524469797852624,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9648143930767479,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9713401352745615,
- "eval_steps_per_second": 0.458,
- "step": 300
- },
- {
- "epoch": 3.740046838407494,
- "eval_loss": 0.0020659712608903646,
- "eval_runtime": 116.8077,
- "eval_samples_per_second": 2280.056,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9837419742424811,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7114190459251404,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9954100421733855,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.975348704810703,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6966520547866821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.963270232791414,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9687853426826509,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9820016049524246,
- "eval_steps_per_second": 0.462,
- "step": 400
- },
- {
- "epoch": 4.674473067915691,
- "grad_norm": 0.07067500799894333,
- "learning_rate": 2.9402515723270442e-05,
- "loss": 0.0224,
- "step": 500
- },
- {
- "epoch": 4.674473067915691,
- "eval_loss": 0.0016409169184044003,
- "eval_runtime": 117.7739,
- "eval_samples_per_second": 2261.35,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.986370292494274,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.996439193909599,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9792820044518008,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9691467317957321,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.975107979086156,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9834919179181474,
- "eval_steps_per_second": 0.459,
- "step": 500
- },
- {
- "epoch": 5.608899297423887,
- "eval_loss": 0.0014551315689459443,
- "eval_runtime": 117.5801,
- "eval_samples_per_second": 2265.077,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9884729470957083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7460525035858154,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9969945004512654,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9824360661365067,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7435637712478638,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9738614226726382,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9805847418912745,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9842943941304597,
- "eval_steps_per_second": 0.459,
- "step": 600
- },
- {
- "epoch": 6.543325526932084,
- "eval_loss": 0.0013776659034192562,
- "eval_runtime": 117.6764,
- "eval_samples_per_second": 2263.223,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9893740847820374,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.997357375070481,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9838035826704058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9758996171607873,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9822857142857143,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9853261492605755,
- "eval_steps_per_second": 0.459,
- "step": 700
- },
- {
- "epoch": 7.477751756440281,
- "eval_loss": 0.0013444514479488134,
- "eval_runtime": 117.3408,
- "eval_samples_per_second": 2269.696,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9898246536252018,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7261425852775574,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9975494130839752,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9844654628833477,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7227741479873657,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9769000718683564,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9845218986470993,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9844090335893615,
- "eval_steps_per_second": 0.46,
- "step": 800
- }
- ],
- "logging_steps": 500,
- "max_steps": 1060,
- "num_input_tokens_seen": 0,
- "num_train_epochs": 10,
- "save_steps": 100,
- "stateful_callbacks": {
- "EarlyStoppingCallback": {
- "args": {
- "early_stopping_patience": 2,
- "early_stopping_threshold": 0.0
- },
- "attributes": {
- "early_stopping_patience_counter": 0
- }
- },
- "TrainerControl": {
- "args": {
- "should_epoch_stop": false,
- "should_evaluate": false,
- "should_log": false,
- "should_save": true,
- "should_training_stop": false
- },
- "attributes": {}
- }
- },
- "total_flos": 0.0,
- "train_batch_size": 5000,
- "trial_name": null,
- "trial_params": null
-}
diff --git a/checkpoint-800/training_args.bin b/checkpoint-800/training_args.bin
deleted file mode 100644
index d9dc73de16e341766a62f00cd26c21c6f69c3391..0000000000000000000000000000000000000000
--- a/checkpoint-800/training_args.bin
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:319aaa354e613c6db82c6bb78290f3da04198ef2c7a75b61b314fa305ed33c45
-size 6033
diff --git a/checkpoint-800/unigram.json b/checkpoint-800/unigram.json
deleted file mode 100644
index 2faa9ec874108d53a017ff2c7ab98d155fb21a82..0000000000000000000000000000000000000000
--- a/checkpoint-800/unigram.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:da145b5e7700ae40f16691ec32a0b1fdc1ee3298db22a31ea55f57a966c4a65d
-size 14763260
diff --git a/checkpoint-900/1_Pooling/config.json b/checkpoint-900/1_Pooling/config.json
deleted file mode 100644
index a97f8d140b6aee43dfac9fc4521b2842657c5608..0000000000000000000000000000000000000000
--- a/checkpoint-900/1_Pooling/config.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "word_embedding_dimension": 384,
- "pooling_mode_cls_token": false,
- "pooling_mode_mean_tokens": true,
- "pooling_mode_max_tokens": false,
- "pooling_mode_mean_sqrt_len_tokens": false,
- "pooling_mode_weightedmean_tokens": false,
- "pooling_mode_lasttoken": false,
- "include_prompt": true
-}
\ No newline at end of file
diff --git a/checkpoint-900/README.md b/checkpoint-900/README.md
deleted file mode 100644
index edcbb47da51ba9dc956dcfb12775f1b93e52b793..0000000000000000000000000000000000000000
--- a/checkpoint-900/README.md
+++ /dev/null
@@ -1,465 +0,0 @@
----
-language:
-- en
-license: apache-2.0
-tags:
-- sentence-transformers
-- sentence-similarity
-- feature-extraction
-- generated_from_trainer
-- dataset_size:2130620
-- loss:ContrastiveLoss
-base_model: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
-widget:
-- source_sentence: مانوئلا دی سنتا
- sentences:
- - Renko Kitagawa
- - هانس هيرمان وير
- - Ди Чента, Мануэла
-- source_sentence: يورى جافريلوف
- sentences:
- - Wiktor Pinczuk
- - Natalia Germanovna DIRKS
- - Світлана Євгенівна Савицька
-- source_sentence: Џуди Колинс
- sentences:
- - Collins
- - Aisha Muhammed Abdul Salam
- - Phonic Boy On Dope
-- source_sentence: ויליאם בלייר
- sentences:
- - The Hon. Mr Justice Blair
- - Queen Ingrid of Denmark
- - Herman van Rompuy
-- source_sentence: Saif al-Arab GADAFI
- sentences:
- - Максім Недасекаў
- - Mervyn Allister King
- - Paul d. scully-power
-pipeline_tag: sentence-similarity
-library_name: sentence-transformers
-metrics:
-- cosine_accuracy
-- cosine_accuracy_threshold
-- cosine_f1
-- cosine_f1_threshold
-- cosine_precision
-- cosine_recall
-- cosine_ap
-- cosine_mcc
-model-index:
-- name: sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
- results:
- - task:
- type: binary-classification
- name: Binary Classification
- dataset:
- name: sentence transformers paraphrase multilingual MiniLM L12 v2
- type: sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2
- metrics:
- - type: cosine_accuracy
- value: 0.9902752224683663
- name: Cosine Accuracy
- - type: cosine_accuracy_threshold
- value: 0.685534656047821
- name: Cosine Accuracy Threshold
- - type: cosine_f1
- value: 0.9852413242919824
- name: Cosine F1
- - type: cosine_f1_threshold
- value: 0.6582455635070801
- name: Cosine F1 Threshold
- - type: cosine_precision
- value: 0.9794924087922049
- name: Cosine Precision
- - type: cosine_recall
- value: 0.9910581222056631
- name: Cosine Recall
- - type: cosine_ap
- value: 0.9977460917001926
- name: Cosine Ap
- - type: cosine_mcc
- value: 0.9780277137066985
- name: Cosine Mcc
----
-
-# sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2-address-matcher-original
-
-This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
-
-## Model Details
-
-### Model Description
-- **Model Type:** Sentence Transformer
-- **Base model:** [sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2)
-- **Maximum Sequence Length:** 128 tokens
-- **Output Dimensionality:** 384 dimensions
-- **Similarity Function:** Cosine Similarity
-
-- **Language:** en
-- **License:** apache-2.0
-
-### Model Sources
-
-- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
-- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
-- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
-
-### Full Model Architecture
-
-```
-SentenceTransformer(
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False}) with Transformer model: BertModel
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
-)
-```
-
-## Usage
-
-### Direct Usage (Sentence Transformers)
-
-First install the Sentence Transformers library:
-
-```bash
-pip install -U sentence-transformers
-```
-
-Then you can load this model and run inference.
-```python
-from sentence_transformers import SentenceTransformer
-
-# Download from the 🤗 Hub
-model = SentenceTransformer("sentence_transformers_model_id")
-# Run inference
-sentences = [
- 'Saif al-Arab GADAFI',
- 'Максім Недасекаў',
- 'Mervyn Allister King',
-]
-embeddings = model.encode(sentences)
-print(embeddings.shape)
-# [3, 384]
-
-# Get the similarity scores for the embeddings
-similarities = model.similarity(embeddings, embeddings)
-print(similarities.shape)
-# [3, 3]
-```
-
-
-
-
-
-
-
-## Evaluation
-
-### Metrics
-
-#### Binary Classification
-
-* Dataset: `sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2`
-* Evaluated with [BinaryClassificationEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator)
-
-| Metric | Value |
-|:--------------------------|:-----------|
-| cosine_accuracy | 0.9903 |
-| cosine_accuracy_threshold | 0.6855 |
-| cosine_f1 | 0.9852 |
-| cosine_f1_threshold | 0.6582 |
-| cosine_precision | 0.9795 |
-| cosine_recall | 0.9911 |
-| **cosine_ap** | **0.9977** |
-| cosine_mcc | 0.978 |
-
-
-
-
-
-## Training Details
-
-### Training Dataset
-
-#### Unnamed Dataset
-
-* Size: 2,130,620 training samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.28 tokens
- max: 57 tokens
| - min: 3 tokens
- mean: 9.11 tokens
- max: 65 tokens
| - min: 0.0
- mean: 0.34
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:----------------------------|:-------------------------------|:-----------------|
- | ג'ק וייט
| Jack White
| 1.0
|
- | Абдуллоҳ Гул
| Савицкая Светлана
| 0.0
|
- | ショーン・ジャスティン・ペン
| شان پن
| 1.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Evaluation Dataset
-
-#### Unnamed Dataset
-
-* Size: 266,328 evaluation samples
-* Columns: sentence1
, sentence2
, and label
-* Approximate statistics based on the first 1000 samples:
- | | sentence1 | sentence2 | label |
- |:--------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:---------------------------------------------------------------|
- | type | string | string | float |
- | details | - min: 3 tokens
- mean: 9.27 tokens
- max: 79 tokens
| - min: 3 tokens
- mean: 8.99 tokens
- max: 61 tokens
| - min: 0.0
- mean: 0.32
- max: 1.0
|
-* Samples:
- | sentence1 | sentence2 | label |
- |:---------------------------------------------|:-----------------------------------------------|:-----------------|
- | Анатолий Николаевич Герасимов
| Anatoli Nikolajewitsch Gerassimow
| 1.0
|
- | Igor Stanislavovitsj Prokopenko
| Angelo Lauricella
| 0.0
|
- | Кофе, Линда
| Святлана Яўгенаўна Савіцкая
| 0.0
|
-* Loss: [ContrastiveLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters:
- ```json
- {
- "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE",
- "margin": 0.5,
- "size_average": true
- }
- ```
-
-### Training Hyperparameters
-#### Non-Default Hyperparameters
-
-- `eval_strategy`: steps
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `gradient_accumulation_steps`: 4
-- `weight_decay`: 0.02
-- `num_train_epochs`: 10
-- `warmup_ratio`: 0.1
-- `fp16`: True
-- `load_best_model_at_end`: True
-- `optim`: adafactor
-- `gradient_checkpointing`: True
-
-#### All Hyperparameters
-Click to expand
-
-- `overwrite_output_dir`: False
-- `do_predict`: False
-- `eval_strategy`: steps
-- `prediction_loss_only`: True
-- `per_device_train_batch_size`: 5000
-- `per_device_eval_batch_size`: 5000
-- `per_gpu_train_batch_size`: None
-- `per_gpu_eval_batch_size`: None
-- `gradient_accumulation_steps`: 4
-- `eval_accumulation_steps`: None
-- `torch_empty_cache_steps`: None
-- `learning_rate`: 5e-05
-- `weight_decay`: 0.02
-- `adam_beta1`: 0.9
-- `adam_beta2`: 0.999
-- `adam_epsilon`: 1e-08
-- `max_grad_norm`: 1.0
-- `num_train_epochs`: 10
-- `max_steps`: -1
-- `lr_scheduler_type`: linear
-- `lr_scheduler_kwargs`: {}
-- `warmup_ratio`: 0.1
-- `warmup_steps`: 0
-- `log_level`: passive
-- `log_level_replica`: warning
-- `log_on_each_node`: True
-- `logging_nan_inf_filter`: True
-- `save_safetensors`: True
-- `save_on_each_node`: False
-- `save_only_model`: False
-- `restore_callback_states_from_checkpoint`: False
-- `no_cuda`: False
-- `use_cpu`: False
-- `use_mps_device`: False
-- `seed`: 42
-- `data_seed`: None
-- `jit_mode_eval`: False
-- `use_ipex`: False
-- `bf16`: False
-- `fp16`: True
-- `fp16_opt_level`: O1
-- `half_precision_backend`: auto
-- `bf16_full_eval`: False
-- `fp16_full_eval`: False
-- `tf32`: None
-- `local_rank`: 0
-- `ddp_backend`: None
-- `tpu_num_cores`: None
-- `tpu_metrics_debug`: False
-- `debug`: []
-- `dataloader_drop_last`: False
-- `dataloader_num_workers`: 0
-- `dataloader_prefetch_factor`: None
-- `past_index`: -1
-- `disable_tqdm`: False
-- `remove_unused_columns`: True
-- `label_names`: None
-- `load_best_model_at_end`: True
-- `ignore_data_skip`: False
-- `fsdp`: []
-- `fsdp_min_num_params`: 0
-- `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
-- `tp_size`: 0
-- `fsdp_transformer_layer_cls_to_wrap`: None
-- `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
-- `deepspeed`: None
-- `label_smoothing_factor`: 0.0
-- `optim`: adafactor
-- `optim_args`: None
-- `adafactor`: False
-- `group_by_length`: False
-- `length_column_name`: length
-- `ddp_find_unused_parameters`: None
-- `ddp_bucket_cap_mb`: None
-- `ddp_broadcast_buffers`: False
-- `dataloader_pin_memory`: True
-- `dataloader_persistent_workers`: False
-- `skip_memory_metrics`: True
-- `use_legacy_prediction_loop`: False
-- `push_to_hub`: False
-- `resume_from_checkpoint`: None
-- `hub_model_id`: None
-- `hub_strategy`: every_save
-- `hub_private_repo`: None
-- `hub_always_push`: False
-- `gradient_checkpointing`: True
-- `gradient_checkpointing_kwargs`: None
-- `include_inputs_for_metrics`: False
-- `include_for_metrics`: []
-- `eval_do_concat_batches`: True
-- `fp16_backend`: auto
-- `push_to_hub_model_id`: None
-- `push_to_hub_organization`: None
-- `mp_parameters`:
-- `auto_find_batch_size`: False
-- `full_determinism`: False
-- `torchdynamo`: None
-- `ray_scope`: last
-- `ddp_timeout`: 1800
-- `torch_compile`: False
-- `torch_compile_backend`: None
-- `torch_compile_mode`: None
-- `include_tokens_per_second`: False
-- `include_num_input_tokens_seen`: False
-- `neftune_noise_alpha`: None
-- `optim_target_modules`: None
-- `batch_eval_metrics`: False
-- `eval_on_start`: False
-- `use_liger_kernel`: False
-- `eval_use_gather_object`: False
-- `average_tokens_across_devices`: False
-- `prompts`: None
-- `batch_sampler`: batch_sampler
-- `multi_dataset_batch_sampler`: proportional
-
-
-
-### Training Logs
-| Epoch | Step | Training Loss | Validation Loss | sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap |
-|:------:|:----:|:-------------:|:---------------:|:---------------------------------------------------------------------:|
-| -1 | -1 | - | - | 0.7195 |
-| 0.9368 | 100 | - | 0.0083 | 0.9597 |
-| 1.8712 | 200 | - | 0.0043 | 0.9877 |
-| 2.8056 | 300 | - | 0.0028 | 0.9936 |
-| 3.7400 | 400 | - | 0.0021 | 0.9954 |
-| 4.6745 | 500 | 0.0224 | 0.0016 | 0.9964 |
-| 5.6089 | 600 | - | 0.0015 | 0.9970 |
-| 6.5433 | 700 | - | 0.0014 | 0.9974 |
-| 7.4778 | 800 | - | 0.0013 | 0.9975 |
-| 8.4122 | 900 | - | 0.0013 | 0.9977 |
-
-
-### Framework Versions
-- Python: 3.12.9
-- Sentence Transformers: 3.4.1
-- Transformers: 4.51.3
-- PyTorch: 2.7.0+cu126
-- Accelerate: 1.6.0
-- Datasets: 3.6.0
-- Tokenizers: 0.21.1
-
-## Citation
-
-### BibTeX
-
-#### Sentence Transformers
-```bibtex
-@inproceedings{reimers-2019-sentence-bert,
- title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
- author = "Reimers, Nils and Gurevych, Iryna",
- booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
- month = "11",
- year = "2019",
- publisher = "Association for Computational Linguistics",
- url = "https://arxiv.org/abs/1908.10084",
-}
-```
-
-#### ContrastiveLoss
-```bibtex
-@inproceedings{hadsell2006dimensionality,
- author={Hadsell, R. and Chopra, S. and LeCun, Y.},
- booktitle={2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)},
- title={Dimensionality Reduction by Learning an Invariant Mapping},
- year={2006},
- volume={2},
- number={},
- pages={1735-1742},
- doi={10.1109/CVPR.2006.100}
-}
-```
-
-
-
-
-
-
\ No newline at end of file
diff --git a/checkpoint-900/config.json b/checkpoint-900/config.json
deleted file mode 100644
index 26e48501fdf44110239e00ad4d438aee8679504a..0000000000000000000000000000000000000000
--- a/checkpoint-900/config.json
+++ /dev/null
@@ -1,25 +0,0 @@
-{
- "architectures": [
- "BertModel"
- ],
- "attention_probs_dropout_prob": 0.1,
- "classifier_dropout": null,
- "gradient_checkpointing": false,
- "hidden_act": "gelu",
- "hidden_dropout_prob": 0.1,
- "hidden_size": 384,
- "initializer_range": 0.02,
- "intermediate_size": 1536,
- "layer_norm_eps": 1e-12,
- "max_position_embeddings": 512,
- "model_type": "bert",
- "num_attention_heads": 12,
- "num_hidden_layers": 12,
- "pad_token_id": 0,
- "position_embedding_type": "absolute",
- "torch_dtype": "float32",
- "transformers_version": "4.51.3",
- "type_vocab_size": 2,
- "use_cache": true,
- "vocab_size": 250037
-}
diff --git a/checkpoint-900/config_sentence_transformers.json b/checkpoint-900/config_sentence_transformers.json
deleted file mode 100644
index dcf436801f55bd22a257de2aad7eef5cfd06efaa..0000000000000000000000000000000000000000
--- a/checkpoint-900/config_sentence_transformers.json
+++ /dev/null
@@ -1,10 +0,0 @@
-{
- "__version__": {
- "sentence_transformers": "3.4.1",
- "transformers": "4.51.3",
- "pytorch": "2.7.0+cu126"
- },
- "prompts": {},
- "default_prompt_name": null,
- "similarity_fn_name": "cosine"
-}
\ No newline at end of file
diff --git a/checkpoint-900/model.safetensors b/checkpoint-900/model.safetensors
deleted file mode 100644
index 5927fa0a868d92d34994480521a0b84c9b1253da..0000000000000000000000000000000000000000
--- a/checkpoint-900/model.safetensors
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:e5f8133057b269c25406bb9c6bc18cb3251d44b999f2caa81b590b81adc1534a
-size 470637416
diff --git a/checkpoint-900/modules.json b/checkpoint-900/modules.json
deleted file mode 100644
index f7640f94e81bb7f4f04daf1668850b38763a13d9..0000000000000000000000000000000000000000
--- a/checkpoint-900/modules.json
+++ /dev/null
@@ -1,14 +0,0 @@
-[
- {
- "idx": 0,
- "name": "0",
- "path": "",
- "type": "sentence_transformers.models.Transformer"
- },
- {
- "idx": 1,
- "name": "1",
- "path": "1_Pooling",
- "type": "sentence_transformers.models.Pooling"
- }
-]
\ No newline at end of file
diff --git a/checkpoint-900/optimizer.pt b/checkpoint-900/optimizer.pt
deleted file mode 100644
index 44c865af0a73ca355984e6f52bd0274f91ee6e6b..0000000000000000000000000000000000000000
--- a/checkpoint-900/optimizer.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:54f996c5a0e8fbb19d9bf29ee41dbdf5f346a266fbaae119071d5b149b2c87a4
-size 1715019
diff --git a/checkpoint-900/rng_state.pth b/checkpoint-900/rng_state.pth
deleted file mode 100644
index 061a035e9e0d61883526c4bc25d153a28da2e12b..0000000000000000000000000000000000000000
--- a/checkpoint-900/rng_state.pth
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:714b560c1f244925481f6c9bd4d3fecfa16c5979f2f92a244f976f05529bdfd0
-size 14645
diff --git a/checkpoint-900/scaler.pt b/checkpoint-900/scaler.pt
deleted file mode 100644
index 1e4437a955590eb751b51104943fa84acda739f5..0000000000000000000000000000000000000000
--- a/checkpoint-900/scaler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:4a0b4230f34cfc1b81dc2c15ef8d265bdd348193f5a746ca2018df11549c7ac0
-size 1383
diff --git a/checkpoint-900/scheduler.pt b/checkpoint-900/scheduler.pt
deleted file mode 100644
index 3a45e5ac3857c0ff7f904160c334def573e3f023..0000000000000000000000000000000000000000
--- a/checkpoint-900/scheduler.pt
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:d2742f07e7e0a5eee305ae9b6ef159da498981c08dfbb1b6f435f9e063ddc810
-size 1465
diff --git a/checkpoint-900/sentence_bert_config.json b/checkpoint-900/sentence_bert_config.json
deleted file mode 100644
index 5fd10429389515d3e5cccdeda08cae5fea1ae82e..0000000000000000000000000000000000000000
--- a/checkpoint-900/sentence_bert_config.json
+++ /dev/null
@@ -1,4 +0,0 @@
-{
- "max_seq_length": 128,
- "do_lower_case": false
-}
\ No newline at end of file
diff --git a/checkpoint-900/special_tokens_map.json b/checkpoint-900/special_tokens_map.json
deleted file mode 100644
index b1879d702821e753ffe4245048eee415d54a9385..0000000000000000000000000000000000000000
--- a/checkpoint-900/special_tokens_map.json
+++ /dev/null
@@ -1,51 +0,0 @@
-{
- "bos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "cls_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "eos_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "mask_token": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "pad_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "sep_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- },
- "unk_token": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false
- }
-}
diff --git a/checkpoint-900/tokenizer.json b/checkpoint-900/tokenizer.json
deleted file mode 100644
index e3420945e193cc0791136cdc6e5cd69801c838af..0000000000000000000000000000000000000000
--- a/checkpoint-900/tokenizer.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:cad551d5600a84242d0973327029452a1e3672ba6313c2a3c3d69c4310e12719
-size 17082987
diff --git a/checkpoint-900/tokenizer_config.json b/checkpoint-900/tokenizer_config.json
deleted file mode 100644
index facf4436a8f11c26085c16a14f4e576853927a9e..0000000000000000000000000000000000000000
--- a/checkpoint-900/tokenizer_config.json
+++ /dev/null
@@ -1,65 +0,0 @@
-{
- "added_tokens_decoder": {
- "0": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "1": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "2": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "3": {
- "content": "",
- "lstrip": false,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- },
- "250001": {
- "content": "",
- "lstrip": true,
- "normalized": false,
- "rstrip": false,
- "single_word": false,
- "special": true
- }
- },
- "bos_token": "",
- "clean_up_tokenization_spaces": false,
- "cls_token": "",
- "do_lower_case": true,
- "eos_token": "",
- "extra_special_tokens": {},
- "mask_token": "",
- "max_length": 128,
- "model_max_length": 128,
- "pad_to_multiple_of": null,
- "pad_token": "",
- "pad_token_type_id": 0,
- "padding_side": "right",
- "sep_token": "",
- "stride": 0,
- "strip_accents": null,
- "tokenize_chinese_chars": true,
- "tokenizer_class": "BertTokenizer",
- "truncation_side": "right",
- "truncation_strategy": "longest_first",
- "unk_token": ""
-}
diff --git a/checkpoint-900/trainer_state.json b/checkpoint-900/trainer_state.json
deleted file mode 100644
index ec677bb6dc669732fc0f6684bba468776a00b483..0000000000000000000000000000000000000000
--- a/checkpoint-900/trainer_state.json
+++ /dev/null
@@ -1,194 +0,0 @@
-{
- "best_global_step": 900,
- "best_metric": 0.0012511691311374307,
- "best_model_checkpoint": "data/fine-tuned-sbert-sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2-original-adafactor/checkpoint-900",
- "epoch": 8.412177985948478,
- "eval_steps": 100,
- "global_step": 900,
- "is_hyper_param_search": false,
- "is_local_process_zero": true,
- "is_world_process_zero": true,
- "log_history": [
- {
- "epoch": 0.936768149882904,
- "eval_loss": 0.008251233026385307,
- "eval_runtime": 117.4457,
- "eval_samples_per_second": 2267.669,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9330529793864755,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6639679670333862,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9596591982248662,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.8990018609372358,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6536919474601746,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.8488676021429209,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.8846836847946726,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.913791126905881,
- "eval_steps_per_second": 0.46,
- "step": 100
- },
- {
- "epoch": 1.8711943793911008,
- "eval_loss": 0.004326523281633854,
- "eval_runtime": 118.308,
- "eval_samples_per_second": 2251.141,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9683099913640971,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.987669070948898,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9520018198362147,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6799858808517456,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9284143244509058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9445886468795847,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9595322710076808,
- "eval_steps_per_second": 0.456,
- "step": 200
- },
- {
- "epoch": 2.8056206088992974,
- "eval_loss": 0.002782753435894847,
- "eval_runtime": 117.8399,
- "eval_samples_per_second": 2260.083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9790110013892539,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7040826678276062,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9935758649482886,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9680662667809197,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7029732465744019,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9524469797852624,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9648143930767479,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9713401352745615,
- "eval_steps_per_second": 0.458,
- "step": 300
- },
- {
- "epoch": 3.740046838407494,
- "eval_loss": 0.0020659712608903646,
- "eval_runtime": 116.8077,
- "eval_samples_per_second": 2280.056,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9837419742424811,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7114190459251404,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9954100421733855,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.975348704810703,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6966520547866821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.963270232791414,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9687853426826509,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9820016049524246,
- "eval_steps_per_second": 0.462,
- "step": 400
- },
- {
- "epoch": 4.674473067915691,
- "grad_norm": 0.07067500799894333,
- "learning_rate": 2.9402515723270442e-05,
- "loss": 0.0224,
- "step": 500
- },
- {
- "epoch": 4.674473067915691,
- "eval_loss": 0.0016409169184044003,
- "eval_runtime": 117.7739,
- "eval_samples_per_second": 2261.35,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.986370292494274,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.996439193909599,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9792820044518008,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7391290664672852,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9691467317957321,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.975107979086156,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9834919179181474,
- "eval_steps_per_second": 0.459,
- "step": 500
- },
- {
- "epoch": 5.608899297423887,
- "eval_loss": 0.0014551315689459443,
- "eval_runtime": 117.5801,
- "eval_samples_per_second": 2265.077,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9884729470957083,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7460525035858154,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9969945004512654,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9824360661365067,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7435637712478638,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9738614226726382,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9805847418912745,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9842943941304597,
- "eval_steps_per_second": 0.459,
- "step": 600
- },
- {
- "epoch": 6.543325526932084,
- "eval_loss": 0.0013776659034192562,
- "eval_runtime": 117.6764,
- "eval_samples_per_second": 2263.223,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9893740847820374,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.997357375070481,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9838035826704058,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7209540009498596,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9758996171607873,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9822857142857143,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9853261492605755,
- "eval_steps_per_second": 0.459,
- "step": 700
- },
- {
- "epoch": 7.477751756440281,
- "eval_loss": 0.0013444514479488134,
- "eval_runtime": 117.3408,
- "eval_samples_per_second": 2269.696,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9898246536252018,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.7261425852775574,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9975494130839752,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9844654628833477,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.7227741479873657,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9769000718683564,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9845218986470993,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9844090335893615,
- "eval_steps_per_second": 0.46,
- "step": 800
- },
- {
- "epoch": 8.412177985948478,
- "eval_loss": 0.0012511691311374307,
- "eval_runtime": 117.668,
- "eval_samples_per_second": 2263.385,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy": 0.9902752224683663,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_accuracy_threshold": 0.685534656047821,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_ap": 0.9977460917001926,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1": 0.9852413242919824,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_f1_threshold": 0.6582455635070801,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_mcc": 0.9780277137066985,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_precision": 0.9794924087922049,
- "eval_sentence-transformers-paraphrase-multilingual-MiniLM-L12-v2_cosine_recall": 0.9910581222056631,
- "eval_steps_per_second": 0.459,
- "step": 900
- }
- ],
- "logging_steps": 500,
- "max_steps": 1060,
- "num_input_tokens_seen": 0,
- "num_train_epochs": 10,
- "save_steps": 100,
- "stateful_callbacks": {
- "EarlyStoppingCallback": {
- "args": {
- "early_stopping_patience": 2,
- "early_stopping_threshold": 0.0
- },
- "attributes": {
- "early_stopping_patience_counter": 0
- }
- },
- "TrainerControl": {
- "args": {
- "should_epoch_stop": false,
- "should_evaluate": false,
- "should_log": false,
- "should_save": true,
- "should_training_stop": false
- },
- "attributes": {}
- }
- },
- "total_flos": 0.0,
- "train_batch_size": 5000,
- "trial_name": null,
- "trial_params": null
-}
diff --git a/checkpoint-900/training_args.bin b/checkpoint-900/training_args.bin
deleted file mode 100644
index d9dc73de16e341766a62f00cd26c21c6f69c3391..0000000000000000000000000000000000000000
--- a/checkpoint-900/training_args.bin
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:319aaa354e613c6db82c6bb78290f3da04198ef2c7a75b61b314fa305ed33c45
-size 6033
diff --git a/checkpoint-900/unigram.json b/checkpoint-900/unigram.json
deleted file mode 100644
index 2faa9ec874108d53a017ff2c7ab98d155fb21a82..0000000000000000000000000000000000000000
--- a/checkpoint-900/unigram.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:da145b5e7700ae40f16691ec32a0b1fdc1ee3298db22a31ea55f57a966c4a65d
-size 14763260
diff --git a/model.safetensors b/model.safetensors
deleted file mode 100644
index 89f9b51779c36bf737f8aa0a3f8c108b9baf5fb7..0000000000000000000000000000000000000000
--- a/model.safetensors
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:4f4d32a2eafc839cb2ab10b136bf98c4d30bdad7f85e5f55ceafdf3a54a9e859
-size 470637416
diff --git a/tokenizer.json b/tokenizer.json
deleted file mode 100644
index e3420945e193cc0791136cdc6e5cd69801c838af..0000000000000000000000000000000000000000
--- a/tokenizer.json
+++ /dev/null
@@ -1,3 +0,0 @@
-version https://git-lfs.github.com/spec/v1
-oid sha256:cad551d5600a84242d0973327029452a1e3672ba6313c2a3c3d69c4310e12719
-size 17082987