taeuk1 commited on
Commit
a8b54ba
·
verified ·
1 Parent(s): 95711ff

add binary vulnerability model

Browse files
README.md ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ base_model: microsoft/codebert-base
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - f1
9
+ model-index:
10
+ - name: codebert-bin
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # codebert-bin
18
+
19
+ This model is a fine-tuned version of [microsoft/codebert-base](https://huggingface.co/microsoft/codebert-base) on the None dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.0
22
+ - Accuracy: 1.0
23
+ - F1: 1.0
24
+ - Roc Auc: 1.0
25
+
26
+ ## Model description
27
+
28
+ More information needed
29
+
30
+ ## Intended uses & limitations
31
+
32
+ More information needed
33
+
34
+ ## Training and evaluation data
35
+
36
+ More information needed
37
+
38
+ ## Training procedure
39
+
40
+ ### Training hyperparameters
41
+
42
+ The following hyperparameters were used during training:
43
+ - learning_rate: 2e-05
44
+ - train_batch_size: 16
45
+ - eval_batch_size: 16
46
+ - seed: 42
47
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
+ - lr_scheduler_type: linear
49
+ - num_epochs: 4
50
+
51
+ ### Training results
52
+
53
+
54
+
55
+ ### Framework versions
56
+
57
+ - Transformers 4.55.0
58
+ - Pytorch 2.7.0+cu128
59
+ - Datasets 4.0.0
60
+ - Tokenizers 0.21.2
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RobertaForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "non_vuln",
14
+ "1": "vuln"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 3072,
18
+ "label2id": {
19
+ "non_vuln": 0,
20
+ "vuln": 1
21
+ },
22
+ "layer_norm_eps": 1e-05,
23
+ "max_position_embeddings": 514,
24
+ "model_type": "roberta",
25
+ "num_attention_heads": 12,
26
+ "num_hidden_layers": 12,
27
+ "output_past": true,
28
+ "pad_token_id": 1,
29
+ "position_embedding_type": "absolute",
30
+ "torch_dtype": "float32",
31
+ "transformers_version": "4.55.0",
32
+ "type_vocab_size": 1,
33
+ "use_cache": true,
34
+ "vocab_size": 50265
35
+ }
final_metrics.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss": 0.0,
3
+ "eval_accuracy": 1.0,
4
+ "eval_f1": 1.0,
5
+ "eval_roc_auc": 1.0,
6
+ "eval_runtime": 408.2672,
7
+ "eval_samples_per_second": 42.71,
8
+ "eval_steps_per_second": 2.67,
9
+ "epoch": 4.0
10
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
metrics.csv ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,step,split,loss,accuracy,f1,roc_auc
2
+ 0.05734602592040371,500,train,0.034,,,
3
+ 0.11469205184080743,1000,train,0.0001,,,
4
+ 0.17203807776121116,1500,train,0.0,,,
5
+ 0.22938410368161485,2000,train,0.0,,,
6
+ 0.2867301296020186,2500,train,0.0,,,
7
+ 0.3440761555224223,3000,train,0.0,,,
8
+ 0.401422181442826,3500,train,0.0,,,
9
+ 0.4587682073632297,4000,train,0.0,,,
10
+ 0.5161142332836335,4500,train,0.0,,,
11
+ 0.5734602592040372,5000,train,0.0,,,
12
+ 0.6308062851244409,5500,train,0.0,,,
13
+ 0.6881523110448446,6000,train,0.0,,,
14
+ 0.7454983369652483,6500,train,0.0,,,
15
+ 0.802844362885652,7000,train,0.0,,,
16
+ 0.8601903888060557,7500,train,0.0,,,
17
+ 0.9175364147264594,8000,train,0.0,,,
18
+ 0.9748824406468631,8500,train,0.0,,,
19
+ 1.032228466567267,9000,train,0.0,,,
20
+ 1.0895744924876707,9500,train,0.0,,,
21
+ 1.1469205184080744,10000,train,0.0,,,
22
+ 1.2042665443284781,10500,train,0.0,,,
23
+ 1.2616125702488818,11000,train,0.0,,,
24
+ 1.3189585961692853,11500,train,0.0,,,
25
+ 1.3763046220896893,12000,train,0.0,,,
26
+ 1.4336506480100928,12500,train,0.0,,,
27
+ 1.4909966739304967,13000,train,0.0,,,
28
+ 1.5483426998509002,13500,train,0.0,,,
29
+ 1.6056887257713042,14000,train,0.0,,,
30
+ 1.6630347516917077,14500,train,0.0,,,
31
+ 1.7203807776121116,15000,train,0.0,,,
32
+ 1.777726803532515,15500,train,0.0,,,
33
+ 1.835072829452919,16000,train,0.0,,,
34
+ 1.8924188553733226,16500,train,0.0,,,
35
+ 1.9497648812937265,17000,train,0.0,,,
36
+ 2.00711090721413,17500,train,0.0,,,
37
+ 2.064456933134534,18000,train,0.0,,,
38
+ 2.1218029590549374,18500,train,0.0,,,
39
+ 2.1791489849753414,19000,train,0.0,,,
40
+ 2.236495010895745,19500,train,0.0,,,
41
+ 2.293841036816149,20000,train,0.0,,,
42
+ 2.3511870627365523,20500,train,0.0,,,
43
+ 2.4085330886569563,21000,train,0.0,,,
44
+ 2.4658791145773598,21500,train,0.0,,,
45
+ 2.5232251404977637,22000,train,0.0,,,
46
+ 2.580571166418167,22500,train,0.0,,,
47
+ 2.6379171923385707,23000,train,0.0,,,
48
+ 2.6952632182589746,23500,train,0.0,,,
49
+ 2.7526092441793786,24000,train,0.0,,,
50
+ 2.809955270099782,24500,train,0.0,,,
51
+ 2.8673012960201856,25000,train,0.0,,,
52
+ 2.9246473219405895,25500,train,0.0,,,
53
+ 2.9819933478609935,26000,train,0.0,,,
54
+ 3.039339373781397,26500,train,0.0,,,
55
+ 3.0966853997018005,27000,train,0.0,,,
56
+ 3.1540314256222044,27500,train,0.0,,,
57
+ 3.211377451542608,28000,train,0.0,,,
58
+ 3.268723477463012,28500,train,0.0,,,
59
+ 3.3260695033834153,29000,train,0.0,,,
60
+ 3.3834155293038193,29500,train,0.0,,,
61
+ 3.440761555224223,30000,train,0.0,,,
62
+ 3.4981075811446267,30500,train,0.0,,,
63
+ 3.55545360706503,31000,train,0.0,,,
64
+ 3.612799632985434,31500,train,0.0,,,
65
+ 3.6701456589058377,32000,train,0.0,,,
66
+ 3.7274916848262416,32500,train,0.0,,,
67
+ 3.784837710746645,33000,train,0.0,,,
68
+ 3.842183736667049,33500,train,0.0,,,
69
+ 3.8995297625874525,34000,train,0.0,,,
70
+ 3.9568757885078565,34500,train,0.0,,,
71
+ 4.0,34876,eval,0.0,1.0,1.0,1.0
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8eb4944872d2fb9f6c02f497cd853979857fdc1134e06db9d0b7fd8df1f16ed2
3
+ size 498612824
runs/Aug11_11-25-38_ip-172-31-31-181.ap-northeast-2.compute.internal/events.out.tfevents.1754911539.ip-172-31-31-181.ap-northeast-2.compute.internal.65513.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc0d60ba92d1046a2a797f2e1a85e507d56970fad30868c5bb7d03bacb782cd2
3
+ size 5122
runs/Aug11_11-27-05_ip-172-31-31-181.ap-northeast-2.compute.internal/events.out.tfevents.1754911626.ip-172-31-31-181.ap-northeast-2.compute.internal.65607.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0d78eeb385fc651b31f34fb88b4d360007197aa0d76b89e8ee4a320e8bc622c
3
+ size 20189
runs/Aug11_11-27-05_ip-172-31-31-181.ap-northeast-2.compute.internal/events.out.tfevents.1754959428.ip-172-31-31-181.ap-northeast-2.compute.internal.65607.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da606df193e383d750528ad0ffeafc662aa86f071683b2e5de23d8b7aeb349f0
3
+ size 516
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": false,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "extra_special_tokens": {},
51
+ "mask_token": "<mask>",
52
+ "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "RobertaTokenizer",
56
+ "trim_offsets": true,
57
+ "unk_token": "<unk>"
58
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a2183ab1fed67759b3e2e4bb336206f3c7835af748e2007b948d9338eaad553
3
+ size 5777
vocab.json ADDED
The diff for this file is too large to render. See raw diff