Deaksh commited on
Commit
ba26c9a
·
verified ·
1 Parent(s): 5aa671e

Upload folder using huggingface_hub

Browse files
checkpoint-26/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "B-ADE",
16
+ "1": "B-AGE",
17
+ "2": "B-DRUG",
18
+ "3": "I-ADE",
19
+ "4": "I-AGE",
20
+ "5": "I-DRUG",
21
+ "6": "O"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "label2id": {
27
+ "B-ADE": 0,
28
+ "B-AGE": 1,
29
+ "B-DRUG": 2,
30
+ "I-ADE": 3,
31
+ "I-AGE": 4,
32
+ "I-DRUG": 5,
33
+ "O": 6
34
+ },
35
+ "layer_norm_eps": 1e-12,
36
+ "max_position_embeddings": 512,
37
+ "model_type": "bert",
38
+ "num_attention_heads": 12,
39
+ "num_hidden_layers": 12,
40
+ "pad_token_id": 0,
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.3.0",
43
+ "type_vocab_size": 2,
44
+ "use_cache": false,
45
+ "vocab_size": 28996
46
+ }
checkpoint-26/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9c2496469f79c1fb87c1591444607745d8e6108ca4adb7d45140bf2dff46af5
3
+ size 430923564
checkpoint-26/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:845bc460a8c73f457650d48c009c3db103f3efc570a6bf9285776f8bb6992256
3
+ size 861970123
checkpoint-26/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a5a7422dbe51806d882a2479da08afcb5a46f17dd87b6371d4e78b65d56b8ea
3
+ size 14645
checkpoint-26/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ebfcfb4944bd4fb2e03e1e04992fc304c0f65cf8dd985b64119612b6df24f2d
3
+ size 1465
checkpoint-26/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-26/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
checkpoint-26/trainer_state.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 26,
3
+ "best_metric": 0.06289733201265335,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final/checkpoint-26",
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 26,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_f1": 0.0,
15
+ "eval_loss": 0.06289733201265335,
16
+ "eval_precision": 0.0,
17
+ "eval_recall": 0.0,
18
+ "eval_runtime": 0.7698,
19
+ "eval_samples_per_second": 33.775,
20
+ "eval_steps_per_second": 5.196,
21
+ "step": 26
22
+ }
23
+ ],
24
+ "logging_steps": 50,
25
+ "max_steps": 78,
26
+ "num_input_tokens_seen": 0,
27
+ "num_train_epochs": 3,
28
+ "save_steps": 500,
29
+ "stateful_callbacks": {
30
+ "TrainerControl": {
31
+ "args": {
32
+ "should_epoch_stop": false,
33
+ "should_evaluate": false,
34
+ "should_log": false,
35
+ "should_save": true,
36
+ "should_training_stop": false
37
+ },
38
+ "attributes": {}
39
+ }
40
+ },
41
+ "total_flos": 27176091131904.0,
42
+ "train_batch_size": 4,
43
+ "trial_name": null,
44
+ "trial_params": null
45
+ }
checkpoint-26/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2e33e0687b2225b37b3b842824292cee58a877110d3e61b73275919a0a2f4b1
3
+ size 5201
checkpoint-52/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "B-ADE",
16
+ "1": "B-AGE",
17
+ "2": "B-DRUG",
18
+ "3": "I-ADE",
19
+ "4": "I-AGE",
20
+ "5": "I-DRUG",
21
+ "6": "O"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "label2id": {
27
+ "B-ADE": 0,
28
+ "B-AGE": 1,
29
+ "B-DRUG": 2,
30
+ "I-ADE": 3,
31
+ "I-AGE": 4,
32
+ "I-DRUG": 5,
33
+ "O": 6
34
+ },
35
+ "layer_norm_eps": 1e-12,
36
+ "max_position_embeddings": 512,
37
+ "model_type": "bert",
38
+ "num_attention_heads": 12,
39
+ "num_hidden_layers": 12,
40
+ "pad_token_id": 0,
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.3.0",
43
+ "type_vocab_size": 2,
44
+ "use_cache": false,
45
+ "vocab_size": 28996
46
+ }
checkpoint-52/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06f3e92579bd3ee0f5b4a8dc8d5709068f2ef100851ffb0bba272814494b1170
3
+ size 430923564
checkpoint-52/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:356b5d063ccb0d1429b245b503f250811f8e547f4d06bc214165902cb83c3524
3
+ size 861970123
checkpoint-52/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7e6c0e8e273c2f6b50af5caac50a1b7f627da4cf5222dd263a292a90c1d573e
3
+ size 14645
checkpoint-52/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bea3c51af45d25a7261b4b0df286c324e8be04554f45df38e96d91bc84fdda87
3
+ size 1465
checkpoint-52/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-52/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
checkpoint-52/trainer_state.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 52,
3
+ "best_metric": 0.049389105290174484,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final/checkpoint-52",
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 52,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_f1": 0.0,
15
+ "eval_loss": 0.06289733201265335,
16
+ "eval_precision": 0.0,
17
+ "eval_recall": 0.0,
18
+ "eval_runtime": 0.7698,
19
+ "eval_samples_per_second": 33.775,
20
+ "eval_steps_per_second": 5.196,
21
+ "step": 26
22
+ },
23
+ {
24
+ "epoch": 1.9230769230769231,
25
+ "grad_norm": 0.09649301320314407,
26
+ "learning_rate": 1.858974358974359e-05,
27
+ "loss": 0.15085140228271485,
28
+ "step": 50
29
+ },
30
+ {
31
+ "epoch": 2.0,
32
+ "eval_f1": 0.1904761904761905,
33
+ "eval_loss": 0.049389105290174484,
34
+ "eval_precision": 1.0,
35
+ "eval_recall": 0.10526315789473684,
36
+ "eval_runtime": 0.7878,
37
+ "eval_samples_per_second": 33.002,
38
+ "eval_steps_per_second": 5.077,
39
+ "step": 52
40
+ }
41
+ ],
42
+ "logging_steps": 50,
43
+ "max_steps": 78,
44
+ "num_input_tokens_seen": 0,
45
+ "num_train_epochs": 3,
46
+ "save_steps": 500,
47
+ "stateful_callbacks": {
48
+ "TrainerControl": {
49
+ "args": {
50
+ "should_epoch_stop": false,
51
+ "should_evaluate": false,
52
+ "should_log": false,
53
+ "should_save": true,
54
+ "should_training_stop": false
55
+ },
56
+ "attributes": {}
57
+ }
58
+ },
59
+ "total_flos": 54352182263808.0,
60
+ "train_batch_size": 4,
61
+ "trial_name": null,
62
+ "trial_params": null
63
+ }
checkpoint-52/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2e33e0687b2225b37b3b842824292cee58a877110d3e61b73275919a0a2f4b1
3
+ size 5201
checkpoint-78/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "B-ADE",
16
+ "1": "B-AGE",
17
+ "2": "B-DRUG",
18
+ "3": "I-ADE",
19
+ "4": "I-AGE",
20
+ "5": "I-DRUG",
21
+ "6": "O"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "label2id": {
27
+ "B-ADE": 0,
28
+ "B-AGE": 1,
29
+ "B-DRUG": 2,
30
+ "I-ADE": 3,
31
+ "I-AGE": 4,
32
+ "I-DRUG": 5,
33
+ "O": 6
34
+ },
35
+ "layer_norm_eps": 1e-12,
36
+ "max_position_embeddings": 512,
37
+ "model_type": "bert",
38
+ "num_attention_heads": 12,
39
+ "num_hidden_layers": 12,
40
+ "pad_token_id": 0,
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.3.0",
43
+ "type_vocab_size": 2,
44
+ "use_cache": false,
45
+ "vocab_size": 28996
46
+ }
checkpoint-78/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4d7f02d1f3246386da560af52aee7ca7aecdb423d0707a4c5b6dd8eef7dabcc
3
+ size 430923564
checkpoint-78/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90c4a48b2641c9e4428ed7beef0d4df29041494b046b605b0b37dbe3fd833630
3
+ size 861970123
checkpoint-78/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79ed8d45a4499bcffc47aab390786f5185fc3d7b706539fb18806d87c243fb0d
3
+ size 14645
checkpoint-78/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bf819bf62e0d94d7d30d87fe36e9317a4be1fad217925e872ff99a8214e2e9d
3
+ size 1465
checkpoint-78/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-78/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
checkpoint-78/trainer_state.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 78,
3
+ "best_metric": 0.04243722930550575,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/ADEGuard/ner_biobert_final/checkpoint-78",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 78,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.0,
14
+ "eval_f1": 0.0,
15
+ "eval_loss": 0.06289733201265335,
16
+ "eval_precision": 0.0,
17
+ "eval_recall": 0.0,
18
+ "eval_runtime": 0.7698,
19
+ "eval_samples_per_second": 33.775,
20
+ "eval_steps_per_second": 5.196,
21
+ "step": 26
22
+ },
23
+ {
24
+ "epoch": 1.9230769230769231,
25
+ "grad_norm": 0.09649301320314407,
26
+ "learning_rate": 1.858974358974359e-05,
27
+ "loss": 0.15085140228271485,
28
+ "step": 50
29
+ },
30
+ {
31
+ "epoch": 2.0,
32
+ "eval_f1": 0.1904761904761905,
33
+ "eval_loss": 0.049389105290174484,
34
+ "eval_precision": 1.0,
35
+ "eval_recall": 0.10526315789473684,
36
+ "eval_runtime": 0.7878,
37
+ "eval_samples_per_second": 33.002,
38
+ "eval_steps_per_second": 5.077,
39
+ "step": 52
40
+ },
41
+ {
42
+ "epoch": 3.0,
43
+ "eval_f1": 0.1904761904761905,
44
+ "eval_loss": 0.04243722930550575,
45
+ "eval_precision": 1.0,
46
+ "eval_recall": 0.10526315789473684,
47
+ "eval_runtime": 0.8159,
48
+ "eval_samples_per_second": 31.867,
49
+ "eval_steps_per_second": 4.903,
50
+ "step": 78
51
+ }
52
+ ],
53
+ "logging_steps": 50,
54
+ "max_steps": 78,
55
+ "num_input_tokens_seen": 0,
56
+ "num_train_epochs": 3,
57
+ "save_steps": 500,
58
+ "stateful_callbacks": {
59
+ "TrainerControl": {
60
+ "args": {
61
+ "should_epoch_stop": false,
62
+ "should_evaluate": false,
63
+ "should_log": false,
64
+ "should_save": true,
65
+ "should_training_stop": true
66
+ },
67
+ "attributes": {}
68
+ }
69
+ },
70
+ "total_flos": 81528273395712.0,
71
+ "train_batch_size": 4,
72
+ "trial_name": null,
73
+ "trial_params": null
74
+ }
checkpoint-78/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2e33e0687b2225b37b3b842824292cee58a877110d3e61b73275919a0a2f4b1
3
+ size 5201
config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": null,
8
+ "classifier_dropout": null,
9
+ "dtype": "float32",
10
+ "eos_token_id": null,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "B-ADE",
16
+ "1": "B-AGE",
17
+ "2": "B-DRUG",
18
+ "3": "I-ADE",
19
+ "4": "I-AGE",
20
+ "5": "I-DRUG",
21
+ "6": "O"
22
+ },
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "label2id": {
27
+ "B-ADE": 0,
28
+ "B-AGE": 1,
29
+ "B-DRUG": 2,
30
+ "I-ADE": 3,
31
+ "I-AGE": 4,
32
+ "I-DRUG": 5,
33
+ "O": 6
34
+ },
35
+ "layer_norm_eps": 1e-12,
36
+ "max_position_embeddings": 512,
37
+ "model_type": "bert",
38
+ "num_attention_heads": 12,
39
+ "num_hidden_layers": 12,
40
+ "pad_token_id": 0,
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.3.0",
43
+ "type_vocab_size": 2,
44
+ "use_cache": false,
45
+ "vocab_size": 28996
46
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4d7f02d1f3246386da560af52aee7ca7aecdb423d0707a4c5b6dd8eef7dabcc
3
+ size 430923564
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "is_local": false,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2e33e0687b2225b37b3b842824292cee58a877110d3e61b73275919a0a2f4b1
3
+ size 5201