karths commited on
Commit
5ebfe65
·
verified ·
1 Parent(s): 3aea6a0

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ reduced_main_data.csv filter=lfs diff=lfs merge=lfs -text
37
+ test_data_for_future_evaluation.csv filter=lfs diff=lfs merge=lfs -text
38
+ test_top_repo_data.csv filter=lfs diff=lfs merge=lfs -text
39
+ top_repo_data.csv filter=lfs diff=lfs merge=lfs -text
checkpoint-1295/config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilroberta-base",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 514,
17
+ "model_type": "roberta",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 6,
20
+ "pad_token_id": 1,
21
+ "position_embedding_type": "absolute",
22
+ "problem_type": "single_label_classification",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.35.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50265
28
+ }
checkpoint-1295/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71762df7a4007e0cd16a712ecf2c803d4334d29df44b46f7ea5eef40f09eb07f
3
+ size 328492280
checkpoint-1295/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f232a70322a5f80d2df3fb410c2d207ecc26bbc8182be23ddfd71321bc81425e
3
+ size 657047610
checkpoint-1295/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c81138287efd999bff45e325170cf8ffeb3e537bcbe17f3d643783d70f2d90dc
3
+ size 14244
checkpoint-1295/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37d10fb60cc0b5e5d712c71b21e1cf036d736b3a00613110e07f765e44005b6c
3
+ size 1064
checkpoint-1295/trainer_state.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 5.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1295,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.93,
13
+ "learning_rate": 4.9800000000000004e-05,
14
+ "loss": 0.0383,
15
+ "step": 500
16
+ },
17
+ {
18
+ "epoch": 3.86,
19
+ "learning_rate": 1.5330327944282913e-05,
20
+ "loss": 0.0473,
21
+ "step": 1000
22
+ }
23
+ ],
24
+ "logging_steps": 500,
25
+ "max_steps": 1295,
26
+ "num_train_epochs": 5,
27
+ "save_steps": 500,
28
+ "total_flos": 1.096565126074368e+16,
29
+ "trial_name": null,
30
+ "trial_params": null
31
+ }
checkpoint-1295/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86cac2c00747813b4bbd3037c16347abce94f1c04f3fda003b7f9078ae13ee2e
3
+ size 4664
classification_report.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"0": {"precision": 0.995985727029438, "recall": 0.9942119323241317, "f1-score": 0.9950980392156862, "support": 2246}, "1": {"precision": 0.9931470743278862, "recall": 0.9952456418383518, "f1-score": 0.9941952506596307, "support": 1893}, "accuracy": 0.9946847064508335, "macro avg": {"precision": 0.9945664006786621, "recall": 0.9947287870812418, "f1-score": 0.9946466449376584, "support": 4139}, "weighted avg": {"precision": 0.9946874497730868, "recall": 0.9946847064508335, "f1-score": 0.9946851426859417, "support": 4139}}
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilroberta-base",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 514,
17
+ "model_type": "roberta",
18
+ "num_attention_heads": 12,
19
+ "num_hidden_layers": 6,
20
+ "pad_token_id": 1,
21
+ "position_embedding_type": "absolute",
22
+ "problem_type": "single_label_classification",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.35.0",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50265
28
+ }
confusion_matrix.png ADDED
detailed_confusion_matrix.png ADDED
fold_results.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "0": {
3
+ "eval_loss": 0.5350394248962402,
4
+ "eval_precision": 0.8132691278758695,
5
+ "eval_recall": 0.8029582673005811,
6
+ "eval_acc": 0.825561729886446,
7
+ "eval_mcc": 0.6482610618664716,
8
+ "eval_f1": 0.8080808080808081,
9
+ "eval_auc": 0.9079088303488646,
10
+ "eval_runtime": 45.6804,
11
+ "eval_samples_per_second": 90.608,
12
+ "eval_steps_per_second": 1.423,
13
+ "epoch": 5.0
14
+ },
15
+ "1": {
16
+ "eval_loss": 0.17293670773506165,
17
+ "eval_precision": 0.9322647362978284,
18
+ "eval_recall": 0.9524564183835182,
19
+ "eval_acc": 0.9466054602561005,
20
+ "eval_mcc": 0.8927884505222072,
21
+ "eval_f1": 0.9422524170368435,
22
+ "eval_auc": 0.9888017154638711,
23
+ "eval_runtime": 44.3128,
24
+ "eval_samples_per_second": 93.404,
25
+ "eval_steps_per_second": 1.467,
26
+ "epoch": 5.0
27
+ },
28
+ "2": {
29
+ "eval_loss": 0.038910072296857834,
30
+ "eval_precision": 0.9889357218124342,
31
+ "eval_recall": 0.9915478077126255,
32
+ "eval_acc": 0.991060642667311,
33
+ "eval_mcc": 0.9819969315787961,
34
+ "eval_f1": 0.9902400422052229,
35
+ "eval_auc": 0.9991457490430837,
36
+ "eval_runtime": 43.5039,
37
+ "eval_samples_per_second": 95.141,
38
+ "eval_steps_per_second": 1.494,
39
+ "epoch": 5.0
40
+ },
41
+ "3": {
42
+ "eval_loss": 0.02266552299261093,
43
+ "eval_precision": 0.9905263157894737,
44
+ "eval_recall": 0.99418911780243,
45
+ "eval_acc": 0.9929934766851897,
46
+ "eval_mcc": 0.9858940992776017,
47
+ "eval_f1": 0.9923543369364619,
48
+ "eval_auc": 0.9997593891164853,
49
+ "eval_runtime": 43.2885,
50
+ "eval_samples_per_second": 95.614,
51
+ "eval_steps_per_second": 1.502,
52
+ "epoch": 5.0
53
+ },
54
+ "4": {
55
+ "eval_loss": 0.02025098353624344,
56
+ "eval_precision": 0.9931470743278862,
57
+ "eval_recall": 0.9952456418383518,
58
+ "eval_acc": 0.9946847064508335,
59
+ "eval_mcc": 0.9892951744325653,
60
+ "eval_f1": 0.9941952506596307,
61
+ "eval_auc": 0.9998631128697893,
62
+ "eval_runtime": 43.5093,
63
+ "eval_samples_per_second": 95.129,
64
+ "eval_steps_per_second": 1.494,
65
+ "epoch": 5.0
66
+ }
67
+ }
metrics.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"precision": 0.9931470743278862, "recall": 0.9952456418383518, "acc": 0.9946847064508335, "mcc": 0.9892951744325653, "f1": 0.9941952506596307, "auc": 0.9998631128697893}
metrics_all_fold.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "precision": [
3
+ 0.8132691278758695,
4
+ 0.9322647362978284,
5
+ 0.9889357218124342,
6
+ 0.9905263157894737,
7
+ 0.9931470743278862
8
+ ],
9
+ "recall": [
10
+ 0.8029582673005811,
11
+ 0.9524564183835182,
12
+ 0.9915478077126255,
13
+ 0.99418911780243,
14
+ 0.9952456418383518
15
+ ],
16
+ "f1": [
17
+ 0.8080808080808081,
18
+ 0.9422524170368435,
19
+ 0.9902400422052229,
20
+ 0.9923543369364619,
21
+ 0.9941952506596307
22
+ ],
23
+ "auc": [
24
+ 0.9079088303488646,
25
+ 0.9888017154638711,
26
+ 0.9991457490430837,
27
+ 0.9997593891164853,
28
+ 0.9998631128697893
29
+ ],
30
+ "acc": [
31
+ 0.825561729886446,
32
+ 0.9466054602561005,
33
+ 0.991060642667311,
34
+ 0.9929934766851897,
35
+ 0.9946847064508335
36
+ ],
37
+ "mcc": [
38
+ 0.6482610618664716,
39
+ 0.8927884505222072,
40
+ 0.9819969315787961,
41
+ 0.9858940992776017,
42
+ 0.9892951744325653
43
+ ]
44
+ }
metrics_ci_bounds.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "precision": {
3
+ "ci_lower": 0.8477971084000429,
4
+ "ci_upper": 1.0394600820413542
5
+ },
6
+ "recall": {
7
+ "ci_lower": 0.8446704293600269,
8
+ "ci_upper": 1.0498884718549757
9
+ },
10
+ "f1": {
11
+ "ci_lower": 0.8463580286574575,
12
+ "ci_upper": 1.0444911133101293
13
+ },
14
+ "auc": {
15
+ "ci_lower": 0.9293436948474769,
16
+ "ci_upper": 1.0288478238893604
17
+ },
18
+ "acc": {
19
+ "ci_lower": 0.8601554679308618,
20
+ "ci_upper": 1.0402069384474903
21
+ },
22
+ "mcc": {
23
+ "ci_lower": 0.7181147649015004,
24
+ "ci_upper": 1.0811795221695566
25
+ }
26
+ }
metrics_mean.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "precision": 0.9436285952206985,
3
+ "recall": 0.9472794506075013,
4
+ "f1": 0.9454245709837934,
5
+ "auc": 0.9790957593684186,
6
+ "acc": 0.9501812031891761,
7
+ "mcc": 0.8996471435355284
8
+ }
metrics_std.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "precision": 0.0771798867964995,
3
+ "recall": 0.08263831551523768,
4
+ "f1": 0.07978530622587837,
5
+ "auc": 0.040068862907287915,
6
+ "acc": 0.07250410367744298,
7
+ "mcc": 0.1462008875965148
8
+ }
metrics_visualisation.png ADDED
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71762df7a4007e0cd16a712ecf2c803d4334d29df44b46f7ea5eef40f09eb07f
3
+ size 328492280
precision_recall_curve.png ADDED
reduced_main_data.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2524e9883e79fbb5a0d8df8b8ae42bc8bcec231b87b4ac0ed87adfad6574c84b
3
+ size 66623480
roc_curve.png ADDED
test_data_for_future_evaluation.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b56d2a39247149624eba9c7d76c0b9449d9c4b3ef7b1b831cb293a5b5ae62d92
3
+ size 16548925
test_top_repo_data.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c88a9c9a4333860e88c3d35f029aec742bd73dc3d8a53b0619af1a8992b9962
3
+ size 31962041
top_repo_data.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cc1c649501a6cf87f7b064645b05d5cb8ae6b71052feb3f280af457e0ba80ef
3
+ size 23750158
tracker_carbon_statistics.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cloud_provider": "",
3
+ "cloud_region": "",
4
+ "codecarbon_version": "2.3.4",
5
+ "country_iso_code": "NOR",
6
+ "country_name": "Norway",
7
+ "cpu_count": 192,
8
+ "cpu_energy": 0.15044465764479648,
9
+ "cpu_model": "AMD EPYC 7642 48-Core Processor",
10
+ "cpu_power": 73.92069695676452,
11
+ "duration": 3457.3639595508575,
12
+ "emissions": 0.028483843035557094,
13
+ "emissions_rate": 8.238601249044489e-06,
14
+ "energy_consumed": 1.0338962989312919,
15
+ "gpu_count": 4,
16
+ "gpu_energy": 0.5268581956528848,
17
+ "gpu_model": "4 x NVIDIA GeForce RTX 3090",
18
+ "gpu_power": 526.2912881226167,
19
+ "latitude": 59.9016,
20
+ "longitude": 10.7343,
21
+ "on_cloud": "N",
22
+ "os": "Linux-4.18.0-513.18.1.el8_9.x86_64-x86_64-with-glibc2.28",
23
+ "project_name": "codecarbon",
24
+ "pue": 1.0,
25
+ "python_version": "3.10.8",
26
+ "ram_energy": 0.3565934456336106,
27
+ "ram_power": 377.6938190460205,
28
+ "ram_total_size": 1007.1835174560547,
29
+ "region": "oslo county",
30
+ "run_id": "aedaa80b-949e-4b9a-bd8c-8534e37a0135",
31
+ "timestamp": "2024-04-05T21:48:47",
32
+ "tracking_mode": "machine"
33
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86cac2c00747813b4bbd3037c16347abce94f1c04f3fda003b7f9078ae13ee2e
3
+ size 4664