SashaSheykina commited on
Commit
23d7a01
1 Parent(s): 91b014d

Training complete

Browse files
README.md CHANGED
@@ -15,16 +15,16 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [microsoft/codebert-base](https://huggingface.co/microsoft/codebert-base) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 18.2530
19
- - eval_rouge1: 0.0557
20
- - eval_rouge2: 0.0041
21
- - eval_rougeL: 0.0397
22
- - eval_bleu: 0.8313
23
- - eval_meteor: 0.0884
24
- - eval_codebleu: 0.2407
25
- - eval_runtime: 150.2026
26
- - eval_samples_per_second: 0.067
27
- - eval_steps_per_second: 0.013
28
  - step: 0
29
 
30
  ## Model description
@@ -45,8 +45,8 @@ More information needed
45
 
46
  The following hyperparameters were used during training:
47
  - learning_rate: 2e-05
48
- - train_batch_size: 8
49
- - eval_batch_size: 8
50
  - seed: 42
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
 
15
 
16
  This model is a fine-tuned version of [microsoft/codebert-base](https://huggingface.co/microsoft/codebert-base) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - eval_loss: 17.9546
19
+ - eval_rouge1: 0.0741
20
+ - eval_rouge2: 0.0047
21
+ - eval_rougeL: 0.0528
22
+ - eval_bleu: 1.0975
23
+ - eval_meteor: 0.0974
24
+ - eval_codebleu: {'codebleu': 0.2174788437391285, 'ngram_match_score': 0.0007172304133318851, 'weighted_ngram_match_score': 0.0015773932015006452, 'syntax_match_score': 0.07692307692307693, 'dataflow_match_score': 0.7906976744186046}
25
+ - eval_runtime: 40.1888
26
+ - eval_samples_per_second: 0.249
27
+ - eval_steps_per_second: 0.025
28
  - step: 0
29
 
30
  ## Model description
 
45
 
46
  The following hyperparameters were used during training:
47
  - learning_rate: 2e-05
48
+ - train_batch_size: 32
49
+ - eval_batch_size: 64
50
  - seed: 42
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
config.json CHANGED
@@ -12,6 +12,7 @@
12
  "hidden_size": 768,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 3072,
 
15
  "layer_norm_eps": 1e-05,
16
  "max_position_embeddings": 514,
17
  "model_type": "roberta",
 
12
  "hidden_size": 768,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 3072,
15
+ "is_decoder": true,
16
  "layer_norm_eps": 1e-05,
17
  "max_position_embeddings": 514,
18
  "model_type": "roberta",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:916dc519bdf65a462f90cb40768d07e186ba8a812a9687855fb967bd09f1873d
3
  size 498813948
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:221ef71ce2f5de5811df7c6ef40413e590833be4b780f714c6f9f41e4b52c141
3
  size 498813948
runs/Jul26_17-48-33_30aead859e7a/events.out.tfevents.1722016153.30aead859e7a.53813.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:920c1d3e92e4b7649b41a6e46a96942133b6009fe833cb4eb531d212d1917e44
3
+ size 530
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 128,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 128
12
- },
13
- "direction": "Left",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 2,
16
- "pad_type_id": 0,
17
- "pad_token": "</s>"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c63e64576549df8ac6dd7cfd5063324309b17aa0470e25f17ff207a0c85ea682
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aab5322319da1eef8a45a74e25425c600d2eccd011bc4dd74b8d87c79f119789
3
  size 5304