malekradwan130 commited on
Commit
ebeee82
·
verified ·
1 Parent(s): b9b181d

Upload folder using huggingface_hub

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7353692ff981605c93cf22b096dbd246c5cf5bf5fc543b490cf3abeac82a7a31
3
  size 59001752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3136b9dabce6a4ae406f1d3498508a41caaee6e48e4f89e5e650e30fb89be72a
3
  size 59001752
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:149be6cff081d78c7d5b1ec35545a1530746b35e6a028f2e35055b6d9fbeb68c
3
  size 118086731
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:006d09f6e0c431d7ae5b9eb4fbaae5108b8d1e413f666ee8fca34360e86c1c0e
3
  size 118086731
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a3365689e6d26d46f2bfa2d42b0b54103889b4f6a6bad2f60268c5b6325623d
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e2db19db6959d99d7cf488b002f6d817e98d33ad082cb99f1481611b97a8b6d
3
  size 14645
scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3656c0abce9576d0b5083188f2c7d6efcbee0134d27c94bb8b76920474ad16de
3
  size 1383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bd565e4d2ccaef156f334400d2b3c0f963cb4267afcfb22f72d1c02240578ac
3
  size 1383
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bebaf2e9a58bbd47b2799ce9fd04cb408b13ca7015b9a1b4e2b2bcad498171ac
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5367250aa3dbe18a71a87c36f9a3373d602c91495e0cb8690451d95250cc0609
3
  size 1465
trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.035555028814387934,
6
  "eval_steps": 500,
7
- "global_step": 600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -92,6 +92,34 @@
92
  "learning_rate": 0.00048228253140554636,
93
  "loss": 1.0983,
94
  "step": 600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
95
  }
96
  ],
97
  "logging_steps": 50,
@@ -111,7 +139,7 @@
111
  "attributes": {}
112
  }
113
  },
114
- "total_flos": 4.1133136674816e+16,
115
  "train_batch_size": 2,
116
  "trial_name": null,
117
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.04740670508585058,
6
  "eval_steps": 500,
7
+ "global_step": 800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
92
  "learning_rate": 0.00048228253140554636,
93
  "loss": 1.0983,
94
  "step": 600
95
+ },
96
+ {
97
+ "epoch": 0.03851794788225359,
98
+ "grad_norm": 0.09815791249275208,
99
+ "learning_rate": 0.0004808011377103579,
100
+ "loss": 1.1235,
101
+ "step": 650
102
+ },
103
+ {
104
+ "epoch": 0.04148086695011926,
105
+ "grad_norm": 0.07672577351331711,
106
+ "learning_rate": 0.0004793197440151695,
107
+ "loss": 1.0744,
108
+ "step": 700
109
+ },
110
+ {
111
+ "epoch": 0.04444378601798492,
112
+ "grad_norm": 0.08625241369009018,
113
+ "learning_rate": 0.0004778679781938848,
114
+ "loss": 1.0897,
115
+ "step": 750
116
+ },
117
+ {
118
+ "epoch": 0.04740670508585058,
119
+ "grad_norm": 0.10188218951225281,
120
+ "learning_rate": 0.0004763865844986964,
121
+ "loss": 1.0737,
122
+ "step": 800
123
  }
124
  ],
125
  "logging_steps": 50,
 
139
  "attributes": {}
140
  }
141
  },
142
+ "total_flos": 5.4844182233088e+16,
143
  "train_batch_size": 2,
144
  "trial_name": null,
145
  "trial_params": null