DISCLAIMER: For those of you who are downloading this model, it is not finished, the results are poor.
Question Answering Model applying fine tuning to a GPT2 text generator model in a Catalan Dataset "projecte-aina/catalanqa".
Results over the first epoch
200it [01:14, 2.29it/s]Train: wpb=10, num_updates=200, accuracy=2.5, loss=0.97
500it [02:57, 3.06it/s]Train: wpb=10, num_updates=500, accuracy=3.1, loss=0.98
1000it [05:47, 2.72it/s]Train: wpb=10, num_updates=1000, accuracy=3.7, loss=0.91
2000it [11:29, 3.32it/s]Train: wpb=10, num_updates=2000, accuracy=3.7, loss=0.85
3000it [16:48, 3.90it/s]Train: wpb=10, num_updates=3000, accuracy=3.7, loss=0.82
4000it [22:10, 3.06it/s]Train: wpb=10, num_updates=4000, accuracy=3.9, loss=0.79
5000it [27:24, 3.50it/s]Train: wpb=10, num_updates=5000, accuracy=4.1, loss=0.77
6000it [32:41, 2.19it/s]Train: wpb=10, num_updates=6000, accuracy=4.5, loss=0.76
7000it [37:56, 3.03it/s]Train: wpb=10, num_updates=7000, accuracy=4.6, loss=0.75
8000it [43:06, 3.73it/s]Train: wpb=10, num_updates=8000, accuracy=4.8, loss=0.74
9000it [48:28, 2.85it/s]Train: wpb=10, num_updates=9000, accuracy=4.9, loss=0.73
10000it [53:43, 2.89it/s]Train: wpb=10, num_updates=10000, accuracy=5.1, loss=0.73
11000it [59:09, 3.10it/s]Train: wpb=10, num_updates=11000, accuracy=5.2, loss=0.73
12000it [1:04:37, 2.64it/s]Train: wpb=10, num_updates=12000, accuracy=5.3, loss=0.72
13000it [1:10:02, 2.66it/s]Train: wpb=10, num_updates=13000, accuracy=5.4, loss=0.72
14000it [1:15:15, 2.68it/s]Train: wpb=10, num_updates=14000, accuracy=5.4, loss=0.72
14150it [1:16:05, 3.10it/s]
Train: wpb=9, num_updates=14150, accuracy=5.4, loss=0.72
| epoch 000 | train accuracy=5.4%, train loss=0.72
| epoch 000 | valid accuracy=7.6%, valid loss=0.69
200it [01:16, 2.21it/s]Train: wpb=10, num_updates=200, accuracy=4.5, loss=0.68
500it [03:02, 2.94it/s]Train: wpb=10, num_updates=500, accuracy=4.3, loss=0.74
1000it [05:59, 2.60it/s]Train: wpb=10, num_updates=1000, accuracy=4.9, loss=0.74
2000it [11:53, 3.18it/s]Train: wpb=10, num_updates=2000, accuracy=4.8, loss=0.74
3000it [17:24, 3.80it/s]Train: wpb=10, num_updates=3000, accuracy=4.8, loss=0.73
4000it [22:58, 2.96it/s]Train: wpb=10, num_updates=4000, accuracy=4.9, loss=0.72
5000it [28:23, 3.43it/s]Train: wpb=10, num_updates=5000, accuracy=5.0, loss=0.71
6000it [33:52, 2.15it/s]Train: wpb=10, num_updates=6000, accuracy=5.2, loss=0.70
7000it [39:18, 2.92it/s]Train: wpb=10, num_updates=7000, accuracy=5.3, loss=0.70
8000it [44:39, 3.63it/s]Train: wpb=10, num_updates=8000, accuracy=5.4, loss=0.69
9000it [50:13, 2.74it/s]Train: wpb=10, num_updates=9000, accuracy=5.5, loss=0.69
10000it [55:39, 2.84it/s]Train: wpb=10, num_updates=10000, accuracy=5.7, loss=0.69
11000it [1:01:16, 3.00it/s]Train: wpb=10, num_updates=11000, accuracy=5.7, loss=0.69
12000it [1:06:57, 2.54it/s]Train: wpb=10, num_updates=12000, accuracy=5.8, loss=0.69
13000it [1:12:33, 2.56it/s]Train: wpb=10, num_updates=13000, accuracy=5.8, loss=0.69
14000it [1:17:58, 2.56it/s]Train: wpb=10, num_updates=14000, accuracy=5.9, loss=0.69
14150it [1:18:49, 2.99it/s]
Train: wpb=9, num_updates=14150, accuracy=5.9, loss=0.69
| epoch 001 | train accuracy=5.9%, train loss=0.69
| epoch 001 | valid accuracy=7.7%, valid loss=0.69
- Downloads last month
- 16