File size: 101,596 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
{
    "paper_id": "2020",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T04:33:23.322568Z"
    },
    "title": "The ELTE.DH Pilot Corpus -Creating a Handcrafted Gigaword Web Corpus with Metadata",
    "authors": [
        {
            "first": "Bal\u00e1zs",
            "middle": [],
            "last": "Indig",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
                "location": {
                    "addrLine": "M\u00fazeum krt. 6-8",
                    "postCode": "H-1088",
                    "settlement": "Budapest",
                    "country": "Hungary"
                }
            },
            "email": "[email protected]"
        },
        {
            "first": "\u00c1rp\u00e1d",
            "middle": [],
            "last": "Knap",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
                "location": {
                    "addrLine": "1/A",
                    "postCode": "H-1117",
                    "settlement": "Budapest",
                    "country": "Hungary"
                }
            },
            "email": "[email protected]"
        },
        {
            "first": "Zs\u00f3fia",
            "middle": [],
            "last": "S\u00e1rk\u00f6zi-Lindner",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
                "location": {
                    "addrLine": "M\u00fazeum krt. 6-8",
                    "postCode": "H-1088",
                    "settlement": "Budapest",
                    "country": "Hungary"
                }
            },
            "email": ""
        },
        {
            "first": "M\u00e1ria",
            "middle": [],
            "last": "Tim\u00e1ri",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
                "location": {
                    "addrLine": "M\u00fazeum krt. 6-8",
                    "postCode": "H-1088",
                    "settlement": "Budapest",
                    "country": "Hungary"
                }
            },
            "email": ""
        },
        {
            "first": "G\u00e1bor",
            "middle": [],
            "last": "Palk\u00f3",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
                "location": {
                    "addrLine": "M\u00fazeum krt. 6-8",
                    "postCode": "H-1088",
                    "settlement": "Budapest",
                    "country": "Hungary"
                }
            },
            "email": "[email protected]"
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "In this article, we present the method we used to create a middle-sized corpus using targeted web crawling. Our corpus contains news portal articles along with their metadata, that can be useful for diverse audiences, ranging from digital humanists to NLP users. The method presented in this paper applies rule-based components that allow the curation of the text and the metadata content. The curated data can thereon serve as a reference for various tasks and measurements. We designed our workflow to encourage modification and customisation. Our concept can also be applied to other genres of portals by using the discovered patterns in the architecture of the portals. We found that for a systematic creation or extension of a similar corpus, our method provides superior accuracy and ease of use compared to The Wayback Machine, while requiring minimal manpower and computational resources. Reproducing the corpus is possible if changes are introduced to the text-extraction process. The standard TEI format and Schema.org encoded metadata is used for the output format, but we stress that placing the corpus in a digital repository system is recommended in order to be able to define semantic relations between the segments and to add rich annotation.",
    "pdf_parse": {
        "paper_id": "2020",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "In this article, we present the method we used to create a middle-sized corpus using targeted web crawling. Our corpus contains news portal articles along with their metadata, that can be useful for diverse audiences, ranging from digital humanists to NLP users. The method presented in this paper applies rule-based components that allow the curation of the text and the metadata content. The curated data can thereon serve as a reference for various tasks and measurements. We designed our workflow to encourage modification and customisation. Our concept can also be applied to other genres of portals by using the discovered patterns in the architecture of the portals. We found that for a systematic creation or extension of a similar corpus, our method provides superior accuracy and ease of use compared to The Wayback Machine, while requiring minimal manpower and computational resources. Reproducing the corpus is possible if changes are introduced to the text-extraction process. The standard TEI format and Schema.org encoded metadata is used for the output format, but we stress that placing the corpus in a digital repository system is recommended in order to be able to define semantic relations between the segments and to add rich annotation.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "In the glossary of the handbook entitled The Digital Humanities, Gardiner and Musto (2015, 250) define web archiving as \"the process of collecting portions of the World Wide Web to ensure the information is preserved in an Archive for future researchers, historians and the public\". It is telling, however, that in the chapter focusing on digital archives as source materials of the present scholarly practices, born-digital archives and web archives are entirely omitted, as the authors solely speak about curated digital collections designed by (digital) archivists for the research community. Web archives are much less organised and curated then digital libraries or databases, and for this reason, are far less usable for (and used by) scholars. If Gardiner and Musto (2015) are right in their choice to emphasise the role of these digital sources in answering present scholarly questions, the fact that web archives do not play a significant role among these sources is a substantial problem for the digital humanities. There are several reasons why web archives are under-represented in the scholarly use of digital sources. The main reason is the lack of high-quality metadata, as source materials must have -among othersa publication date and its authors identified by the archival institution, otherwise, the reference to the material (be it paper-based or born-digital) is questionable 1 . The second reason is the uniqueness and authenticity of the records.",
                "cite_spans": [
                    {
                        "start": 65,
                        "end": 95,
                        "text": "Gardiner and Musto (2015, 250)",
                        "ref_id": null
                    },
                    {
                        "start": 754,
                        "end": 779,
                        "text": "Gardiner and Musto (2015)",
                        "ref_id": "BIBREF2"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "Web archives usually contain many nearly identical versions of the \"same\" resource. This problem is exacerbated by the nearly inseparable dirt (recurring boilerplate text) among relevant content. The drawbacks arising from the unstructured nature of a web archive hinder its integration into the network of digital cultural heritage (DCH). As suggested in (Weber, 2018) , the limitations of web archives can be described along two main dimensions: accuracy and completeness. It is very difficult to tell if an archive actually captures all the content on the web accurately related to a specific topic. Our method, by using websites' own archives, creates \"complete snapshots\" of their real content from time to time, which provides real populational data for the portals included in the project. This also means that the ELTE.DH corpus contains all documents from the selected portals' archives which were publicly available at crawling time. Beyond creating a corpus for NLP applications, our work focuses on providing solutions to the aforementioned issues by developing a trusted digital repository complying with Linked Open Data technology. Our goal with this repository is to meet the essential demands of NLP, DCH and other disciplines uniformly.",
                "cite_spans": [
                    {
                        "start": 356,
                        "end": 369,
                        "text": "(Weber, 2018)",
                        "ref_id": "BIBREF10"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1."
            },
            {
                "text": "When it comes to crawling, web archiving or corpus creation, there are a number of options. The ISO/TR 14873:2013 standard describes the details of such workflows, however, distinct disciplines have come up with their own solutions ignoring this standard or only partially adhering to it. Holding on to the terminologies of the standard, we have conducted selective web archiving that is enriched with more and better metadata compared to general crawling. We argue that our method has a smaller footprint while remaining easy to manage. This makes the whole workflow sustainable and scalable. In the following sections, we will review the already available tools and formats to place our solution among them.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Background",
                "sec_num": "2."
            },
            {
                "text": "The standardisation process of web archiving practices, initiated and controlled mainly by national libraries (Oury and Poll, 2013) , does not provide comprehensive guidelines to the standardised encoding of the texts extracted from web archiving activity. The situation is much better on the level of metadata. The technical report of Statistics and Quality Indicators for Web Archiving stresses the importance of different metadata types for curating web resources 2 : \"Long term preservation also includes keeping safe the metadata associated with the resources in the Web Archive, which are critical for supporting collection management, access and preservation activities\" (ISO/TC 46/SC 8 N). The Metadata Encoding and Transmission Standard (METS) distinguishes four metadata types to be used in curated collections sourced from web archiving: (a) Descriptive metadata, (b) Structural metadata, (c) Provenance metadata, (d) Rights metadata. This is the theoretical standpoint, but since the creation of such metadata requires a lot of manual work, it is impossible to find a collection of archived web documents that complies with these requirements on metadata entirely. Therefore, there is virtually no reliable digital cultural heritage source for researchers. In contrast, there are metadata standards which cover finegrained requirements. The only standard that could gain large-scale adoption is Dublin Core, which is not refined enough to comply with the aforementioned standards. Our repository uses Schema.org, a metadata standard we have chosen for several reasons:",
                "cite_spans": [
                    {
                        "start": 110,
                        "end": 131,
                        "text": "(Oury and Poll, 2013)",
                        "ref_id": "BIBREF7"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metadata",
                "sec_num": "2.1."
            },
            {
                "text": "\u2022 Schema.org is designed explicitly for storing information about web resources",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metadata",
                "sec_num": "2.1."
            },
            {
                "text": "\u2022 It has a dynamic, community based development (in contrast with robust standards, such as METS)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metadata",
                "sec_num": "2.1."
            },
            {
                "text": "\u2022 It is increasingly popular on the web, which makes it easy to extract metadata from the HTML source",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metadata",
                "sec_num": "2.1."
            },
            {
                "text": "\u2022 It is compatible with semantic web technology (Linked Open Data)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metadata",
                "sec_num": "2.1."
            },
            {
                "text": "\u2022 It has a growing usage in the digital cultural heritage domain (e.g. Europeana)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metadata",
                "sec_num": "2.1."
            },
            {
                "text": "The Szeged corpus is the largest, manually annotated corpus (Vincze et al., 2010) in the Hungarian language containing 1.2 million tokens, KorKorpusz (31,492 tokens) is similar but smaller corpus based on a recent pilot project (Vad\u00e1sz, 2020) . The first Hungarian gigaword corpus was the Hungarian Gigaword Corpus (Oravecz et al., 2014) with 1,532,933,778 tokens. Both aforementioned corpora contain text only from curated sources (newspapers, literary texts, social media, legal texts, etc.) that are not entirely from the Internet. The first Hungarian web corpus that was created by Kornai and his colleagues (Hal\u00e1csy et al., 2004) is called the Hungarian Webcorpus. It was later superseded by the 1.2 billion token P\u00e1zm\u00e1ny corpus 3 (Endr\u00e9dy and Pr\u00f3sz\u00e9ky, 2016) and the 2.5 billion token HuTenTen corpus (Jakub\u00ed\u010dek et al., 2013) , two larger corpora entirely from the web. Nowadays, large corpora are utilising the Common Crawl archive like the OSCAR corpus (Ortiz Su\u00e1rez et al., 2019) with 5.16 billion (2.33 billion deduplicated) words in Hungarian. However, the documents presented in these corpora often contain gaps due to deduplication. All of these corpora -except the ones based on Common Crawl -have the same flaw, namely that after their creation and publication, several errors were discovered in the tools used to create them, and these errors could not be corrected as expected. The reason being that their original source HTML files have been deleted -and these are unavailable in an unmodified form on the web.",
                "cite_spans": [
                    {
                        "start": 60,
                        "end": 81,
                        "text": "(Vincze et al., 2010)",
                        "ref_id": "BIBREF21"
                    },
                    {
                        "start": 228,
                        "end": 242,
                        "text": "(Vad\u00e1sz, 2020)",
                        "ref_id": "BIBREF20"
                    },
                    {
                        "start": 315,
                        "end": 337,
                        "text": "(Oravecz et al., 2014)",
                        "ref_id": "BIBREF18"
                    },
                    {
                        "start": 612,
                        "end": 634,
                        "text": "(Hal\u00e1csy et al., 2004)",
                        "ref_id": "BIBREF14"
                    },
                    {
                        "start": 736,
                        "end": 764,
                        "text": "(Endr\u00e9dy and Pr\u00f3sz\u00e9ky, 2016)",
                        "ref_id": "BIBREF13"
                    },
                    {
                        "start": 807,
                        "end": 831,
                        "text": "(Jakub\u00ed\u010dek et al., 2013)",
                        "ref_id": "BIBREF15"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Existing Hungarian Corpora",
                "sec_num": "2.2."
            },
            {
                "text": "Since then, there have been numerous attempts to create web-based corpora, but these were not published and could not arouse public interest, as web corpora and crawling became increasingly common tools. The speciality of the corpus and the method presented in this paper lies in the fact that it unites the experience from the above mentioned corpora into a manually curated gigaword web corpus, which includes metadata and can be built from the source of the downloaded web pages in a reproducible manner.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Existing Hungarian Corpora",
                "sec_num": "2.2."
            },
            {
                "text": "To put our method into a larger perspective, in the following sections we will describe the process of corpus creation in an abstract workflow (see Figure 1. ), where the elements have to be further specified by certain design decisions.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 148,
                        "end": 157,
                        "text": "Figure 1.",
                        "ref_id": "FIGREF0"
                    }
                ],
                "eq_spans": [],
                "section": "From the web to a corpus",
                "sec_num": "3."
            },
            {
                "text": "Classical web crawling can be characterised by only a few parameters that are set at the start of the process. These parameters include the initial seed of URLs where to start the crawl from, the maximal depth, and the breadth to restrict the crawler's movement. In some cases a set of targeted domains is also specified. Although there are only a few widely used crawler engines, it is hard to characterize these as most web libraries (e.g. Python requests, wget, etc.) can be used for crawling nowadays and the desired output varies from corpora to \"exact offline duplicates\" of websites. Here we would like to mention three crawler engines: both Heritix 4 and Apache Nutch (Laliwala and Shaikh, 2013) are used in the Internet Archive and Common Crawl projects. The third crawler engine is Spiderling (Suchomel and Pomik\u00e1lek, 2012) , which was developed by the authors of Sketch Engine (Kilgarriff et al., 2014) . These crawlers are fast, generalised tools, but for targeted or spe- cialised crawling they became tedious to use. This may explain the numerous different libraries used for crawling. Nowadays, we do not even have to use a crawler as we can begin the process with Common Crawl, The Internet Archive and similar resources. In this case, the first step is to clean the actual data, and remove collected garbage. Due to the nature of the internet, there are numerous aggressive SEO traps out in the web that are used to optimise the page rank in search engines that end up in the archive. These traps are placed in such a manner that they can divert crawler bots from their original course when these bots stumble upon them. Such general bots cannot distinguish \"normal\" pages from these traps, a task that humans are able to carry out in a matter of seconds. Another common problem using these sources is the need for deduplication (see Section 3.3.), which causes the waste of resources on both sides (crawler and deduplicator).",
                "cite_spans": [
                    {
                        "start": 676,
                        "end": 703,
                        "text": "(Laliwala and Shaikh, 2013)",
                        "ref_id": "BIBREF17"
                    },
                    {
                        "start": 803,
                        "end": 833,
                        "text": "(Suchomel and Pomik\u00e1lek, 2012)",
                        "ref_id": "BIBREF19"
                    },
                    {
                        "start": 888,
                        "end": 913,
                        "text": "(Kilgarriff et al., 2014)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Web Crawler",
                "sec_num": "3.1."
            },
            {
                "text": "To overcome these problems, Indig et al. (2019) built a new NLP-oriented crawler tool called Corpus Builder, which utilises a very different approach from the above: a twolevel crawling method. By using targeted crawling of large or medium portals, they claim that with their method, it is possible to crawl only the targeted portals virtually without duplication, with a small footprint and in a sustainable manner. Their main idea is the exploitation of two different levels of recurring patterns (\"archives\" and \"article\" pages) that can be discovered by analysing the structure of different web pages.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Web Crawler",
                "sec_num": "3.1."
            },
            {
                "text": "The first and obvious level is the recurring boilerplate content on the actual web pages of a portal. Objects of the same type are usually stored in the database back end and generated on demand into specifically formatted web pages. This is the internal nature of the data. In this paper, we call these pages \"articles\" regardless of their content type: whether they represent news articles, images with captions in a gallery, product descriptions and customer reviews in a webshop, posts in a forum or blog, etc. The output pages for these content types look the same in narrower time frames for a certain portal, but they can be very different from website to website. These pages are generated on the serverside, so we must collect HTMLs and extract their content. If we collect a mass amount of web pages representing the same objects from the same portal using the same template selectively or classify them after collection, the (uniform) boilerplate can be easily removed with a few simple rules per class, therefore this task does not require complex tools.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Article Crawler",
                "sec_num": "3.1.1."
            },
            {
                "text": "The second level arises from the first: how can we systematically collect such web pages on a given portal? The answer is very simple: portals created for human readers are very likely to have some kind of a \"table of contents\", \"product or topic list\" or an \"article archive\", which display all available pages in a structured form. Because objects are traditionally stored in taxonomies -e.g. temporal (years, months) or other feature-based (colour, shape, price, etc.)that can be enumerated and each object has finite number of possible values. If we enumerate articles for right feature values, we will gather links to all pages of the same layout systematically from the given portal.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Archive Crawler",
                "sec_num": "3.1.2."
            },
            {
                "text": "Using the two-step method described above, it is possible to gather a massive number of web pages even from only a small number of portals that will have virtually no duplication and effectively zero garbage pages in contrast to the general crawling methodology. This method has been successfully tested on three Hungarian news portals (Indig et al., 2019) , while the further generalisation of the method for the steps following the crawling of different portals with different schemes and layouts requires further elaboration. Indig et al. (2019) assembled the minimal number of parameters that are needed to handle such portals in a unified framework. The major highlights of the configuration are showcased as the following:",
                "cite_spans": [
                    {
                        "start": 336,
                        "end": 356,
                        "text": "(Indig et al., 2019)",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 529,
                        "end": 548,
                        "text": "Indig et al. (2019)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Possible Parameters for Portals",
                "sec_num": "3.1.3."
            },
            {
                "text": "\u2022 The date of the first and last article, or page number where applicable",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Possible Parameters for Portals",
                "sec_num": "3.1.3."
            },
            {
                "text": "\u2022 The archive URL format with the placeholders to be substituted",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Possible Parameters for Portals",
                "sec_num": "3.1.3."
            },
            {
                "text": "\u2022 The function for finding the next-page URL for the archive where applicable",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Possible Parameters for Portals",
                "sec_num": "3.1.3."
            },
            {
                "text": "\u2022 The function to get the article URLs from the archive pages",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Possible Parameters for Portals",
                "sec_num": "3.1.3."
            },
            {
                "text": "\u2022 Boolean variables to answer the following questions: is the archive paginated, infinite scrolling, or datebased?",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Possible Parameters for Portals",
                "sec_num": "3.1.3."
            },
            {
                "text": "One can distinguish between crawl-based (applicable for the current crawl), portal-based (which applies to the crawled portal regardless of crawl settings), and portalspecific configurations. Our method follows the latter direction for crawling. For problems not addressed by Indig et al. (2019) we present our solutions in Section 4.",
                "cite_spans": [
                    {
                        "start": 276,
                        "end": 295,
                        "text": "Indig et al. (2019)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Possible Parameters for Portals",
                "sec_num": "3.1.3."
            },
            {
                "text": "There are a lot of pages that present the same type of objects or articles surrounded by the same boilerplate content (i.e. scheme or template) -menu, advertisements, etc. in a portal. In most cases, this boilerplate content can be characterised by not having multiple paragraphs with many lines of text, but containing short texts, as well as many links and pictures (Pomik\u00e1lek, 2011) . The process of boilerplate removal can be broken down into two steps presented in the following sections.",
                "cite_spans": [
                    {
                        "start": 368,
                        "end": 385,
                        "text": "(Pomik\u00e1lek, 2011)",
                        "ref_id": "BIBREF8"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Boilerplate Removal",
                "sec_num": "3.2."
            },
            {
                "text": "By normalisation we mean the reformatting of visual elements into a simpler visual annotation (e.g. the elements of Markdown language or XML tags) to create a common ground for the text of different portals in the resulting corpus. Normalisation is not a trivial task: most tools extract paragraphs as plain text, however, visual formatting elements are essential for humans and may also help the machine reader, therefore these elements should be kept and standardised.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Normalisation",
                "sec_num": "3.2.1."
            },
            {
                "text": "Curated metadata is the cornerstone of proper (web) archiving. It can be regarded as gold standard labels for each document, which can later be utilised for training or benchmarking ML algorithms (i.e. authorship attribution, keyword extraction, topic modelling, etc.). There are automatic tools for extracting metadata from the crawled web pages such as the Web Curator Tool 5 or Apache Tika 6 . These tools extract standards compliant descriptive metadata automatically from the crawled web pages, but they are very complex and it is difficult to understand and improve their method for the specific portals. Moreover, they are plagued with the same problems as other boilerplate removal tools (see Section 3.2.3.): their heuristics and output formats are wired in by design and it is very hard to change these without major conflicts. When the these programs yield deficient output for the targeted portals -for example due to the lack of knowledge about the typographical rules of the language, or when the output is missing some important variables, -it is inevitable to implement a custom metadata extractor methodology. We decided to use this method to allow future modifications, to be able to compare results with the presented generic tools (see Section 3.2.3.), and also to demonstrate how easily our method can be implemented. Our findings will be described in Section 4..",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Metadata extraction",
                "sec_num": "3.2.2."
            },
            {
                "text": "As web page layouts, coding styles, and HTML standards differ throughout the portals and were used differently over the years, the boilerplate removal task is mostly solved by clever heuristics, which makes it hard for the users to create general measurements and comparisons between them. It is also hard to set their parameters, fix, extend or modify their functionality. Some tools are designed to remove boilerplate from a single page, while others use multiple pages of the same layout to delete recurring elements (Endr\u00e9dy and Nov\u00e1k, 2013) . In this paper, we could not survey all the available methods, therefore we are comparing JusText (Pomik\u00e1lek, 2011) , a tool created directly for NLP-centric web crawling and Newspaper3k (Ou-Yang, 2013), created especially for news portal crawling. Both modules are still popular and widely-used because of their simplicity and effectiveness. They both remove formatting and yield plain text paragraphs, but the latter tool supports extracting metadata from web pages and has other features to simplify the crawling process for beginners. We followed the route marked by Indig et al. (2019) and created our own handcrafted boilerplate removal rules. At first we found ourselves in a dilemma about choosing between regular expressions and HTML parsers. Regular expressions are simple, and it is also easier to train machines to create them, while HTML parsers are easier to create, read and maintain for humans, but are harder to automate. As some of the portals recently added to the corpus have very complex layouts, it is not feasible to further extend the number of portals using regular expressions. For example, it may be impossible or become very unpractical to encode various attributes and their combinations (which might be in arbitrary order due to the structure of HTML). We compared the aforementioned methods on our gold standard data set 7 . This measurement is presented in Section 5., followed by other details of our method.",
                "cite_spans": [
                    {
                        "start": 520,
                        "end": 545,
                        "text": "(Endr\u00e9dy and Nov\u00e1k, 2013)",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 645,
                        "end": 662,
                        "text": "(Pomik\u00e1lek, 2011)",
                        "ref_id": "BIBREF8"
                    },
                    {
                        "start": 1118,
                        "end": 1137,
                        "text": "Indig et al. (2019)",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 1899,
                        "end": 1900,
                        "text": "7",
                        "ref_id": "BIBREF5"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Existing Tools and Techniques",
                "sec_num": "3.2.3."
            },
            {
                "text": "Sometimes the exact same content -available on several domains -can be stored in the web archive multiple times, but, of course, we need one intstance only. There are great tools for deduplication (like Onion (Pomik\u00e1lek, 2011) ), but their waste of valuable resources, such as disk space and network bandwidth is not ideal. When using targeted crawling, such as Indig et al. 2019, we can select only those distinct URLs which are needed and so bypass the process of deduplication.",
                "cite_spans": [
                    {
                        "start": 209,
                        "end": 226,
                        "text": "(Pomik\u00e1lek, 2011)",
                        "ref_id": "BIBREF8"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Deduplication and NLP",
                "sec_num": "3.3."
            },
            {
                "text": "The main problem with deduplication -besides wasting resources -is that some parts of a document or the whole document may become missing because it had been recognised and deleted as a duplicate. This undermines the completeness of the crawling which is the the first priority for some user groups (e.g. humanists and sociologists). The publicly available corpora that were created for NLP purposes have further disabilities: their sentences are scrambled to avoid the infringement of copyright laws. This makes the analysis of full documents -an emergent trend -impossible. The role -and legal privilege -of national libraries is to preserve documents in its entirety, even for born-digital materials. This role can be fulfilled with our method, in contrast to the traditional ones. Different levels of NLP annotation can optionally be applied before or between the deduplication with the plethora of available tools. Until recently, texts have been stored only after this step in some format, however, the increasing performance of NLP tools makes it advisable to store crawled content also in raw format (e.g. WARC files) to be able to correct errors found in the processing pipeline. This is mainly important to humanists, sociologists and other scholars outside NLP where the specific text is the subject of analysis, in contrast to NLP, where only the amount of text matters.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Deduplication and NLP",
                "sec_num": "3.3."
            },
            {
                "text": "The process of creating the output from the HTML files can be split into four steps for easier maintainability:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "\u2022 Simplification of HTML by finding the tightest bounding HTML tag of the whole text content and decomposing unneeded subtrees 8",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "\u2022 Extraction of paragraphs and metadata from the HTML tree keeping only specific -intended -formatting",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "\u2022 Rewriting elements to a unified format by standardising site-specific formatting",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "\u2022 Writing the output file according to the expected format. In this step, the fields get their final place and canonical names",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "The first three steps contain well-defined portal specific instructions, while the fourth is only dependent on the output format, which -as it is totally separated from the otherscan comply with the actual purpose and front end in the future. Some user groups have special requirements, such as full documents and metadata, while others only require the raw text. Nonetheless, both requirements can be achieved at the same time.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "In the field of NLP, three main use cases exist. To search patterns in large corpora, the classic vertical format used primarily by the Sketch Engine (Kilgarriff et al., 2014) is recommended. If the aim is to process the corpus with a wide variety of standard NLP tools, the CoNLL-U format 9 is adequate. If the goal is to put documents to a full text search engine or into a language model, it is necessary to comply with the input expectation of such software, which is usually raw text.",
                "cite_spans": [
                    {
                        "start": 150,
                        "end": 175,
                        "text": "(Kilgarriff et al., 2014)",
                        "ref_id": "BIBREF16"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "In the field of digital humanities, -especially in philology,the XML document markup language and the Text Encoding Initiative (TEI) recommendation have become dominant over the decades (Schreibman et al., 2008) . TEI makes the versioning and annotation of the enriched articles possible in an easy and reliable way, and it is also capable of storing metadata and the body of the document structurally in one file. This format satisfies NLP users as well, while opening the resulting corpus for other audiences including librarians, humanists and sociologists. TEI also allows the verification of the authenticity of the source text by the metadata and increases the reproducibility of research which has an increasing importance in the 'distant reading' paradigm (Da, 2019) . Text can be converted to a simpler form corresponding to the actual use case, while keeping the master copy untouched, in a similarly to how it is done with images by resizing and cropping them on demand dynamically.",
                "cite_spans": [
                    {
                        "start": 186,
                        "end": 211,
                        "text": "(Schreibman et al., 2008)",
                        "ref_id": "BIBREF9"
                    },
                    {
                        "start": 764,
                        "end": 774,
                        "text": "(Da, 2019)",
                        "ref_id": "BIBREF0"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Final Format and Front End",
                "sec_num": "3.4."
            },
            {
                "text": "We examined several Hungarian news portals and increased the number of examined portals to six, compared to the three portals examined by Indig et al. (2019) in order to test how the presented method can be applied to portals of different structures. First, we selected mainstream Hungarian news portals, because these contain a vast number of articles. As a secondary priority, we included portals that are special from the perspective of used web technology and architecture. We wanted to reach a milestone, where adding new portals and maintaining the existing ones is a routine task that can be handled by as little manpower as possible. In this section, we describe the main highlights of our crawling method compared to (Indig et al., 2019 ) (for further comparisons see Section 3.)",
                "cite_spans": [
                    {
                        "start": 138,
                        "end": 157,
                        "text": "Indig et al. (2019)",
                        "ref_id": "BIBREF3"
                    },
                    {
                        "start": 726,
                        "end": 745,
                        "text": "(Indig et al., 2019",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method",
                "sec_num": "4."
            },
            {
                "text": "We decided to change the regular expressions used in Corpusbuilder (Indig et al., 2019) for Python functions, which use an HTML parser to handle the input as an HTML tree. Using HTML trees enabled us both to simplify many regular expression patterns and to support many different layouts. With this change, the accuracy of extracting article URLs from the page archives has dramatically increased, as we found that on some portals different columns may be hosted on different domains, or -while using the same site template -they may not match the expressions written for extracting URLs. This can be recognised by tree searching expressions more easily than with regular expressions. This, of course, sacrifices speed for clarity and precision, but saves us from HTML fragments slipping through regular expressions.",
                "cite_spans": [
                    {
                        "start": 67,
                        "end": 87,
                        "text": "(Indig et al., 2019)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "HTML Parsers vs. Regular Expressions",
                "sec_num": "4.1."
            },
            {
                "text": "The date-based pagination handling logic (Indig et al., 2019) was separated from other pagination methods, as it allows sorting and can be used to filter crawling by specific date intervals, since we found that date-based pagination can be and is combined freely with the other methods. We also introduced support for open (date) intervals. Our other significant change was in handling infinite scrolling 10 and active archives 11 together in an easy-tounderstand form by extracting the page URLs before determining the URL of the next archive page. We have broken down the possible patterns of finding the next archive page URL to the following cases:",
                "cite_spans": [
                    {
                        "start": 41,
                        "end": 61,
                        "text": "(Indig et al., 2019)",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Refined Archive Crawler",
                "sec_num": "4.2."
            },
            {
                "text": "\u2022 There is no pagination at all",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Refined Archive Crawler",
                "sec_num": "4.2."
            },
            {
                "text": "\u2022 There is a next page link which we need to use",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Refined Archive Crawler",
                "sec_num": "4.2."
            },
            {
                "text": "\u2022 There is infinite scrolling: we use the page number from the base value to \"infinity\" where no more article URLs are detected",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Refined Archive Crawler",
                "sec_num": "4.2."
            },
            {
                "text": "\u2022 There is page numbering: we use the page number from the base value to a portal-specific maximum",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Refined Archive Crawler",
                "sec_num": "4.2."
            },
            {
                "text": "\u2022 There is page numbering, but we expect the archive to expand during crawling (can be identified by finding known article URLs during crawling)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Refined Archive Crawler",
                "sec_num": "4.2."
            },
            {
                "text": "By using these features, all examined portals could be handled, therefore we narrowed down our experiments to six portals that showcase all of the described features, and allows them to be thoroughly tested.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Refined Archive Crawler",
                "sec_num": "4.2."
            },
            {
                "text": "Metadata can be extracted from multiple sources from an article page. We identified and handled the following:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Advanced Metadata Extraction",
                "sec_num": "4.3."
            },
            {
                "text": "\u2022 Date, title and column are frequently encoded in the URLs",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Advanced Metadata Extraction",
                "sec_num": "4.3."
            },
            {
                "text": "\u2022 HTML meta tag properties which can be encoded according to various conventions (like Dublin Core, Schema.org, etc.) that are mainly included for Search Engine Optimization (SEO) purposes",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Advanced Metadata Extraction",
                "sec_num": "4.3."
            },
            {
                "text": "\u2022 The increasingly popular JSON-LD, storing properties that were previously stored as meta tags, but in a more structured form",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Advanced Metadata Extraction",
                "sec_num": "4.3."
            },
            {
                "text": "\u2022 From the content of the HTML itself, where it is included to be displayed for the user There are several portals that use more than one of the above sources of metadata. We also found examples where different sources yielded contradicting results or missing values, these are probably due to bugs in the websites' engines. Older articles tend to have more of these errors as they were probably converted from a previous layout and the conversion introduced such errors 12 . Some portals partially or fully generate metadata dynamically by using JavaScript and non-standard data-sources. This practically makes it impossible to extract such metadata with traditional tools and forces us to use a portal-specific solution.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Advanced Metadata Extraction",
                "sec_num": "4.3."
            },
            {
                "text": "To handle millions of pages without reading them -through \"distant reading\" -, we invented utilities to examine, analyse and normalise the tags and the scheme used by a portal, and then freely convert it to the new and customisable output format. We started with cutting the HTML to the relevant part, as mentioned in Section 3.4.. The first utility function helps to filter out tags that do not contain any text. Next, we introduced placeholders to simplify some elements (e.g. links). The second function aids in simplifying the tags by manually selecting groups that belong to the same class (e.g. formatting, embedded content, etc.), but are specialised to the portal's scheme. This method is quite effective even without portal-specific parameters. Table 1 shows how the number of tags (from one of the examined domains) is reduced after using these tools allowing further fine-grained modifications in an iterative manner.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 754,
                        "end": 761,
                        "text": "Table 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Converting HTML to the Output Format",
                "sec_num": "4.4."
            },
            {
                "text": "No. of tags % all tags 33,466 100 text containing tags 18,517 55 after simplify tags 359 10 relevant tags 267 7 Table 1 : Illustration of how the number of tags to be analysed manually decreases in magnitude.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 112,
                        "end": 119,
                        "text": "Table 1",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Converting HTML to the Output Format",
                "sec_num": "4.4."
            },
            {
                "text": "Possible layouts for all URLs of a domain were described with the help of a tree-representation: the subtrees of the contents' tightest bounding HTML tag for all pages were merged, counting the frequencies of each element and the cumulative length of their immediate text. It was also marked if a specific tag had no child elements in the tree. The resulting frequency distribution allows efficient examination and handling of subtrees for all URLs at once. In order to be able to make decisions concerning the remaining tags, we built a tag dictionary. To each tag (or simplified tag), we assigned the average length of the contained text, the average number of descendants, and the average length of the immediate text supplemented with a sample of occurrences (URLs). This dictionary was augmented with the operation to perform at each occurrence of that specific tag. As we formalised the operators, their execution was made by the code automatically generated from the dictionary. These steps can be iterated to gain more insight on the portal's scheme and finally arrive to the desired form.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "The Tree Representation",
                "sec_num": "4.4.1."
            },
            {
                "text": "When standardising and rewriting elements, we found the following operators useful:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "\u2022 decomposing (deleting the tag with its contents, e.g. advertisements, boilerplate)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "\u2022 unwrapping (deleting the tag, keeping its contents, e.g. text anchors)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "\u2022 unwrapping all descendants (simplifying a block)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "\u2022 rewriting tags context-free",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "\u2022 splitting tags to super-subordinate pairs (e.g. when the content and formatting properties are in the same tag)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "\u2022 rewriting tags context-specific (special blocks)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "These operators can be applied sequentially in the proper order for every URL. We narrowed down the various layouts (e.g. left, right, top block) into a few, portal independent types of blocks that we intended to keep. The contextspecific rules mark the root tag for each block we found, so their subtrees can be handled by independent dictionaries in the same way. The analysis of the visual layout of the examined portals shows that there are no blocks embedded into other blocks. This property allows us to rely on the described two-level transformation with a low number of distinct tag dictionaries modified by the defined operators. To conclude, normalising the tags and then rewriting them to the final schema are independent steps which can be achieved with successive approximation in an iterative manner. This allows us fine-grained control to change design decisions or customise the output (TEI XML in our case) easily at all times.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Rewriting Rules and Transformation Methods",
                "sec_num": "4.4.2."
            },
            {
                "text": "We ran our crawling on a low-end desktop machine (Intel i3, 4 GB RAM) for 30 days on a 100 MB/s connection (with rate-limiting to avoid the hammering of the remote servers) using circa 100 GB of disk space to demonstrate the effectivity of the method presentes here. It is not possible to compare this method's crawling performance to other general crawler engines mentioned earlier, as the workflow and methodology differ significantly (see Section 3.1.). It is possible, however, to compare the crawling accuracy to the most widely used archiving practice: the Internet Archive (see Section 5.2.). It is also possible to compare our sitespecific rule-based boilerplate removal and metadata extractor functions to the mainstream crawling methods (see Section 3.2.3.). The goal of the compared tools and their design differs significantly so the way how to make an objective comparison was not at all obvious. When comparing our method with the aforementioned tools, we strived to highlight performance differences due to design, while separating them from the strengths and weaknesses stemming from the methods themselves.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Evaluation",
                "sec_num": "5."
            },
            {
                "text": "We extracted a total of 2,227,180 articles from six Hungarian news portals, this signifies 984 million words (without tokenisation) of extracted text without metadata from November 1998 until September 2019. We visualised the annual distribution of articles to see the estimated growth in the number of articles and the expected number of articles per year (see Figure 2) . The figure shows a clearly growing tendency in the number of articles published on the crawled portals during the last twenty years -except 2019, which does not qualify as a full year at the time of measurement. In the case of the six portals, this means that more than 200 articles have been published on average every day in the recent years. These numbers tells us that by adding new portals the quantity of the crawled articles and the volume of the corpus can be increased quickly and easily with low human resource investment and a lightweight technical infrastructure.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 362,
                        "end": 371,
                        "text": "Figure 2)",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "The data set",
                "sec_num": "5.1."
            },
            {
                "text": "1 9 9 7 1 9 9 9 2 0 0 1 2 0 0 3 2 0 0 5 2 0 0 7 2 0 0 9 2 0 1 1 2 0 1 3 2 0 1 5 2 0 1 7 2 0 1 9 No. of Articles Figure 2 : The annual distribution of 2,227,180 articles from six portals from November 1998 to September 2019. The number of articles per year is increasing. The decrease at 2019 is due to the fact that it is not a full year at the time of measurement.",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 112,
                        "end": 120,
                        "text": "Figure 2",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "The data set",
                "sec_num": "5.1."
            },
            {
                "text": "In Table 2 , we can see the performance of the boilerplate removal tools in different scenarios. We examined Jus-Text and Newspaper3k on the full HTML code, the article body and constrained to the original and the cleaned up paragraphs. We wanted to check whether an educated initial guess (on the text's location) helps these programs or not. As the former package does not extract metadata separately, we present numbers with metadata and provide the number of words without metadata in brackets. The numbers have some small differences that suggests that a more detailed evaluation of the content is needed. We also compared the actual values of the extracted metadata (author, publication date, title) in terms of precision and recall for Newspaper3k (see Table 3 ). Our educated initial guess does not help metadata extraction, but for the text extraction it has a potential because it rules out unwanted content in one step. It is clear from these that our method is superior to the compared ones, however, a content-based comparison of the extracted paragraphs is needed in order to be able to evaluate the mentioned methods objectively. We argue that if full articles are chosen, the precision provided by our method is needed to ensure that the right amount and quality of texts can be extracted with the compared methods. ",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 3,
                        "end": 10,
                        "text": "Table 2",
                        "ref_id": "TABREF2"
                    },
                    {
                        "start": 760,
                        "end": 767,
                        "text": "Table 3",
                        "ref_id": "TABREF3"
                    }
                ],
                "eq_spans": [],
                "section": "The data set",
                "sec_num": "5.1."
            },
            {
                "text": "We compared our results of the six crawled news portals to the Internet Archive as the \"standard\" source of web archiving. We evaluated whether the same set of URLs could be acquired using the Internet Archive, and also compared the number of crawled articles by portals with data downloaded from The Wayback Machine. In the following step, based on the mime type attribute, we removed all URLs from the Internet Archive data sets that represent content other than articles (e.g. images, scripts, etc.). Using the status code variable, we omitted all URLs that were not successfully downloaded for some reason (e.g. 404 errors and redirections). From our crawl we selected the timestamp of the last article downloaded for each domain, and removed all URLs from the Internet Archive data that were crawled after that date. At this point, we still had hundreds of thousands of URLs in the Internet Archive data sets that represented e.g. certain taxonomy pages (date, category, author, search, etc.) or any kind of content other than single articles. Thus, we introduced a domain-level cleaning function for each crawled website, in order to remove all URLs representing content other than articles. This proved to be a difficult, time-consuming, iterative task, as in case of some websites, the URL structure changed multiple times over the years, making it nearly impossible to retrospectively identify URLs that certainly lead to articles. This is one important aspect why our method is much easier to use (even retrospectively), when the goal is to produce a clean corpus, without duplicated content. In the case of several websites, the URL structure was not logically constructed (e.g. tag archives have the same URL structure as articles; randomly generated version numbers appear at the end of some of the URLs, but not all of them; etc.), therefore in some cases, we had to restrict the comparison to certain columns of the portal, as it was very difficult to clean the data sets in a more generalised way. Our next step was to normalise all URLs in both crawls. We removed http, https, www from the beginning, and port numbers (e.g. \":80\") and parameters from the end of the URL strings. Using these normalised URLs, we created two dictionaries to store the URLs themselves and their slugs -the last part of the URL (after the last /) -for each portal. For some portals the URLs could not be used for a valid comparison, because the URL structure has changed over time, but not the slug, therefore -in these cases -we used the slug for our comparison.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Crawling Compared to Archive.org",
                "sec_num": "5.2."
            },
            {
                "text": "With the steps described above, we reduced the number of Archive.org URLs from 8.9 million to only 1.2 million for the six crawled portals. After removing entries with wrong status codes 75.4%, after mime-type-based cleaning 53.7% of the URLs remained. While only 0.7% of URLs were removed in the date-based cleaning phase, after running website-specific cleaning functions and compiling the final list of URLs, just 13.5% of the initial number of URLs remained. We found that 846,343 articles are present both in our crawl and in the Internet Archive's data, while 1,082,484 articles are only present in the ELTE.DH corpus. A further 315,649 articles are only found in Archive.org's data. More work is needed in order to eliminate all possible bad URLs, however, it is safe to say that by using our crawler it is easier to achieve the same results than finding and downloading all relevant content from Archive.org.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Crawling Compared to Archive.org",
                "sec_num": "5.2."
            },
            {
                "text": "We have demonstrated that by using a low-end machinewhich has similar computational power as our smartphones, the storage capacity of our pendrives nowadays -and minimal manpower it is possible to create a gold-standard quality gigaword corpus with metadata which suits many audiences at the same time 13 . As the presented work was only a pilot study to design and stabilise the workflow on many candidate pages, we plan to apply this methodology on several more websites, and start serving requests on sitespecific crawling to provide data for research in multiple disciplines in a future version of this corpus. In conjunction with the previously outlined plans, we intend to support national libraries with our research as they are responsible of keeping the data of our present for the future researchers who can thus provide objective and balanced research. One obvious step in this direction is to conduct research on how to keep the authenticity of web archives and how to eliminate the risks of tampering, retroactive modification and deletion of content which undermine scholarly credibility. We plan to utilise digital fingerprinting, signatures and blockchain technology on downloaded documents in order to keep them safe, while making them available for the widest possible audience. 13 The software is published under the GNU Lesser General Public License v3.0 at https://github. com/ELTE-DH/WebArticleCurator and https: //doi.org/10.5281/zenodo.3755323",
                "cite_spans": [
                    {
                        "start": 1297,
                        "end": 1299,
                        "text": "13",
                        "ref_id": null
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion and Future Work",
                "sec_num": "6."
            },
            {
                "text": "Winters (2017, 240) deals with the problem of website dates in detail.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "http://netpreserve.org/resources/IIPC_ project-SO_TR_14873__E__2012-10-02_DRAFT. pdf",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "The P\u00e1zm\u00e1ny corpus was the first Hungarian corpus which separated edited text (news articles) from unedited text (comments).4 https://github.com/internetarchive/ heritrix3",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "https://webcuratortool.readthedocs.io/ 6 http://tika.apache.org/",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "Some elements were kept or thrown away by design decision that may not match with the compared tools or future use cases. However, we support the change of these decisions by the user.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "There are three classes of decomposing rules: a) general rules used for every portal, b) \"must-have\" portal-specific rules, c) rules which follow certain design decisions about the data to be extracted.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "https://universaldependencies.org/ format.html",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "A technique used to dynamically add new content to the page when the user scrolls down.11 If new elements are added to the archive during crawling, the list of articles will be divided to pages in a way that their content URLs will appear on different pages than as expected. This makes it impossible to handle archive pages' URLs as permalinks.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "This can be solved by crawling articles as soon as possible after their publication.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "The computational case against computational literary studies",
                "authors": [
                    {
                        "first": "N",
                        "middle": [
                            "Z"
                        ],
                        "last": "Da",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "Critical inquiry",
                "volume": "45",
                "issue": "3",
                "pages": "601--639",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Da, N. Z. (2019). The computational case against compu- tational literary studies. Critical inquiry, 45(3):601-639.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "More effective boilerplate removal-the goldminer algorithm",
                "authors": [
                    {
                        "first": "I",
                        "middle": [],
                        "last": "Endr\u00e9dy",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Nov\u00e1k",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "Polibits",
                "volume": "1",
                "issue": "48",
                "pages": "79--83",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Endr\u00e9dy, I. and Nov\u00e1k, A. (2013). More effective boilerplate removal-the goldminer algorithm. Polibits, 1(48):79-83.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "The Digital Humanities: A Primer for Students and Scholars",
                "authors": [
                    {
                        "first": "E",
                        "middle": [],
                        "last": "Gardiner",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [
                            "G"
                        ],
                        "last": "Musto",
                        "suffix": ""
                    }
                ],
                "year": 2015,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Gardiner, E. and Musto, R. G. (2015). The Digital Human- ities: A Primer for Students and Scholars. Cambridge University Press.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Crawling in reverse -lightweight targeted crawling of news portals",
                "authors": [
                    {
                        "first": "B",
                        "middle": [],
                        "last": "Indig",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "K\u00e1konyi",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Nov\u00e1k",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "Proceedings of the 9th Language & Technology Conference: Human Language Technologies as a Challenge for Computer Science and Linguistics",
                "volume": "",
                "issue": "",
                "pages": "81--87",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Indig, B., K\u00e1konyi, T., and Nov\u00e1k, A. (2019). Crawling in reverse -lightweight targeted crawling of news portals. In Marek Kubis, editor, Proceedings of the 9th Language & Technology Conference: Human Language Technolo- gies as a Challenge for Computer Science and Linguis- tics, pages 81-87, Pozna\u0144, Poland, may. Wydawnictwo Nauka i Innowacje.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "Asynchronous Pipeline for Processing Huge Corpora on Medium to Low Resource Infrastructures",
                "authors": [
                    {
                        "first": "P",
                        "middle": [
                            "J"
                        ],
                        "last": "Ortiz Su\u00e1rez",
                        "suffix": ""
                    },
                    {
                        "first": "B",
                        "middle": [],
                        "last": "Sagot",
                        "suffix": ""
                    },
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Romary",
                        "suffix": ""
                    }
                ],
                "year": 2019,
                "venue": "7th Workshop on the Challenges in the Management of Large Corpora",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Ortiz Su\u00e1rez, P. J., Sagot, B., and Romary, L. (2019). Asynchronous Pipeline for Processing Huge Corpora on Medium to Low Resource Infrastructures. In Pi- otr Ba\u0144ski, et al., editors, 7th Workshop on the Chal- lenges in the Management of Large Corpora (CMLC-",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "Leibniz-Institut f\u00fcr Deutsche Sprache",
                "authors": [],
                "year": null,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": ", Cardiff, United Kingdom, July. Leibniz-Institut f\u00fcr Deutsche Sprache.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Newspaper3k: Article scraping and curation",
                "authors": [
                    {
                        "first": "L",
                        "middle": [],
                        "last": "Ou-Yang",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Ou-Yang, L. (2013). Newspaper3k: Article scraping and curation. https://github.com/codelucas/ newspaper.",
                "links": null
            },
            "BIBREF7": {
                "ref_id": "b7",
                "title": "Counting the uncountable: statistics for web archives",
                "authors": [
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Oury",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Poll",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "Performance Measurement and Metrics",
                "volume": "14",
                "issue": "2",
                "pages": "132--141",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Oury, C. and Poll, R. (2013). Counting the uncountable: statistics for web archives. Performance Measurement and Metrics, 14(2):132-141.",
                "links": null
            },
            "BIBREF8": {
                "ref_id": "b8",
                "title": "Removing boilerplate and duplicate content from web corpora",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Pomik\u00e1lek",
                        "suffix": ""
                    }
                ],
                "year": 2011,
                "venue": "Masaryk university, Faculty of informatics",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Pomik\u00e1lek, J. (2011). Removing boilerplate and duplicate content from web corpora. Ph.D. thesis, Masaryk uni- versity, Faculty of informatics, Brno, Czech Republic.",
                "links": null
            },
            "BIBREF9": {
                "ref_id": "b9",
                "title": "A companion to digital humanities",
                "authors": [
                    {
                        "first": "S",
                        "middle": [],
                        "last": "Schreibman",
                        "suffix": ""
                    },
                    {
                        "first": "R",
                        "middle": [],
                        "last": "Siemens",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Unsworth",
                        "suffix": ""
                    }
                ],
                "year": 2008,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Schreibman, S., Siemens, R., and Unsworth, J. (2008). A companion to digital humanities. John Wiley & Sons.",
                "links": null
            },
            "BIBREF10": {
                "ref_id": "b10",
                "title": "Methods and approaches to using web archives in computational communication research",
                "authors": [
                    {
                        "first": "M",
                        "middle": [
                            "S"
                        ],
                        "last": "Weber",
                        "suffix": ""
                    }
                ],
                "year": 2018,
                "venue": "Communication Methods and Measures",
                "volume": "12",
                "issue": "2-3",
                "pages": "200--215",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Weber, M. S. (2018). Methods and approaches to us- ing web archives in computational communication re- search. Communication Methods and Measures, 12(2- 3):200-215.",
                "links": null
            },
            "BIBREF11": {
                "ref_id": "b11",
                "title": "Coda: Web archives for humanities research -some reflections",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Winters",
                        "suffix": ""
                    }
                ],
                "year": 2017,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "238--248",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Winters, J., (2017). Coda: Web archives for humanities re- search -some reflections, pages 238-248. UCL Press.",
                "links": null
            },
            "BIBREF13": {
                "ref_id": "b13",
                "title": "A p\u00e1zm\u00e1ny korpusz",
                "authors": [
                    {
                        "first": "I",
                        "middle": [],
                        "last": "Endr\u00e9dy",
                        "suffix": ""
                    },
                    {
                        "first": "G",
                        "middle": [],
                        "last": "Pr\u00f3sz\u00e9ky",
                        "suffix": ""
                    }
                ],
                "year": 2016,
                "venue": "Nyelvtudom\u00e1nyi K\u00f6zlem\u00e9nyek",
                "volume": "112",
                "issue": "",
                "pages": "191--205",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Endr\u00e9dy, I. and Pr\u00f3sz\u00e9ky, G. (2016). A p\u00e1zm\u00e1ny korpusz. Nyelvtudom\u00e1nyi K\u00f6zlem\u00e9nyek, 112:191-205.",
                "links": null
            },
            "BIBREF14": {
                "ref_id": "b14",
                "title": "Creating open language resources for Hungarian",
                "authors": [
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Hal\u00e1csy",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Kornai",
                        "suffix": ""
                    },
                    {
                        "first": "L",
                        "middle": [],
                        "last": "N\u00e9meth",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Rung",
                        "suffix": ""
                    },
                    {
                        "first": "I",
                        "middle": [],
                        "last": "Szakad\u00e1t",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Tr\u00f3n",
                        "suffix": ""
                    }
                ],
                "year": 2004,
                "venue": "Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC'04)",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Hal\u00e1csy, P., Kornai, A., N\u00e9meth, L., Rung, A., Szakad\u00e1t, I., and Tr\u00f3n, V. (2004). Creating open language resources for Hungarian. In Proceedings of the Fourth Interna- tional Conference on Language Resources and Evalua- tion (LREC'04), Lisbon, Portugal, May. European Lan- guage Resources Association (ELRA).",
                "links": null
            },
            "BIBREF15": {
                "ref_id": "b15",
                "title": "The tenten corpus family",
                "authors": [
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Jakub\u00ed\u010dek",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Kilgarriff",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Kov\u00e1\u0159",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Rychl\u1ef3",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Suchomel",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "7th International Corpus Linguistics Conference CL",
                "volume": "",
                "issue": "",
                "pages": "125--127",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Jakub\u00ed\u010dek, M., Kilgarriff, A., Kov\u00e1\u0159, V., Rychl\u1ef3, P., and Suchomel, V. (2013). The tenten corpus family. In 7th International Corpus Linguistics Conference CL, pages 125-127.",
                "links": null
            },
            "BIBREF16": {
                "ref_id": "b16",
                "title": "The sketch engine: ten years on. Lexicography",
                "authors": [
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Kilgarriff",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Baisa",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Bu\u0161ta",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Jakub\u00ed\u010dek",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Kov\u00e1\u0159",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Michelfeit",
                        "suffix": ""
                    },
                    {
                        "first": "P",
                        "middle": [],
                        "last": "Rychl\u00fd",
                        "suffix": ""
                    },
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Suchomel",
                        "suffix": ""
                    }
                ],
                "year": 2014,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "7--36",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Kilgarriff, A., Baisa, V., Bu\u0161ta, J., Jakub\u00ed\u010dek, M., Kov\u00e1\u0159, V., Michelfeit, J., Rychl\u00fd, P., and Suchomel, V. (2014). The sketch engine: ten years on. Lexicography, pages 7-36.",
                "links": null
            },
            "BIBREF17": {
                "ref_id": "b17",
                "title": "Web Crawling and Data Mining with Apache Nutch",
                "authors": [
                    {
                        "first": "Z",
                        "middle": [],
                        "last": "Laliwala",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Shaikh",
                        "suffix": ""
                    }
                ],
                "year": 2013,
                "venue": "",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Laliwala, Z. and Shaikh, A. (2013). Web Crawling and Data Mining with Apache Nutch. Packt Publishing.",
                "links": null
            },
            "BIBREF18": {
                "ref_id": "b18",
                "title": "The Hungarian Gigaword corpus",
                "authors": [
                    {
                        "first": "C",
                        "middle": [],
                        "last": "Oravecz",
                        "suffix": ""
                    },
                    {
                        "first": "T",
                        "middle": [],
                        "last": "V\u00e1radi",
                        "suffix": ""
                    },
                    {
                        "first": "B",
                        "middle": [],
                        "last": "Sass",
                        "suffix": ""
                    }
                ],
                "year": 2014,
                "venue": "Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)",
                "volume": "",
                "issue": "",
                "pages": "1719--1723",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Oravecz, C., V\u00e1radi, T., and Sass, B. (2014). The Hungar- ian Gigaword corpus. In Proceedings of the Ninth Inter- national Conference on Language Resources and Eval- uation (LREC'14), pages 1719-1723, Reykjavik, Ice- land, May. European Language Resources Association (ELRA).",
                "links": null
            },
            "BIBREF19": {
                "ref_id": "b19",
                "title": "Efficient web crawling for large text corpora",
                "authors": [
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Suchomel",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Pomik\u00e1lek",
                        "suffix": ""
                    }
                ],
                "year": 2012,
                "venue": "Proceedings of the seventh Web as Corpus Workshop (WAC7)",
                "volume": "",
                "issue": "",
                "pages": "39--43",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Suchomel, V. and Pomik\u00e1lek, J. (2012). Efficient web crawling for large text corpora. In Adam Kilgarriff et al., editors, Proceedings of the seventh Web as Corpus Work- shop (WAC7), pages 39-43, Lyon.",
                "links": null
            },
            "BIBREF20": {
                "ref_id": "b20",
                "title": "KorKorpusz: k\u00e9zzel annot\u00e1lt, t\u00f6bbr\u00e9teg\u0171 pilotkorpusz\u00e9p\u00edt\u00e9se",
                "authors": [
                    {
                        "first": "N",
                        "middle": [],
                        "last": "Vad\u00e1sz",
                        "suffix": ""
                    }
                ],
                "year": 2020,
                "venue": "XVI. Magyar Sz\u00e1m\u00edt\u00f3g\u00e9pes Nyelv\u00e9szeti Konferencia (MSZNY 2020)",
                "volume": "",
                "issue": "",
                "pages": "141--154",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Vad\u00e1sz, N. (2020). KorKorpusz: k\u00e9zzel annot\u00e1lt, t\u00f6bbr\u00e9teg\u0171 pilotkorpusz\u00e9p\u00edt\u00e9se. In G\u00e1bor Berend, et al., editors, XVI. Magyar Sz\u00e1m\u00edt\u00f3g\u00e9pes Nyelv\u00e9szeti Konfer- encia (MSZNY 2020), pages 141-154, Szeged. Szegedi Tudom\u00e1nyegyetem, TTIK, Informatikai Int\u00e9zet.",
                "links": null
            },
            "BIBREF21": {
                "ref_id": "b21",
                "title": "Hungarian Dependency Treebank",
                "authors": [
                    {
                        "first": "V",
                        "middle": [],
                        "last": "Vincze",
                        "suffix": ""
                    },
                    {
                        "first": "D",
                        "middle": [],
                        "last": "Szauter",
                        "suffix": ""
                    },
                    {
                        "first": "A",
                        "middle": [],
                        "last": "Alm\u00e1si",
                        "suffix": ""
                    },
                    {
                        "first": "Gy",
                        "middle": [],
                        "last": "M\u00f3ra",
                        "suffix": ""
                    },
                    {
                        "first": "Z",
                        "middle": [],
                        "last": "Alexin",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Csirik",
                        "suffix": ""
                    }
                ],
                "year": 2010,
                "venue": "Proceedings of LREC 2010",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Vincze, V., Szauter, D., Alm\u00e1si, A., M\u00f3ra, Gy., Alexin, Z., and Csirik, J. (2010). Hungarian Dependency Tree- bank. In Proceedings of LREC 2010, Valletta, Malta, May. ELRA.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "text": "The abstract workflow of web corpus creation. Parallelogram-shaped boxes denote the optional phases, the grey background denotes the produced data.",
                "uris": null,
                "num": null,
                "type_str": "figure"
            },
            "TABREF2": {
                "text": "The extracted text from different parts of the HTML with different tools in million words. Newspaper3k and our method is displayed with and without metadata.",
                "content": "<table><tr><td/><td>Full</td><td>Article</td></tr><tr><td/><td>HTML</td><td>Body</td></tr><tr><td>Newspaper3k (precision)</td><td>0.77</td><td>0.69</td></tr><tr><td>Newspaper3k (recall)</td><td>0.52</td><td>0.26</td></tr></table>",
                "type_str": "table",
                "num": null,
                "html": null
            },
            "TABREF3": {
                "text": "",
                "content": "<table><tr><td>: The content-wise comparison of metadata (author,</td></tr><tr><td>title, publication date) extracted by Newspaper3k and our</td></tr><tr><td>method (=1.0).</td></tr></table>",
                "type_str": "table",
                "num": null,
                "html": null
            }
        }
    }
}