File size: 101,596 Bytes
6fa4bc9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 |
{
"paper_id": "2020",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T04:33:23.322568Z"
},
"title": "The ELTE.DH Pilot Corpus -Creating a Handcrafted Gigaword Web Corpus with Metadata",
"authors": [
{
"first": "Bal\u00e1zs",
"middle": [],
"last": "Indig",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
"location": {
"addrLine": "M\u00fazeum krt. 6-8",
"postCode": "H-1088",
"settlement": "Budapest",
"country": "Hungary"
}
},
"email": "[email protected]"
},
{
"first": "\u00c1rp\u00e1d",
"middle": [],
"last": "Knap",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
"location": {
"addrLine": "1/A",
"postCode": "H-1117",
"settlement": "Budapest",
"country": "Hungary"
}
},
"email": "[email protected]"
},
{
"first": "Zs\u00f3fia",
"middle": [],
"last": "S\u00e1rk\u00f6zi-Lindner",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
"location": {
"addrLine": "M\u00fazeum krt. 6-8",
"postCode": "H-1088",
"settlement": "Budapest",
"country": "Hungary"
}
},
"email": ""
},
{
"first": "M\u00e1ria",
"middle": [],
"last": "Tim\u00e1ri",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
"location": {
"addrLine": "M\u00fazeum krt. 6-8",
"postCode": "H-1088",
"settlement": "Budapest",
"country": "Hungary"
}
},
"email": ""
},
{
"first": "G\u00e1bor",
"middle": [],
"last": "Palk\u00f3",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "E\u00f6tv\u00f6s Lor\u00e1nd University",
"location": {
"addrLine": "M\u00fazeum krt. 6-8",
"postCode": "H-1088",
"settlement": "Budapest",
"country": "Hungary"
}
},
"email": "[email protected]"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "In this article, we present the method we used to create a middle-sized corpus using targeted web crawling. Our corpus contains news portal articles along with their metadata, that can be useful for diverse audiences, ranging from digital humanists to NLP users. The method presented in this paper applies rule-based components that allow the curation of the text and the metadata content. The curated data can thereon serve as a reference for various tasks and measurements. We designed our workflow to encourage modification and customisation. Our concept can also be applied to other genres of portals by using the discovered patterns in the architecture of the portals. We found that for a systematic creation or extension of a similar corpus, our method provides superior accuracy and ease of use compared to The Wayback Machine, while requiring minimal manpower and computational resources. Reproducing the corpus is possible if changes are introduced to the text-extraction process. The standard TEI format and Schema.org encoded metadata is used for the output format, but we stress that placing the corpus in a digital repository system is recommended in order to be able to define semantic relations between the segments and to add rich annotation.",
"pdf_parse": {
"paper_id": "2020",
"_pdf_hash": "",
"abstract": [
{
"text": "In this article, we present the method we used to create a middle-sized corpus using targeted web crawling. Our corpus contains news portal articles along with their metadata, that can be useful for diverse audiences, ranging from digital humanists to NLP users. The method presented in this paper applies rule-based components that allow the curation of the text and the metadata content. The curated data can thereon serve as a reference for various tasks and measurements. We designed our workflow to encourage modification and customisation. Our concept can also be applied to other genres of portals by using the discovered patterns in the architecture of the portals. We found that for a systematic creation or extension of a similar corpus, our method provides superior accuracy and ease of use compared to The Wayback Machine, while requiring minimal manpower and computational resources. Reproducing the corpus is possible if changes are introduced to the text-extraction process. The standard TEI format and Schema.org encoded metadata is used for the output format, but we stress that placing the corpus in a digital repository system is recommended in order to be able to define semantic relations between the segments and to add rich annotation.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "In the glossary of the handbook entitled The Digital Humanities, Gardiner and Musto (2015, 250) define web archiving as \"the process of collecting portions of the World Wide Web to ensure the information is preserved in an Archive for future researchers, historians and the public\". It is telling, however, that in the chapter focusing on digital archives as source materials of the present scholarly practices, born-digital archives and web archives are entirely omitted, as the authors solely speak about curated digital collections designed by (digital) archivists for the research community. Web archives are much less organised and curated then digital libraries or databases, and for this reason, are far less usable for (and used by) scholars. If Gardiner and Musto (2015) are right in their choice to emphasise the role of these digital sources in answering present scholarly questions, the fact that web archives do not play a significant role among these sources is a substantial problem for the digital humanities. There are several reasons why web archives are under-represented in the scholarly use of digital sources. The main reason is the lack of high-quality metadata, as source materials must have -among othersa publication date and its authors identified by the archival institution, otherwise, the reference to the material (be it paper-based or born-digital) is questionable 1 . The second reason is the uniqueness and authenticity of the records.",
"cite_spans": [
{
"start": 65,
"end": 95,
"text": "Gardiner and Musto (2015, 250)",
"ref_id": null
},
{
"start": 754,
"end": 779,
"text": "Gardiner and Musto (2015)",
"ref_id": "BIBREF2"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "Web archives usually contain many nearly identical versions of the \"same\" resource. This problem is exacerbated by the nearly inseparable dirt (recurring boilerplate text) among relevant content. The drawbacks arising from the unstructured nature of a web archive hinder its integration into the network of digital cultural heritage (DCH). As suggested in (Weber, 2018) , the limitations of web archives can be described along two main dimensions: accuracy and completeness. It is very difficult to tell if an archive actually captures all the content on the web accurately related to a specific topic. Our method, by using websites' own archives, creates \"complete snapshots\" of their real content from time to time, which provides real populational data for the portals included in the project. This also means that the ELTE.DH corpus contains all documents from the selected portals' archives which were publicly available at crawling time. Beyond creating a corpus for NLP applications, our work focuses on providing solutions to the aforementioned issues by developing a trusted digital repository complying with Linked Open Data technology. Our goal with this repository is to meet the essential demands of NLP, DCH and other disciplines uniformly.",
"cite_spans": [
{
"start": 356,
"end": 369,
"text": "(Weber, 2018)",
"ref_id": "BIBREF10"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1."
},
{
"text": "When it comes to crawling, web archiving or corpus creation, there are a number of options. The ISO/TR 14873:2013 standard describes the details of such workflows, however, distinct disciplines have come up with their own solutions ignoring this standard or only partially adhering to it. Holding on to the terminologies of the standard, we have conducted selective web archiving that is enriched with more and better metadata compared to general crawling. We argue that our method has a smaller footprint while remaining easy to manage. This makes the whole workflow sustainable and scalable. In the following sections, we will review the already available tools and formats to place our solution among them.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Background",
"sec_num": "2."
},
{
"text": "The standardisation process of web archiving practices, initiated and controlled mainly by national libraries (Oury and Poll, 2013) , does not provide comprehensive guidelines to the standardised encoding of the texts extracted from web archiving activity. The situation is much better on the level of metadata. The technical report of Statistics and Quality Indicators for Web Archiving stresses the importance of different metadata types for curating web resources 2 : \"Long term preservation also includes keeping safe the metadata associated with the resources in the Web Archive, which are critical for supporting collection management, access and preservation activities\" (ISO/TC 46/SC 8 N). The Metadata Encoding and Transmission Standard (METS) distinguishes four metadata types to be used in curated collections sourced from web archiving: (a) Descriptive metadata, (b) Structural metadata, (c) Provenance metadata, (d) Rights metadata. This is the theoretical standpoint, but since the creation of such metadata requires a lot of manual work, it is impossible to find a collection of archived web documents that complies with these requirements on metadata entirely. Therefore, there is virtually no reliable digital cultural heritage source for researchers. In contrast, there are metadata standards which cover finegrained requirements. The only standard that could gain large-scale adoption is Dublin Core, which is not refined enough to comply with the aforementioned standards. Our repository uses Schema.org, a metadata standard we have chosen for several reasons:",
"cite_spans": [
{
"start": 110,
"end": 131,
"text": "(Oury and Poll, 2013)",
"ref_id": "BIBREF7"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Metadata",
"sec_num": "2.1."
},
{
"text": "\u2022 Schema.org is designed explicitly for storing information about web resources",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metadata",
"sec_num": "2.1."
},
{
"text": "\u2022 It has a dynamic, community based development (in contrast with robust standards, such as METS)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metadata",
"sec_num": "2.1."
},
{
"text": "\u2022 It is increasingly popular on the web, which makes it easy to extract metadata from the HTML source",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metadata",
"sec_num": "2.1."
},
{
"text": "\u2022 It is compatible with semantic web technology (Linked Open Data)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metadata",
"sec_num": "2.1."
},
{
"text": "\u2022 It has a growing usage in the digital cultural heritage domain (e.g. Europeana)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metadata",
"sec_num": "2.1."
},
{
"text": "The Szeged corpus is the largest, manually annotated corpus (Vincze et al., 2010) in the Hungarian language containing 1.2 million tokens, KorKorpusz (31,492 tokens) is similar but smaller corpus based on a recent pilot project (Vad\u00e1sz, 2020) . The first Hungarian gigaword corpus was the Hungarian Gigaword Corpus (Oravecz et al., 2014) with 1,532,933,778 tokens. Both aforementioned corpora contain text only from curated sources (newspapers, literary texts, social media, legal texts, etc.) that are not entirely from the Internet. The first Hungarian web corpus that was created by Kornai and his colleagues (Hal\u00e1csy et al., 2004) is called the Hungarian Webcorpus. It was later superseded by the 1.2 billion token P\u00e1zm\u00e1ny corpus 3 (Endr\u00e9dy and Pr\u00f3sz\u00e9ky, 2016) and the 2.5 billion token HuTenTen corpus (Jakub\u00ed\u010dek et al., 2013) , two larger corpora entirely from the web. Nowadays, large corpora are utilising the Common Crawl archive like the OSCAR corpus (Ortiz Su\u00e1rez et al., 2019) with 5.16 billion (2.33 billion deduplicated) words in Hungarian. However, the documents presented in these corpora often contain gaps due to deduplication. All of these corpora -except the ones based on Common Crawl -have the same flaw, namely that after their creation and publication, several errors were discovered in the tools used to create them, and these errors could not be corrected as expected. The reason being that their original source HTML files have been deleted -and these are unavailable in an unmodified form on the web.",
"cite_spans": [
{
"start": 60,
"end": 81,
"text": "(Vincze et al., 2010)",
"ref_id": "BIBREF21"
},
{
"start": 228,
"end": 242,
"text": "(Vad\u00e1sz, 2020)",
"ref_id": "BIBREF20"
},
{
"start": 315,
"end": 337,
"text": "(Oravecz et al., 2014)",
"ref_id": "BIBREF18"
},
{
"start": 612,
"end": 634,
"text": "(Hal\u00e1csy et al., 2004)",
"ref_id": "BIBREF14"
},
{
"start": 736,
"end": 764,
"text": "(Endr\u00e9dy and Pr\u00f3sz\u00e9ky, 2016)",
"ref_id": "BIBREF13"
},
{
"start": 807,
"end": 831,
"text": "(Jakub\u00ed\u010dek et al., 2013)",
"ref_id": "BIBREF15"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Existing Hungarian Corpora",
"sec_num": "2.2."
},
{
"text": "Since then, there have been numerous attempts to create web-based corpora, but these were not published and could not arouse public interest, as web corpora and crawling became increasingly common tools. The speciality of the corpus and the method presented in this paper lies in the fact that it unites the experience from the above mentioned corpora into a manually curated gigaword web corpus, which includes metadata and can be built from the source of the downloaded web pages in a reproducible manner.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Existing Hungarian Corpora",
"sec_num": "2.2."
},
{
"text": "To put our method into a larger perspective, in the following sections we will describe the process of corpus creation in an abstract workflow (see Figure 1. ), where the elements have to be further specified by certain design decisions.",
"cite_spans": [],
"ref_spans": [
{
"start": 148,
"end": 157,
"text": "Figure 1.",
"ref_id": "FIGREF0"
}
],
"eq_spans": [],
"section": "From the web to a corpus",
"sec_num": "3."
},
{
"text": "Classical web crawling can be characterised by only a few parameters that are set at the start of the process. These parameters include the initial seed of URLs where to start the crawl from, the maximal depth, and the breadth to restrict the crawler's movement. In some cases a set of targeted domains is also specified. Although there are only a few widely used crawler engines, it is hard to characterize these as most web libraries (e.g. Python requests, wget, etc.) can be used for crawling nowadays and the desired output varies from corpora to \"exact offline duplicates\" of websites. Here we would like to mention three crawler engines: both Heritix 4 and Apache Nutch (Laliwala and Shaikh, 2013) are used in the Internet Archive and Common Crawl projects. The third crawler engine is Spiderling (Suchomel and Pomik\u00e1lek, 2012) , which was developed by the authors of Sketch Engine (Kilgarriff et al., 2014) . These crawlers are fast, generalised tools, but for targeted or spe- cialised crawling they became tedious to use. This may explain the numerous different libraries used for crawling. Nowadays, we do not even have to use a crawler as we can begin the process with Common Crawl, The Internet Archive and similar resources. In this case, the first step is to clean the actual data, and remove collected garbage. Due to the nature of the internet, there are numerous aggressive SEO traps out in the web that are used to optimise the page rank in search engines that end up in the archive. These traps are placed in such a manner that they can divert crawler bots from their original course when these bots stumble upon them. Such general bots cannot distinguish \"normal\" pages from these traps, a task that humans are able to carry out in a matter of seconds. Another common problem using these sources is the need for deduplication (see Section 3.3.), which causes the waste of resources on both sides (crawler and deduplicator).",
"cite_spans": [
{
"start": 676,
"end": 703,
"text": "(Laliwala and Shaikh, 2013)",
"ref_id": "BIBREF17"
},
{
"start": 803,
"end": 833,
"text": "(Suchomel and Pomik\u00e1lek, 2012)",
"ref_id": "BIBREF19"
},
{
"start": 888,
"end": 913,
"text": "(Kilgarriff et al., 2014)",
"ref_id": "BIBREF16"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The Web Crawler",
"sec_num": "3.1."
},
{
"text": "To overcome these problems, Indig et al. (2019) built a new NLP-oriented crawler tool called Corpus Builder, which utilises a very different approach from the above: a twolevel crawling method. By using targeted crawling of large or medium portals, they claim that with their method, it is possible to crawl only the targeted portals virtually without duplication, with a small footprint and in a sustainable manner. Their main idea is the exploitation of two different levels of recurring patterns (\"archives\" and \"article\" pages) that can be discovered by analysing the structure of different web pages.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Web Crawler",
"sec_num": "3.1."
},
{
"text": "The first and obvious level is the recurring boilerplate content on the actual web pages of a portal. Objects of the same type are usually stored in the database back end and generated on demand into specifically formatted web pages. This is the internal nature of the data. In this paper, we call these pages \"articles\" regardless of their content type: whether they represent news articles, images with captions in a gallery, product descriptions and customer reviews in a webshop, posts in a forum or blog, etc. The output pages for these content types look the same in narrower time frames for a certain portal, but they can be very different from website to website. These pages are generated on the serverside, so we must collect HTMLs and extract their content. If we collect a mass amount of web pages representing the same objects from the same portal using the same template selectively or classify them after collection, the (uniform) boilerplate can be easily removed with a few simple rules per class, therefore this task does not require complex tools.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Article Crawler",
"sec_num": "3.1.1."
},
{
"text": "The second level arises from the first: how can we systematically collect such web pages on a given portal? The answer is very simple: portals created for human readers are very likely to have some kind of a \"table of contents\", \"product or topic list\" or an \"article archive\", which display all available pages in a structured form. Because objects are traditionally stored in taxonomies -e.g. temporal (years, months) or other feature-based (colour, shape, price, etc.)that can be enumerated and each object has finite number of possible values. If we enumerate articles for right feature values, we will gather links to all pages of the same layout systematically from the given portal.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Archive Crawler",
"sec_num": "3.1.2."
},
{
"text": "Using the two-step method described above, it is possible to gather a massive number of web pages even from only a small number of portals that will have virtually no duplication and effectively zero garbage pages in contrast to the general crawling methodology. This method has been successfully tested on three Hungarian news portals (Indig et al., 2019) , while the further generalisation of the method for the steps following the crawling of different portals with different schemes and layouts requires further elaboration. Indig et al. (2019) assembled the minimal number of parameters that are needed to handle such portals in a unified framework. The major highlights of the configuration are showcased as the following:",
"cite_spans": [
{
"start": 336,
"end": 356,
"text": "(Indig et al., 2019)",
"ref_id": "BIBREF3"
},
{
"start": 529,
"end": 548,
"text": "Indig et al. (2019)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The Possible Parameters for Portals",
"sec_num": "3.1.3."
},
{
"text": "\u2022 The date of the first and last article, or page number where applicable",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Possible Parameters for Portals",
"sec_num": "3.1.3."
},
{
"text": "\u2022 The archive URL format with the placeholders to be substituted",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Possible Parameters for Portals",
"sec_num": "3.1.3."
},
{
"text": "\u2022 The function for finding the next-page URL for the archive where applicable",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Possible Parameters for Portals",
"sec_num": "3.1.3."
},
{
"text": "\u2022 The function to get the article URLs from the archive pages",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Possible Parameters for Portals",
"sec_num": "3.1.3."
},
{
"text": "\u2022 Boolean variables to answer the following questions: is the archive paginated, infinite scrolling, or datebased?",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Possible Parameters for Portals",
"sec_num": "3.1.3."
},
{
"text": "One can distinguish between crawl-based (applicable for the current crawl), portal-based (which applies to the crawled portal regardless of crawl settings), and portalspecific configurations. Our method follows the latter direction for crawling. For problems not addressed by Indig et al. (2019) we present our solutions in Section 4.",
"cite_spans": [
{
"start": 276,
"end": 295,
"text": "Indig et al. (2019)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The Possible Parameters for Portals",
"sec_num": "3.1.3."
},
{
"text": "There are a lot of pages that present the same type of objects or articles surrounded by the same boilerplate content (i.e. scheme or template) -menu, advertisements, etc. in a portal. In most cases, this boilerplate content can be characterised by not having multiple paragraphs with many lines of text, but containing short texts, as well as many links and pictures (Pomik\u00e1lek, 2011) . The process of boilerplate removal can be broken down into two steps presented in the following sections.",
"cite_spans": [
{
"start": 368,
"end": 385,
"text": "(Pomik\u00e1lek, 2011)",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Boilerplate Removal",
"sec_num": "3.2."
},
{
"text": "By normalisation we mean the reformatting of visual elements into a simpler visual annotation (e.g. the elements of Markdown language or XML tags) to create a common ground for the text of different portals in the resulting corpus. Normalisation is not a trivial task: most tools extract paragraphs as plain text, however, visual formatting elements are essential for humans and may also help the machine reader, therefore these elements should be kept and standardised.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Normalisation",
"sec_num": "3.2.1."
},
{
"text": "Curated metadata is the cornerstone of proper (web) archiving. It can be regarded as gold standard labels for each document, which can later be utilised for training or benchmarking ML algorithms (i.e. authorship attribution, keyword extraction, topic modelling, etc.). There are automatic tools for extracting metadata from the crawled web pages such as the Web Curator Tool 5 or Apache Tika 6 . These tools extract standards compliant descriptive metadata automatically from the crawled web pages, but they are very complex and it is difficult to understand and improve their method for the specific portals. Moreover, they are plagued with the same problems as other boilerplate removal tools (see Section 3.2.3.): their heuristics and output formats are wired in by design and it is very hard to change these without major conflicts. When the these programs yield deficient output for the targeted portals -for example due to the lack of knowledge about the typographical rules of the language, or when the output is missing some important variables, -it is inevitable to implement a custom metadata extractor methodology. We decided to use this method to allow future modifications, to be able to compare results with the presented generic tools (see Section 3.2.3.), and also to demonstrate how easily our method can be implemented. Our findings will be described in Section 4..",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metadata extraction",
"sec_num": "3.2.2."
},
{
"text": "As web page layouts, coding styles, and HTML standards differ throughout the portals and were used differently over the years, the boilerplate removal task is mostly solved by clever heuristics, which makes it hard for the users to create general measurements and comparisons between them. It is also hard to set their parameters, fix, extend or modify their functionality. Some tools are designed to remove boilerplate from a single page, while others use multiple pages of the same layout to delete recurring elements (Endr\u00e9dy and Nov\u00e1k, 2013) . In this paper, we could not survey all the available methods, therefore we are comparing JusText (Pomik\u00e1lek, 2011) , a tool created directly for NLP-centric web crawling and Newspaper3k (Ou-Yang, 2013), created especially for news portal crawling. Both modules are still popular and widely-used because of their simplicity and effectiveness. They both remove formatting and yield plain text paragraphs, but the latter tool supports extracting metadata from web pages and has other features to simplify the crawling process for beginners. We followed the route marked by Indig et al. (2019) and created our own handcrafted boilerplate removal rules. At first we found ourselves in a dilemma about choosing between regular expressions and HTML parsers. Regular expressions are simple, and it is also easier to train machines to create them, while HTML parsers are easier to create, read and maintain for humans, but are harder to automate. As some of the portals recently added to the corpus have very complex layouts, it is not feasible to further extend the number of portals using regular expressions. For example, it may be impossible or become very unpractical to encode various attributes and their combinations (which might be in arbitrary order due to the structure of HTML). We compared the aforementioned methods on our gold standard data set 7 . This measurement is presented in Section 5., followed by other details of our method.",
"cite_spans": [
{
"start": 520,
"end": 545,
"text": "(Endr\u00e9dy and Nov\u00e1k, 2013)",
"ref_id": "BIBREF1"
},
{
"start": 645,
"end": 662,
"text": "(Pomik\u00e1lek, 2011)",
"ref_id": "BIBREF8"
},
{
"start": 1118,
"end": 1137,
"text": "Indig et al. (2019)",
"ref_id": "BIBREF3"
},
{
"start": 1899,
"end": 1900,
"text": "7",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Existing Tools and Techniques",
"sec_num": "3.2.3."
},
{
"text": "Sometimes the exact same content -available on several domains -can be stored in the web archive multiple times, but, of course, we need one intstance only. There are great tools for deduplication (like Onion (Pomik\u00e1lek, 2011) ), but their waste of valuable resources, such as disk space and network bandwidth is not ideal. When using targeted crawling, such as Indig et al. 2019, we can select only those distinct URLs which are needed and so bypass the process of deduplication.",
"cite_spans": [
{
"start": 209,
"end": 226,
"text": "(Pomik\u00e1lek, 2011)",
"ref_id": "BIBREF8"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Deduplication and NLP",
"sec_num": "3.3."
},
{
"text": "The main problem with deduplication -besides wasting resources -is that some parts of a document or the whole document may become missing because it had been recognised and deleted as a duplicate. This undermines the completeness of the crawling which is the the first priority for some user groups (e.g. humanists and sociologists). The publicly available corpora that were created for NLP purposes have further disabilities: their sentences are scrambled to avoid the infringement of copyright laws. This makes the analysis of full documents -an emergent trend -impossible. The role -and legal privilege -of national libraries is to preserve documents in its entirety, even for born-digital materials. This role can be fulfilled with our method, in contrast to the traditional ones. Different levels of NLP annotation can optionally be applied before or between the deduplication with the plethora of available tools. Until recently, texts have been stored only after this step in some format, however, the increasing performance of NLP tools makes it advisable to store crawled content also in raw format (e.g. WARC files) to be able to correct errors found in the processing pipeline. This is mainly important to humanists, sociologists and other scholars outside NLP where the specific text is the subject of analysis, in contrast to NLP, where only the amount of text matters.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Deduplication and NLP",
"sec_num": "3.3."
},
{
"text": "The process of creating the output from the HTML files can be split into four steps for easier maintainability:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "\u2022 Simplification of HTML by finding the tightest bounding HTML tag of the whole text content and decomposing unneeded subtrees 8",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "\u2022 Extraction of paragraphs and metadata from the HTML tree keeping only specific -intended -formatting",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "\u2022 Rewriting elements to a unified format by standardising site-specific formatting",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "\u2022 Writing the output file according to the expected format. In this step, the fields get their final place and canonical names",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "The first three steps contain well-defined portal specific instructions, while the fourth is only dependent on the output format, which -as it is totally separated from the otherscan comply with the actual purpose and front end in the future. Some user groups have special requirements, such as full documents and metadata, while others only require the raw text. Nonetheless, both requirements can be achieved at the same time.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "In the field of NLP, three main use cases exist. To search patterns in large corpora, the classic vertical format used primarily by the Sketch Engine (Kilgarriff et al., 2014) is recommended. If the aim is to process the corpus with a wide variety of standard NLP tools, the CoNLL-U format 9 is adequate. If the goal is to put documents to a full text search engine or into a language model, it is necessary to comply with the input expectation of such software, which is usually raw text.",
"cite_spans": [
{
"start": 150,
"end": 175,
"text": "(Kilgarriff et al., 2014)",
"ref_id": "BIBREF16"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "In the field of digital humanities, -especially in philology,the XML document markup language and the Text Encoding Initiative (TEI) recommendation have become dominant over the decades (Schreibman et al., 2008) . TEI makes the versioning and annotation of the enriched articles possible in an easy and reliable way, and it is also capable of storing metadata and the body of the document structurally in one file. This format satisfies NLP users as well, while opening the resulting corpus for other audiences including librarians, humanists and sociologists. TEI also allows the verification of the authenticity of the source text by the metadata and increases the reproducibility of research which has an increasing importance in the 'distant reading' paradigm (Da, 2019) . Text can be converted to a simpler form corresponding to the actual use case, while keeping the master copy untouched, in a similarly to how it is done with images by resizing and cropping them on demand dynamically.",
"cite_spans": [
{
"start": 186,
"end": 211,
"text": "(Schreibman et al., 2008)",
"ref_id": "BIBREF9"
},
{
"start": 764,
"end": 774,
"text": "(Da, 2019)",
"ref_id": "BIBREF0"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The Final Format and Front End",
"sec_num": "3.4."
},
{
"text": "We examined several Hungarian news portals and increased the number of examined portals to six, compared to the three portals examined by Indig et al. (2019) in order to test how the presented method can be applied to portals of different structures. First, we selected mainstream Hungarian news portals, because these contain a vast number of articles. As a secondary priority, we included portals that are special from the perspective of used web technology and architecture. We wanted to reach a milestone, where adding new portals and maintaining the existing ones is a routine task that can be handled by as little manpower as possible. In this section, we describe the main highlights of our crawling method compared to (Indig et al., 2019 ) (for further comparisons see Section 3.)",
"cite_spans": [
{
"start": 138,
"end": 157,
"text": "Indig et al. (2019)",
"ref_id": "BIBREF3"
},
{
"start": 726,
"end": 745,
"text": "(Indig et al., 2019",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Method",
"sec_num": "4."
},
{
"text": "We decided to change the regular expressions used in Corpusbuilder (Indig et al., 2019) for Python functions, which use an HTML parser to handle the input as an HTML tree. Using HTML trees enabled us both to simplify many regular expression patterns and to support many different layouts. With this change, the accuracy of extracting article URLs from the page archives has dramatically increased, as we found that on some portals different columns may be hosted on different domains, or -while using the same site template -they may not match the expressions written for extracting URLs. This can be recognised by tree searching expressions more easily than with regular expressions. This, of course, sacrifices speed for clarity and precision, but saves us from HTML fragments slipping through regular expressions.",
"cite_spans": [
{
"start": 67,
"end": 87,
"text": "(Indig et al., 2019)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "HTML Parsers vs. Regular Expressions",
"sec_num": "4.1."
},
{
"text": "The date-based pagination handling logic (Indig et al., 2019) was separated from other pagination methods, as it allows sorting and can be used to filter crawling by specific date intervals, since we found that date-based pagination can be and is combined freely with the other methods. We also introduced support for open (date) intervals. Our other significant change was in handling infinite scrolling 10 and active archives 11 together in an easy-tounderstand form by extracting the page URLs before determining the URL of the next archive page. We have broken down the possible patterns of finding the next archive page URL to the following cases:",
"cite_spans": [
{
"start": 41,
"end": 61,
"text": "(Indig et al., 2019)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "The Refined Archive Crawler",
"sec_num": "4.2."
},
{
"text": "\u2022 There is no pagination at all",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Refined Archive Crawler",
"sec_num": "4.2."
},
{
"text": "\u2022 There is a next page link which we need to use",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Refined Archive Crawler",
"sec_num": "4.2."
},
{
"text": "\u2022 There is infinite scrolling: we use the page number from the base value to \"infinity\" where no more article URLs are detected",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Refined Archive Crawler",
"sec_num": "4.2."
},
{
"text": "\u2022 There is page numbering: we use the page number from the base value to a portal-specific maximum",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Refined Archive Crawler",
"sec_num": "4.2."
},
{
"text": "\u2022 There is page numbering, but we expect the archive to expand during crawling (can be identified by finding known article URLs during crawling)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Refined Archive Crawler",
"sec_num": "4.2."
},
{
"text": "By using these features, all examined portals could be handled, therefore we narrowed down our experiments to six portals that showcase all of the described features, and allows them to be thoroughly tested.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Refined Archive Crawler",
"sec_num": "4.2."
},
{
"text": "Metadata can be extracted from multiple sources from an article page. We identified and handled the following:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Advanced Metadata Extraction",
"sec_num": "4.3."
},
{
"text": "\u2022 Date, title and column are frequently encoded in the URLs",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Advanced Metadata Extraction",
"sec_num": "4.3."
},
{
"text": "\u2022 HTML meta tag properties which can be encoded according to various conventions (like Dublin Core, Schema.org, etc.) that are mainly included for Search Engine Optimization (SEO) purposes",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Advanced Metadata Extraction",
"sec_num": "4.3."
},
{
"text": "\u2022 The increasingly popular JSON-LD, storing properties that were previously stored as meta tags, but in a more structured form",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Advanced Metadata Extraction",
"sec_num": "4.3."
},
{
"text": "\u2022 From the content of the HTML itself, where it is included to be displayed for the user There are several portals that use more than one of the above sources of metadata. We also found examples where different sources yielded contradicting results or missing values, these are probably due to bugs in the websites' engines. Older articles tend to have more of these errors as they were probably converted from a previous layout and the conversion introduced such errors 12 . Some portals partially or fully generate metadata dynamically by using JavaScript and non-standard data-sources. This practically makes it impossible to extract such metadata with traditional tools and forces us to use a portal-specific solution.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Advanced Metadata Extraction",
"sec_num": "4.3."
},
{
"text": "To handle millions of pages without reading them -through \"distant reading\" -, we invented utilities to examine, analyse and normalise the tags and the scheme used by a portal, and then freely convert it to the new and customisable output format. We started with cutting the HTML to the relevant part, as mentioned in Section 3.4.. The first utility function helps to filter out tags that do not contain any text. Next, we introduced placeholders to simplify some elements (e.g. links). The second function aids in simplifying the tags by manually selecting groups that belong to the same class (e.g. formatting, embedded content, etc.), but are specialised to the portal's scheme. This method is quite effective even without portal-specific parameters. Table 1 shows how the number of tags (from one of the examined domains) is reduced after using these tools allowing further fine-grained modifications in an iterative manner.",
"cite_spans": [],
"ref_spans": [
{
"start": 754,
"end": 761,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Converting HTML to the Output Format",
"sec_num": "4.4."
},
{
"text": "No. of tags % all tags 33,466 100 text containing tags 18,517 55 after simplify tags 359 10 relevant tags 267 7 Table 1 : Illustration of how the number of tags to be analysed manually decreases in magnitude.",
"cite_spans": [],
"ref_spans": [
{
"start": 112,
"end": 119,
"text": "Table 1",
"ref_id": null
}
],
"eq_spans": [],
"section": "Converting HTML to the Output Format",
"sec_num": "4.4."
},
{
"text": "Possible layouts for all URLs of a domain were described with the help of a tree-representation: the subtrees of the contents' tightest bounding HTML tag for all pages were merged, counting the frequencies of each element and the cumulative length of their immediate text. It was also marked if a specific tag had no child elements in the tree. The resulting frequency distribution allows efficient examination and handling of subtrees for all URLs at once. In order to be able to make decisions concerning the remaining tags, we built a tag dictionary. To each tag (or simplified tag), we assigned the average length of the contained text, the average number of descendants, and the average length of the immediate text supplemented with a sample of occurrences (URLs). This dictionary was augmented with the operation to perform at each occurrence of that specific tag. As we formalised the operators, their execution was made by the code automatically generated from the dictionary. These steps can be iterated to gain more insight on the portal's scheme and finally arrive to the desired form.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "The Tree Representation",
"sec_num": "4.4.1."
},
{
"text": "When standardising and rewriting elements, we found the following operators useful:",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "\u2022 decomposing (deleting the tag with its contents, e.g. advertisements, boilerplate)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "\u2022 unwrapping (deleting the tag, keeping its contents, e.g. text anchors)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "\u2022 unwrapping all descendants (simplifying a block)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "\u2022 rewriting tags context-free",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "\u2022 splitting tags to super-subordinate pairs (e.g. when the content and formatting properties are in the same tag)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "\u2022 rewriting tags context-specific (special blocks)",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "These operators can be applied sequentially in the proper order for every URL. We narrowed down the various layouts (e.g. left, right, top block) into a few, portal independent types of blocks that we intended to keep. The contextspecific rules mark the root tag for each block we found, so their subtrees can be handled by independent dictionaries in the same way. The analysis of the visual layout of the examined portals shows that there are no blocks embedded into other blocks. This property allows us to rely on the described two-level transformation with a low number of distinct tag dictionaries modified by the defined operators. To conclude, normalising the tags and then rewriting them to the final schema are independent steps which can be achieved with successive approximation in an iterative manner. This allows us fine-grained control to change design decisions or customise the output (TEI XML in our case) easily at all times.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Rewriting Rules and Transformation Methods",
"sec_num": "4.4.2."
},
{
"text": "We ran our crawling on a low-end desktop machine (Intel i3, 4 GB RAM) for 30 days on a 100 MB/s connection (with rate-limiting to avoid the hammering of the remote servers) using circa 100 GB of disk space to demonstrate the effectivity of the method presentes here. It is not possible to compare this method's crawling performance to other general crawler engines mentioned earlier, as the workflow and methodology differ significantly (see Section 3.1.). It is possible, however, to compare the crawling accuracy to the most widely used archiving practice: the Internet Archive (see Section 5.2.). It is also possible to compare our sitespecific rule-based boilerplate removal and metadata extractor functions to the mainstream crawling methods (see Section 3.2.3.). The goal of the compared tools and their design differs significantly so the way how to make an objective comparison was not at all obvious. When comparing our method with the aforementioned tools, we strived to highlight performance differences due to design, while separating them from the strengths and weaknesses stemming from the methods themselves.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Evaluation",
"sec_num": "5."
},
{
"text": "We extracted a total of 2,227,180 articles from six Hungarian news portals, this signifies 984 million words (without tokenisation) of extracted text without metadata from November 1998 until September 2019. We visualised the annual distribution of articles to see the estimated growth in the number of articles and the expected number of articles per year (see Figure 2) . The figure shows a clearly growing tendency in the number of articles published on the crawled portals during the last twenty years -except 2019, which does not qualify as a full year at the time of measurement. In the case of the six portals, this means that more than 200 articles have been published on average every day in the recent years. These numbers tells us that by adding new portals the quantity of the crawled articles and the volume of the corpus can be increased quickly and easily with low human resource investment and a lightweight technical infrastructure.",
"cite_spans": [],
"ref_spans": [
{
"start": 362,
"end": 371,
"text": "Figure 2)",
"ref_id": null
}
],
"eq_spans": [],
"section": "The data set",
"sec_num": "5.1."
},
{
"text": "1 9 9 7 1 9 9 9 2 0 0 1 2 0 0 3 2 0 0 5 2 0 0 7 2 0 0 9 2 0 1 1 2 0 1 3 2 0 1 5 2 0 1 7 2 0 1 9 No. of Articles Figure 2 : The annual distribution of 2,227,180 articles from six portals from November 1998 to September 2019. The number of articles per year is increasing. The decrease at 2019 is due to the fact that it is not a full year at the time of measurement.",
"cite_spans": [],
"ref_spans": [
{
"start": 112,
"end": 120,
"text": "Figure 2",
"ref_id": null
}
],
"eq_spans": [],
"section": "The data set",
"sec_num": "5.1."
},
{
"text": "In Table 2 , we can see the performance of the boilerplate removal tools in different scenarios. We examined Jus-Text and Newspaper3k on the full HTML code, the article body and constrained to the original and the cleaned up paragraphs. We wanted to check whether an educated initial guess (on the text's location) helps these programs or not. As the former package does not extract metadata separately, we present numbers with metadata and provide the number of words without metadata in brackets. The numbers have some small differences that suggests that a more detailed evaluation of the content is needed. We also compared the actual values of the extracted metadata (author, publication date, title) in terms of precision and recall for Newspaper3k (see Table 3 ). Our educated initial guess does not help metadata extraction, but for the text extraction it has a potential because it rules out unwanted content in one step. It is clear from these that our method is superior to the compared ones, however, a content-based comparison of the extracted paragraphs is needed in order to be able to evaluate the mentioned methods objectively. We argue that if full articles are chosen, the precision provided by our method is needed to ensure that the right amount and quality of texts can be extracted with the compared methods. ",
"cite_spans": [],
"ref_spans": [
{
"start": 3,
"end": 10,
"text": "Table 2",
"ref_id": "TABREF2"
},
{
"start": 760,
"end": 767,
"text": "Table 3",
"ref_id": "TABREF3"
}
],
"eq_spans": [],
"section": "The data set",
"sec_num": "5.1."
},
{
"text": "We compared our results of the six crawled news portals to the Internet Archive as the \"standard\" source of web archiving. We evaluated whether the same set of URLs could be acquired using the Internet Archive, and also compared the number of crawled articles by portals with data downloaded from The Wayback Machine. In the following step, based on the mime type attribute, we removed all URLs from the Internet Archive data sets that represent content other than articles (e.g. images, scripts, etc.). Using the status code variable, we omitted all URLs that were not successfully downloaded for some reason (e.g. 404 errors and redirections). From our crawl we selected the timestamp of the last article downloaded for each domain, and removed all URLs from the Internet Archive data that were crawled after that date. At this point, we still had hundreds of thousands of URLs in the Internet Archive data sets that represented e.g. certain taxonomy pages (date, category, author, search, etc.) or any kind of content other than single articles. Thus, we introduced a domain-level cleaning function for each crawled website, in order to remove all URLs representing content other than articles. This proved to be a difficult, time-consuming, iterative task, as in case of some websites, the URL structure changed multiple times over the years, making it nearly impossible to retrospectively identify URLs that certainly lead to articles. This is one important aspect why our method is much easier to use (even retrospectively), when the goal is to produce a clean corpus, without duplicated content. In the case of several websites, the URL structure was not logically constructed (e.g. tag archives have the same URL structure as articles; randomly generated version numbers appear at the end of some of the URLs, but not all of them; etc.), therefore in some cases, we had to restrict the comparison to certain columns of the portal, as it was very difficult to clean the data sets in a more generalised way. Our next step was to normalise all URLs in both crawls. We removed http, https, www from the beginning, and port numbers (e.g. \":80\") and parameters from the end of the URL strings. Using these normalised URLs, we created two dictionaries to store the URLs themselves and their slugs -the last part of the URL (after the last /) -for each portal. For some portals the URLs could not be used for a valid comparison, because the URL structure has changed over time, but not the slug, therefore -in these cases -we used the slug for our comparison.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Crawling Compared to Archive.org",
"sec_num": "5.2."
},
{
"text": "With the steps described above, we reduced the number of Archive.org URLs from 8.9 million to only 1.2 million for the six crawled portals. After removing entries with wrong status codes 75.4%, after mime-type-based cleaning 53.7% of the URLs remained. While only 0.7% of URLs were removed in the date-based cleaning phase, after running website-specific cleaning functions and compiling the final list of URLs, just 13.5% of the initial number of URLs remained. We found that 846,343 articles are present both in our crawl and in the Internet Archive's data, while 1,082,484 articles are only present in the ELTE.DH corpus. A further 315,649 articles are only found in Archive.org's data. More work is needed in order to eliminate all possible bad URLs, however, it is safe to say that by using our crawler it is easier to achieve the same results than finding and downloading all relevant content from Archive.org.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Crawling Compared to Archive.org",
"sec_num": "5.2."
},
{
"text": "We have demonstrated that by using a low-end machinewhich has similar computational power as our smartphones, the storage capacity of our pendrives nowadays -and minimal manpower it is possible to create a gold-standard quality gigaword corpus with metadata which suits many audiences at the same time 13 . As the presented work was only a pilot study to design and stabilise the workflow on many candidate pages, we plan to apply this methodology on several more websites, and start serving requests on sitespecific crawling to provide data for research in multiple disciplines in a future version of this corpus. In conjunction with the previously outlined plans, we intend to support national libraries with our research as they are responsible of keeping the data of our present for the future researchers who can thus provide objective and balanced research. One obvious step in this direction is to conduct research on how to keep the authenticity of web archives and how to eliminate the risks of tampering, retroactive modification and deletion of content which undermine scholarly credibility. We plan to utilise digital fingerprinting, signatures and blockchain technology on downloaded documents in order to keep them safe, while making them available for the widest possible audience. 13 The software is published under the GNU Lesser General Public License v3.0 at https://github. com/ELTE-DH/WebArticleCurator and https: //doi.org/10.5281/zenodo.3755323",
"cite_spans": [
{
"start": 1297,
"end": 1299,
"text": "13",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Conclusion and Future Work",
"sec_num": "6."
},
{
"text": "Winters (2017, 240) deals with the problem of website dates in detail.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "http://netpreserve.org/resources/IIPC_ project-SO_TR_14873__E__2012-10-02_DRAFT. pdf",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "The P\u00e1zm\u00e1ny corpus was the first Hungarian corpus which separated edited text (news articles) from unedited text (comments).4 https://github.com/internetarchive/ heritrix3",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "https://webcuratortool.readthedocs.io/ 6 http://tika.apache.org/",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "Some elements were kept or thrown away by design decision that may not match with the compared tools or future use cases. However, we support the change of these decisions by the user.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "There are three classes of decomposing rules: a) general rules used for every portal, b) \"must-have\" portal-specific rules, c) rules which follow certain design decisions about the data to be extracted.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "https://universaldependencies.org/ format.html",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "A technique used to dynamically add new content to the page when the user scrolls down.11 If new elements are added to the archive during crawling, the list of articles will be divided to pages in a way that their content URLs will appear on different pages than as expected. This makes it impossible to handle archive pages' URLs as permalinks.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
},
{
"text": "This can be solved by crawling articles as soon as possible after their publication.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "",
"sec_num": null
}
],
"back_matter": [],
"bib_entries": {
"BIBREF0": {
"ref_id": "b0",
"title": "The computational case against computational literary studies",
"authors": [
{
"first": "N",
"middle": [
"Z"
],
"last": "Da",
"suffix": ""
}
],
"year": 2019,
"venue": "Critical inquiry",
"volume": "45",
"issue": "3",
"pages": "601--639",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Da, N. Z. (2019). The computational case against compu- tational literary studies. Critical inquiry, 45(3):601-639.",
"links": null
},
"BIBREF1": {
"ref_id": "b1",
"title": "More effective boilerplate removal-the goldminer algorithm",
"authors": [
{
"first": "I",
"middle": [],
"last": "Endr\u00e9dy",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Nov\u00e1k",
"suffix": ""
}
],
"year": 2013,
"venue": "Polibits",
"volume": "1",
"issue": "48",
"pages": "79--83",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Endr\u00e9dy, I. and Nov\u00e1k, A. (2013). More effective boilerplate removal-the goldminer algorithm. Polibits, 1(48):79-83.",
"links": null
},
"BIBREF2": {
"ref_id": "b2",
"title": "The Digital Humanities: A Primer for Students and Scholars",
"authors": [
{
"first": "E",
"middle": [],
"last": "Gardiner",
"suffix": ""
},
{
"first": "R",
"middle": [
"G"
],
"last": "Musto",
"suffix": ""
}
],
"year": 2015,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Gardiner, E. and Musto, R. G. (2015). The Digital Human- ities: A Primer for Students and Scholars. Cambridge University Press.",
"links": null
},
"BIBREF3": {
"ref_id": "b3",
"title": "Crawling in reverse -lightweight targeted crawling of news portals",
"authors": [
{
"first": "B",
"middle": [],
"last": "Indig",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "K\u00e1konyi",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Nov\u00e1k",
"suffix": ""
}
],
"year": 2019,
"venue": "Proceedings of the 9th Language & Technology Conference: Human Language Technologies as a Challenge for Computer Science and Linguistics",
"volume": "",
"issue": "",
"pages": "81--87",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Indig, B., K\u00e1konyi, T., and Nov\u00e1k, A. (2019). Crawling in reverse -lightweight targeted crawling of news portals. In Marek Kubis, editor, Proceedings of the 9th Language & Technology Conference: Human Language Technolo- gies as a Challenge for Computer Science and Linguis- tics, pages 81-87, Pozna\u0144, Poland, may. Wydawnictwo Nauka i Innowacje.",
"links": null
},
"BIBREF4": {
"ref_id": "b4",
"title": "Asynchronous Pipeline for Processing Huge Corpora on Medium to Low Resource Infrastructures",
"authors": [
{
"first": "P",
"middle": [
"J"
],
"last": "Ortiz Su\u00e1rez",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Sagot",
"suffix": ""
},
{
"first": "L",
"middle": [],
"last": "Romary",
"suffix": ""
}
],
"year": 2019,
"venue": "7th Workshop on the Challenges in the Management of Large Corpora",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ortiz Su\u00e1rez, P. J., Sagot, B., and Romary, L. (2019). Asynchronous Pipeline for Processing Huge Corpora on Medium to Low Resource Infrastructures. In Pi- otr Ba\u0144ski, et al., editors, 7th Workshop on the Chal- lenges in the Management of Large Corpora (CMLC-",
"links": null
},
"BIBREF5": {
"ref_id": "b5",
"title": "Leibniz-Institut f\u00fcr Deutsche Sprache",
"authors": [],
"year": null,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": ", Cardiff, United Kingdom, July. Leibniz-Institut f\u00fcr Deutsche Sprache.",
"links": null
},
"BIBREF6": {
"ref_id": "b6",
"title": "Newspaper3k: Article scraping and curation",
"authors": [
{
"first": "L",
"middle": [],
"last": "Ou-Yang",
"suffix": ""
}
],
"year": 2013,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Ou-Yang, L. (2013). Newspaper3k: Article scraping and curation. https://github.com/codelucas/ newspaper.",
"links": null
},
"BIBREF7": {
"ref_id": "b7",
"title": "Counting the uncountable: statistics for web archives",
"authors": [
{
"first": "C",
"middle": [],
"last": "Oury",
"suffix": ""
},
{
"first": "R",
"middle": [],
"last": "Poll",
"suffix": ""
}
],
"year": 2013,
"venue": "Performance Measurement and Metrics",
"volume": "14",
"issue": "2",
"pages": "132--141",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Oury, C. and Poll, R. (2013). Counting the uncountable: statistics for web archives. Performance Measurement and Metrics, 14(2):132-141.",
"links": null
},
"BIBREF8": {
"ref_id": "b8",
"title": "Removing boilerplate and duplicate content from web corpora",
"authors": [
{
"first": "J",
"middle": [],
"last": "Pomik\u00e1lek",
"suffix": ""
}
],
"year": 2011,
"venue": "Masaryk university, Faculty of informatics",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Pomik\u00e1lek, J. (2011). Removing boilerplate and duplicate content from web corpora. Ph.D. thesis, Masaryk uni- versity, Faculty of informatics, Brno, Czech Republic.",
"links": null
},
"BIBREF9": {
"ref_id": "b9",
"title": "A companion to digital humanities",
"authors": [
{
"first": "S",
"middle": [],
"last": "Schreibman",
"suffix": ""
},
{
"first": "R",
"middle": [],
"last": "Siemens",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Unsworth",
"suffix": ""
}
],
"year": 2008,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Schreibman, S., Siemens, R., and Unsworth, J. (2008). A companion to digital humanities. John Wiley & Sons.",
"links": null
},
"BIBREF10": {
"ref_id": "b10",
"title": "Methods and approaches to using web archives in computational communication research",
"authors": [
{
"first": "M",
"middle": [
"S"
],
"last": "Weber",
"suffix": ""
}
],
"year": 2018,
"venue": "Communication Methods and Measures",
"volume": "12",
"issue": "2-3",
"pages": "200--215",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Weber, M. S. (2018). Methods and approaches to us- ing web archives in computational communication re- search. Communication Methods and Measures, 12(2- 3):200-215.",
"links": null
},
"BIBREF11": {
"ref_id": "b11",
"title": "Coda: Web archives for humanities research -some reflections",
"authors": [
{
"first": "J",
"middle": [],
"last": "Winters",
"suffix": ""
}
],
"year": 2017,
"venue": "",
"volume": "",
"issue": "",
"pages": "238--248",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Winters, J., (2017). Coda: Web archives for humanities re- search -some reflections, pages 238-248. UCL Press.",
"links": null
},
"BIBREF13": {
"ref_id": "b13",
"title": "A p\u00e1zm\u00e1ny korpusz",
"authors": [
{
"first": "I",
"middle": [],
"last": "Endr\u00e9dy",
"suffix": ""
},
{
"first": "G",
"middle": [],
"last": "Pr\u00f3sz\u00e9ky",
"suffix": ""
}
],
"year": 2016,
"venue": "Nyelvtudom\u00e1nyi K\u00f6zlem\u00e9nyek",
"volume": "112",
"issue": "",
"pages": "191--205",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Endr\u00e9dy, I. and Pr\u00f3sz\u00e9ky, G. (2016). A p\u00e1zm\u00e1ny korpusz. Nyelvtudom\u00e1nyi K\u00f6zlem\u00e9nyek, 112:191-205.",
"links": null
},
"BIBREF14": {
"ref_id": "b14",
"title": "Creating open language resources for Hungarian",
"authors": [
{
"first": "P",
"middle": [],
"last": "Hal\u00e1csy",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Kornai",
"suffix": ""
},
{
"first": "L",
"middle": [],
"last": "N\u00e9meth",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Rung",
"suffix": ""
},
{
"first": "I",
"middle": [],
"last": "Szakad\u00e1t",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "Tr\u00f3n",
"suffix": ""
}
],
"year": 2004,
"venue": "Proceedings of the Fourth International Conference on Language Resources and Evaluation (LREC'04)",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Hal\u00e1csy, P., Kornai, A., N\u00e9meth, L., Rung, A., Szakad\u00e1t, I., and Tr\u00f3n, V. (2004). Creating open language resources for Hungarian. In Proceedings of the Fourth Interna- tional Conference on Language Resources and Evalua- tion (LREC'04), Lisbon, Portugal, May. European Lan- guage Resources Association (ELRA).",
"links": null
},
"BIBREF15": {
"ref_id": "b15",
"title": "The tenten corpus family",
"authors": [
{
"first": "M",
"middle": [],
"last": "Jakub\u00ed\u010dek",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Kilgarriff",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "Kov\u00e1\u0159",
"suffix": ""
},
{
"first": "P",
"middle": [],
"last": "Rychl\u1ef3",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "Suchomel",
"suffix": ""
}
],
"year": 2013,
"venue": "7th International Corpus Linguistics Conference CL",
"volume": "",
"issue": "",
"pages": "125--127",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Jakub\u00ed\u010dek, M., Kilgarriff, A., Kov\u00e1\u0159, V., Rychl\u1ef3, P., and Suchomel, V. (2013). The tenten corpus family. In 7th International Corpus Linguistics Conference CL, pages 125-127.",
"links": null
},
"BIBREF16": {
"ref_id": "b16",
"title": "The sketch engine: ten years on. Lexicography",
"authors": [
{
"first": "A",
"middle": [],
"last": "Kilgarriff",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "Baisa",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Bu\u0161ta",
"suffix": ""
},
{
"first": "M",
"middle": [],
"last": "Jakub\u00ed\u010dek",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "Kov\u00e1\u0159",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Michelfeit",
"suffix": ""
},
{
"first": "P",
"middle": [],
"last": "Rychl\u00fd",
"suffix": ""
},
{
"first": "V",
"middle": [],
"last": "Suchomel",
"suffix": ""
}
],
"year": 2014,
"venue": "",
"volume": "",
"issue": "",
"pages": "7--36",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Kilgarriff, A., Baisa, V., Bu\u0161ta, J., Jakub\u00ed\u010dek, M., Kov\u00e1\u0159, V., Michelfeit, J., Rychl\u00fd, P., and Suchomel, V. (2014). The sketch engine: ten years on. Lexicography, pages 7-36.",
"links": null
},
"BIBREF17": {
"ref_id": "b17",
"title": "Web Crawling and Data Mining with Apache Nutch",
"authors": [
{
"first": "Z",
"middle": [],
"last": "Laliwala",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Shaikh",
"suffix": ""
}
],
"year": 2013,
"venue": "",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Laliwala, Z. and Shaikh, A. (2013). Web Crawling and Data Mining with Apache Nutch. Packt Publishing.",
"links": null
},
"BIBREF18": {
"ref_id": "b18",
"title": "The Hungarian Gigaword corpus",
"authors": [
{
"first": "C",
"middle": [],
"last": "Oravecz",
"suffix": ""
},
{
"first": "T",
"middle": [],
"last": "V\u00e1radi",
"suffix": ""
},
{
"first": "B",
"middle": [],
"last": "Sass",
"suffix": ""
}
],
"year": 2014,
"venue": "Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC'14)",
"volume": "",
"issue": "",
"pages": "1719--1723",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Oravecz, C., V\u00e1radi, T., and Sass, B. (2014). The Hungar- ian Gigaword corpus. In Proceedings of the Ninth Inter- national Conference on Language Resources and Eval- uation (LREC'14), pages 1719-1723, Reykjavik, Ice- land, May. European Language Resources Association (ELRA).",
"links": null
},
"BIBREF19": {
"ref_id": "b19",
"title": "Efficient web crawling for large text corpora",
"authors": [
{
"first": "V",
"middle": [],
"last": "Suchomel",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Pomik\u00e1lek",
"suffix": ""
}
],
"year": 2012,
"venue": "Proceedings of the seventh Web as Corpus Workshop (WAC7)",
"volume": "",
"issue": "",
"pages": "39--43",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Suchomel, V. and Pomik\u00e1lek, J. (2012). Efficient web crawling for large text corpora. In Adam Kilgarriff et al., editors, Proceedings of the seventh Web as Corpus Work- shop (WAC7), pages 39-43, Lyon.",
"links": null
},
"BIBREF20": {
"ref_id": "b20",
"title": "KorKorpusz: k\u00e9zzel annot\u00e1lt, t\u00f6bbr\u00e9teg\u0171 pilotkorpusz\u00e9p\u00edt\u00e9se",
"authors": [
{
"first": "N",
"middle": [],
"last": "Vad\u00e1sz",
"suffix": ""
}
],
"year": 2020,
"venue": "XVI. Magyar Sz\u00e1m\u00edt\u00f3g\u00e9pes Nyelv\u00e9szeti Konferencia (MSZNY 2020)",
"volume": "",
"issue": "",
"pages": "141--154",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Vad\u00e1sz, N. (2020). KorKorpusz: k\u00e9zzel annot\u00e1lt, t\u00f6bbr\u00e9teg\u0171 pilotkorpusz\u00e9p\u00edt\u00e9se. In G\u00e1bor Berend, et al., editors, XVI. Magyar Sz\u00e1m\u00edt\u00f3g\u00e9pes Nyelv\u00e9szeti Konfer- encia (MSZNY 2020), pages 141-154, Szeged. Szegedi Tudom\u00e1nyegyetem, TTIK, Informatikai Int\u00e9zet.",
"links": null
},
"BIBREF21": {
"ref_id": "b21",
"title": "Hungarian Dependency Treebank",
"authors": [
{
"first": "V",
"middle": [],
"last": "Vincze",
"suffix": ""
},
{
"first": "D",
"middle": [],
"last": "Szauter",
"suffix": ""
},
{
"first": "A",
"middle": [],
"last": "Alm\u00e1si",
"suffix": ""
},
{
"first": "Gy",
"middle": [],
"last": "M\u00f3ra",
"suffix": ""
},
{
"first": "Z",
"middle": [],
"last": "Alexin",
"suffix": ""
},
{
"first": "J",
"middle": [],
"last": "Csirik",
"suffix": ""
}
],
"year": 2010,
"venue": "Proceedings of LREC 2010",
"volume": "",
"issue": "",
"pages": "",
"other_ids": {},
"num": null,
"urls": [],
"raw_text": "Vincze, V., Szauter, D., Alm\u00e1si, A., M\u00f3ra, Gy., Alexin, Z., and Csirik, J. (2010). Hungarian Dependency Tree- bank. In Proceedings of LREC 2010, Valletta, Malta, May. ELRA.",
"links": null
}
},
"ref_entries": {
"FIGREF0": {
"text": "The abstract workflow of web corpus creation. Parallelogram-shaped boxes denote the optional phases, the grey background denotes the produced data.",
"uris": null,
"num": null,
"type_str": "figure"
},
"TABREF2": {
"text": "The extracted text from different parts of the HTML with different tools in million words. Newspaper3k and our method is displayed with and without metadata.",
"content": "<table><tr><td/><td>Full</td><td>Article</td></tr><tr><td/><td>HTML</td><td>Body</td></tr><tr><td>Newspaper3k (precision)</td><td>0.77</td><td>0.69</td></tr><tr><td>Newspaper3k (recall)</td><td>0.52</td><td>0.26</td></tr></table>",
"type_str": "table",
"num": null,
"html": null
},
"TABREF3": {
"text": "",
"content": "<table><tr><td>: The content-wise comparison of metadata (author,</td></tr><tr><td>title, publication date) extracted by Newspaper3k and our</td></tr><tr><td>method (=1.0).</td></tr></table>",
"type_str": "table",
"num": null,
"html": null
}
}
}
} |